Professional Documents
Culture Documents
Methods in Evidence
Based Healthcare
edited by
Andrew Stevens, Keith Abrams,
John Brazier, Ray Fitzpatrick
and Richard Lilford
Edited by
ANDREW STEVENS
KEITH ABRAMS
JOHN BRAZIER
RAY FITZPATRICK
and R I C H A R D L I L F O R D
SAGE Publications
London • Thousand Oaks • New Delhi
Editorial arrangement, Introduction and Part Introductions
© Andrew Stevens, Keith R. Abrams, John Brazier,
Ray Fitzpatrick and Richard J. Lilford 2001
Chapters 1–27 © the contributors 2001
Contributors ix
Introduction
Methods in Evidence Based Healthcare and Health Technology
Assessment: An Overview 1
Andrew Stevens, Ray Fitzpatrick, Keith Abrams, John Brazier
and Richard Lilford
Introduction 91
Ray Fitzpatrick
Introduction 179
John Brazier
Introduction 273
Keith R. Abrams
Introduction 367
Andrew Stevens and Keith R. Abrams
Introduction 449
Andrew Stevens
Index 499
Contributors
Her research interests include provision of services and informal care to older
people with dementia, and the measurement of outcomes of community care.
She is currently a research fellow at the Social Policy Research Unit at the
University of York working on a Department of Health-funded programme of
research into ways of collecting and using information on the outcomes of
community care in routine practice.
Andrew Black was educated in Pakistan, Scotland and Oxford. He did his
Postgraduate clinical training in Oxford, Gloucester, Sydney and Auckland.
He was a Lecturer in Anaesthesia at the University of Oxford from 1977–1982.
Since then he has been a Consultant Senior Lecturer in Anaesthesia at the
University of Bristol. His academic interests are in pulmonary gas exchange,
control of breathing, postoperative analgesia, clinical applications of classifica-
tion approaches including statistical multi-variable and multi-variate modelling
and neural networks, applications of medical simulation and role play in
medical undergraduate education.
Nick Black is Professor of Health Services Research (since 1995) and Head
of the Department of Public Health and Policy at the London School of
Hygiene and Tropical Medicine (since May 1998). After qualifying in
medicine from Birmingham University in 1974, he worked in NHS hospitals
for two-and-a-half years before joining Save the Children Fund and running a
child health programme in Nepal for 18 months. He then underwent post-
graduate training in public health at Oxford including a doctorate on reasons
for the epidemic in surgery for glue ear in the UK. The next three years were
spent half time as a lecturer at the Open University writing a new distance-
learning course ‘Health and Disease’ with a biologist, sociologist and
economist, and half time as a Consultant in Public Health for Oxfordshire
Health Authority. In 1985 he moved to a Senior Lectureship at the LSHTM.
His main research interests are the evaluation and audit of health services
(particularly in the field of surgery), non-randomised methods of evaluation,
R&D policy, and organisational research. He is involved in several aspects of
the NHS R&D programme at national and regional levels. In 1996, together
with Nick Mays (a Visiting Professor in PHP), he founded the Journal of
Health Services Research & Policy. He was promoted to a Readership in 1993
and to his current Chair in 1995.
the uses of routinely assembled data for health services planning and
evaluation, including methods of their analysis and interpretation.
Alison Eastwood is currently a senior research fellow at the NHS Centre for
Reviews and Dissemination. Her background is in statistics and economics,
and throughout her career she has been involved in health services research.
Prior to joining the Centre over five years ago, she was a biostatistician on the
myocardial infarction Patient Outcomes Research Team at Harvard Medical
School in the US.
Health and Epidemiology she completed a review of ethics and RCTs for
the NCCHTA.
rising demand for healthcare, and the evaluation of attempts to meet it. In
1998 he published the SWOOP study, the first ever randomised controlled
trial of a nurse telephone consultation system in primary care, which formed
the evidence base for the new NHS Direct telephone service. He was also one
of the developers of the EORTC PAN-26, a questionnaire for measuring
outcomes in people with carcinoma of the pancreas and, with Professor Mike
Campbell, has published on sample size calculations and analysis of quality
of life data.
Richard Lilford (editor) was born in Cape Town in 1950, received his
schooling in Johannesburg, and became a doctor in that city. He moved to
Contributors xxi
Julia Lowe completed her medical training in the UK and then moved to
Australia. She worked part-time in private practice while her children were
small, completing a masters degree in clinical epidemiology in her spare time.
She moved back to the public sector in 1991 and became Director of General
Medicine at John Hunter teaching hospital in 1996. She has established
research programmes in diabetes and cardiac failure as well as maintaining an
interest in health services research. She is criticism editor for the Cochrane
Heart Group.
of health risk assessment. The work has covered genetic testing in pregnancy,
adulthood and childhood, as well as population based screening programmes.
The conditions studied include heart disease, cervical, breast and bowel
cancer and cystic fibrosis. The aim of this research is to understand responses
as a first step towards evaluating different methods of communicating
information to promote understanding, reduce emotional distress and enhance
health-promoting behaviours. She has published over 100 peer-reviewed
articles in this and related areas and is co-editor of The Troubled Helix (1996;
paperback edition: 1999), a book reviewing the psychological implications of
the new human genetics.
Elaine McColl has a first degree in Geography and Economics and an MSc in
Applied Statistics. She is a native of Dublin and worked for several years in
the Department of Statistics, Trinity College, Dublin, where she taught formal
courses and provided informal advice on survey design and analysis. Since
1987, she has been a member of the Centre for Health Services Research,
University of Newcastle upon Tyne, where she currently holds the post of
Senior Research Associate. Her main research interests are in the assessment
of health-related quality of life and in survey methodology. She has recently
been awarded an NHS Primary Care Career Scientist award, to develop a
programme of research into quality of life in chronic disease. Elaine was
principal investigator for a recently completed literature review Designing
and using patient and staff questionnaires which was funded by the NHS
Health Technology Assessment Programme and upon which the findings
reported here are based.
Lorna McKee is Director of Research at the Department of Management
Studies, University of Aberdeen and was previously principal researcher at
the Centre for Corporate Strategy and Change, University of Warwick. She is
joint author of Shaping Strategic Change (1992) with Andrew Pettigrew and
Ewan Ferlie, and has published in an extensive range of academic and
professional journals. She has recently investigated medical leadership,
primary care management and biomedical innovation transfer processes in
research projects funded by the ESRC, CSO and NHS. She regularly acts as
a consultant to healthcare organisations across the United Kingdom, and
jointly co-ordinates the Multi-Disciplinary Primary Care Leadership Pro-
gramme leading the module on Developing and Leading Effective Teams.
Martin McKee is Professor of European Public Health at the London School
of Hygiene and Tropical Medicine and co-director of the School’s European
Centre on Health of Societies in Transition. His main research interest is
health and healthcare in central and eastern Europe and the former Soviet
Union, but he also has a more general interest in the interpretation and use of
evidence in policy-making.
Klim McPherson is Professor of Public Health Epidemiology in the Cancer
and Public Health Unit at the London School of Hygiene and Tropical
Contributors xxiii
Medicine. He has written over 300 refereed papers during his 30-year career
in epidemiology. His particular interests include the relationship between
breast cancer and hormone supplements such as HRT and oral contraceptives,
the incidence of coronary heart disease, the effects of tobacco and alcohol and
the widening of public health training to include those who do not come
from a conventional and traditional medical background. He is a member of
the editorial board of various refereed journals, including the Journal of
Epidemiology and Community Health, and his membership of different
organisations includes the Board of the Faculty of Public Health Medicine
and the Council of the UK Public Health Association. He has been the author
and co-author of many books and book chapters such as Hormones and
Cancer, ed. by O’Brien and MacLean, published by the RCOG Press
(Chapter 16), a chapter in Health issues related to consumption, ed. by Ian
MacDonald, published by Blackwell Science, and with Prakash Shetty, he
edited Diet Nutrition and Chronic Disease, Lessons from Contrasting Worlds,
published by John Wiley & Sons.
David Moher is an Assistant Professor in the Departments of Pediatrics,
Epidemiology and Community Medicine and Medicine, Faculty of Medicine,
University of Ottawa, and Director, Thomas C. Chalmers Centre for System-
atic Reviews at the Children’s Hospital of Eastern Ontario Research Institute.
During the past ten years his interests have focused on how randomised trials
and meta-analyses of these studies are conducted and reported. He has been
particularly interested in how biases individually and collectively influence
estimates of treatment efficacy.
Michael Moher is currently a GP and Research Fellow at the Department of
Primary Health Care at the University of Oxford. He is currently involved as
medical co-ordinator, in research on the evidence based secondary prevention
of coronary heart disease in primary care. This research is comparing, in a
randomised controlled trial, three methods of promoting secondary prevention
of heart disease in primary care: audit and feedback, the introduction of
structure of records, registers and recall, and the introduction of nurse-run
clinics.
Graham Mowatt has, since April 1997, been Review Group Co-ordinator for
the Cochrane Effective Practice and Organisation of Care Group (EPOC).
The focus of EPOC is on producing systematic reviews of interventions
designed to improve health professional practice and the delivery of effective
health services. This includes various forms of continuing education, quality
assurance, informatics, financial, organisational and regulatory interventions
that can affect the ability of health professionals to deliver services more
effectively and efficiently. Prior to working on the NHS Health Technology
Assessment Programme project on assessing fast-changing technologies he
was employed by Grampian Health Board for four years on a health service
project which was concerned with primary care technological innovation. He
also possesses a Master of Business Administration (MBA) degree.
xxiv The Advanced Handbook of Methods in Evidence Based Healthcare
thesis examines the use of different sources for identifying new healthcare
technologies prior to their widespread adoption by the NHS, and makes
recommendations regarding the establishment and operation of an early
warning system in the United Kingdom.
Paul Roderick trained in general medicine and then public health medicine.
He has been senior lecturer in Public Health Medicine at Southampton
University since 1993. His main research interests are in evaluating the equity
and effectiveness of healthcare delivery, and in defining the epidemiology of
chronic renal failure. He is a member of the Steering Committee of the UK
Renal Registry.
Ian Russell has been Founding Professor of Health Sciences at the University
of York since 1995. He was educated at the Universities of Cambridge
(graduating in Mathematics), Birmingham (Statistics) and Essex (Health
Services Research). He has held academic appointments in the Universities of
Newcastle upon Tyne, North Carolina, Aberdeen (as Director of the Scottish
Health Services Research Unit) and Wales (while Director of R&D for NHS
Wales).
Jennifer Soutter has a first degree in English and Psychology; her PhD is
titled ‘Archetypal Elements in the Poetry of Sylvia Plath’. Returning to work
after completing six years of higher education, she has worked in health
services research for the past ten years, over six of these years being at the
Centre for Health Services Research at the University of Newcastle upon
Tyne. Her current post is as research associate at the Department of Child
Health at the Royal Victoria Infirmary working on a project with children
diagnosed with Duchenne muscular dystrophy. This links with work carried
out with children with life-limiting illnesses some six years ago.
Nick Steen is a medical statistician based at the Centre for Health Services
Research at the University of Newcastle upon Tyne with special interests in
the design and analysis of cluster randomised trials and the development of
health outcome measures. He is currently involved with a number of trials
involving the assessment of health technology.
psychiatry and general practice. His previous research posts were at the
Department of General Practice and the Department of Public Health Sciences,
both within GKT School of Medicine. Work on Chapter 17 was completed
whilst based at the Department of Public Health Sciences.
Marjon van der Pol is an economist who works for the Health Economics
Research Unit at the University of Aberdeen.
EVIDENCE BASED MEDICINE, EVIDENCE many commonly used treatments and investiga-
BASED HEALTHCARE AND HEALTH tions in contemporary healthcare systems have
not been shown to be effective in terms of clear
TECHNOLOGY ASSESSMENT and convincing evidence. A growing literature
on geographical variations, including Wenn-
Evidence based medicine, evidence based health- berg’s notable studies in New England, has
care and health technology assessment are terms underlined how diverse are health professionals’
representing a number of important recent devel- ways of managing similar health problems and
opments in healthcare. The phrase ‘health tech- how great is the deficit of, and the gap between
nology assessment’ was coined by the United scientific evidence and clinical practice.4
States Office of Technology Assessment in 1972 Sackett (1996) describes evidence based
as part of a wide-ranging interest on the part of medicine as ‘the conscientious, explicit, and
Congress in the evaluation of new technologies judicious use of current best evidence in making
across not just healthcare, but all major sectors decisions about the care of individual patients’.5
of importance to the US economy.1 Elsewhere, The term ‘evidence based healthcare’ extends
the term was picked up a decade or so later, such the scope to those making decisions about not
that by 1985, there was an International Society just patients but also populations. By widening
for Health Technology Assessment, an inter- the audience, it has also widened the scope of the
national organisation aiming to promote scien- subject matter to include not only the evidence for
tifically based assessment of technologies in the effectiveness of elements of healthcare, but
healthcare, including drugs, devices, medical also its costs and other ramifications. Further-
and surgical procedures, as well as organisa- more, it is acknowledged in both health technol-
tional and administrative systems. The UK ogy assessment and evidence based healthcare,
adopted the term ‘health technology assessment’ that the health ‘technologies’ of interest are very
with the funding in 1993 of a specific pro- wide, concerning not only pharmaceuticals and
gramme under that name as part of its National healthcare devices, but also procedures, settings,
Health Service Research and Development healthcare personnel and more diffuse innova-
strategy.2 tions including, for example, managerial, IT and
‘Evidence based medicine’ has its roots in clin- policy changes in general. In short, evidence
ical epidemiology, owing much to Cochrane’s based healthcare and health technology assess-
(1972) critique of the effectiveness and effi- ment have converged to cover an interest in the
ciency of health services.3 Cochrane argued that costs and benefits of all activities within the
2 The Advanced Handbook of Methods in Evidence Based Healthcare
healthcare system, and indeed even those for FEATURES OF THE RESEARCH BASE
which responsibility lies outside the boundaries
of the healthcare service.
It is now widely recognised that research for
health technology assessment includes not just
primary research (data gathering), but also sec-
THE HEALTH TECHNOLOGY ASSESSMENT ondary research (a systematic review of primary
REVOLUTION research evidence). Indeed, as the latter becomes
more formalised, disciplined and by conse-
The growth of health technology assessment and quence narrow, tertiary research synthesising
evidence based healthcare has lately become and making accessible multiple primary and
something of a revolution. There has been an secondary research products is also developing.
explosion of journals, conferences, books and The critical features of good primary research
research expenditure on health technology are being increasingly recognised as indispens-
assessment. In almost every OECD country, it able: appropriately posed questions, studies
has become clear that the health services cannot designed to reduce bias, and with patient-
endlessly sustain an acceptance of all innova- relevant outcomes, conducted without bias,
tions, whether they have been evaluated or not, ethically, and with suitable quality control,
and whether they are affordable or not. Health drawing conclusions which flow from results,
Technology Assessment (HTA) has become an and published regardless of the outcome.
important part of a strategy to contain costs and In the same way the critical features of sec-
maximise value from available healthcare ondary and tertiary research in HTA are agreed:
resources. This has led to some countries devel- systematic and reproducible reviews using clear
oping guidelines for the reporting of economic search strategies, formal critical appraisal of
evaluations of pharmaceuticals and other tech- contributory studies, grading of the evidence
nologies, such as in Canada,6 Australia7 and from these studies according to their quality and
the UK.8 A single new pharmaceutical, beta- size, selection of relevant outcomes from pri-
interferon – for multiple sclerosis – has the mary research, and valid methods for integrating
capacity to cost the UK National Health Service the evidence.10
the GDP of many small countries, and cost the
United States Health Services the GDP of a
number of large ones. A similar story is true of
an increasing armoury of new pharmaceuticals. THE METHODOLOGY OF HEALTH
The effectiveness of not just new technolo- TECHNOLOGY ASSESSMENT
gies, but also of many established ones, has been
widely questioned. The problem has been one of The growth of demand for health technology
both ineffective healthcare being promulgated assessment does not of itself guarantee the
into health services, and of therapeutically appropriateness and validity of the material it
advantageous techniques long being ignored. generates. An interest in methods employed to
Antman et al. (1992) demonstrate examples of ensure the quality of evidence has emerged. This
both, with huge time lags between the arrival of is manifest, for example, in the number of
conclusive evidence against antidysrrhythmics, different professions associated with HTA,
and supportive of thrombolytics in myocardial including epidemiologists, health economists,
infarction, and the recognition of these facts in sociologists, statisticians, clinicians and others.
medical text books, quite apart from clinical Each is in a position to supplement – and to
practice.9 some extent confront – the methods used by the
So there has been a sea change in thinking others. And the fertile mix of ideas is becoming
about health care, both clinically and at a policy increasingly apparent. A particular manifesta-
level. There has been a proliferation of attempts tion of the interest in the methods of HTA was
to satisfy the demand for evidence based health- the United Kingdom’s decision specifically to
care, covering the growth and scope of interest have a programme of research on the method-
of funding bodies, the development of network- ology of HTA. It posed a series of questions
ing and co-ordinating bodies, the development which fit into the following broad categories:
of multiple new publications and dissemination
media (from new journals to electronic data- 1 Methods in clinical trials,
bases to compendia of electronic databases), and 2 The place of observational and qualitative
systems for developing skills in interpreting research,
evidence. Most importantly, there have been a 3 The measurement of benefits and costs,
number of innovations in the development of the 4 Analytical (statistical) methods which quan-
research base. tify the uncertainty of the findings,
Introduction 3
Having chosen the technologies and questions of How should new data be collected –
interest, and before gathering data afresh, we through new randomised controlled trials,
must ask how much data/evidence is there or by collecting data on observed activity?
already? (Chapters 6 and 7, Observational methods)
f
Can we be sure we have all relevant evi-
dence? (Chapter 21, Publication bias) If observational methods are deemed
adequate, will routine data do? (Chapters 8
f and 9, Routine data)
Having gathered the existing evidence, how f
can we synthesise it? (Chapter 22, Meta-
analysis) Or might qualitative methods suffice in some
f circumstances? (Chapter 10, Qualitative
methods)
Can we rely on the syntheses? (Chapter 23, f
Assessing the quality of randomised con-
trolled trials in meta-analysis)
Or not using new data at all, but making the
best of consensus in some instances? (Chap-
If the existing data are unsuitable or inadequate, ter 24, Consensus methods)
it follows that new data need to be collected.
f
What measurements are most useful to
make? (Chapter 11, Patient-based outcome If new trials are decided upon – how gen-
measures) eralisable are they? (Chapter 4, General-
isability)
f
f
How should questionnaires be used to gen-
erate these data? (Chapter 15, Question- What are the limiting factors in trial design?
naires) (Chapter 3, Limiting factors in trial design)
f f
4 The Advanced Handbook of Methods in Evidence Based Healthcare
Whatever the methods used for gathering The contributions to this volume, and indeed the
data, how can we best analyse and interpret editorial collaboration in collating the volume,
it? (Chapter 16, Bayesian methods) have distinctive origins that are worth under-
lining. The UK National Health Service estab-
f lished a national Research and Development
(NHS R&D) initiative of which the Health
When data have been collected on both Technology Assessment (HTA) Programme has
quality and length of life how should they been a major component.2 It is hoped that sev-
be analysed? (Chapter 20, Quality of life eral features make this an important develop-
and survival analysis) ment in international, as well as national, terms.
In the first place, an elaborate process was
f established for inviting ideas for research from
the widest range of health professionals, re-
How large should the sample sizes be of the searchers, consumer interest groups and others.
newly collected data? (Chapter 19, Sample Secondly, expert panels were asked carefully to
size issues) develop and prioritise research ideas submitted.
Thirdly, research that has been prioritised for
f funding is commissioned by advertising to
researchers throughout the world to submit ten-
Are there special study designs required ders, with extensive peer review to determine
when the population receiving an inter- research groups to take forward projects. Great
vention is clustered into schools, practices emphasis has been placed on the need for inter-
or other units (Chapter 17, Area-wide disciplinary strengths for research groups to be
interventions) commissioned to undertake work for the HTA
programme. The methodology programme of
It can be seen that this sequence, in order to research has been a distinctive strand of the
follow a logic of asking what the question is and HTA programme as a whole. Because of the
whether it has yet been answered, disrupts the extensive process of consultation to identify
contents order of the book. In the main volume methodological questions and of selecting invest-
we have stayed with the familiar logic of trials igators, and because of the thorough peer-
first, horizons last, but either order would be reviewing of outputs, we are confident that the
sustainable. contributions reported here can be considered an
authoritative summary of current methodo-
logical issues in HTA. Every chapter is a sum-
METHODS FOR UNDERTAKING mary of a more extensive study that has gone
through many steps of evaluation before emerg-
METHODOLOGICAL RESEARCH ing to form a monograph for the NHS R&D
programme. Efforts were also made to achieve
All of the contributions in this volume have had coherence to this body of methodological work
to select methods to generate their suggestions by workshops and other forms of communica-
for methodology in HTA. Principally these have tion to share experiences and results.
been those of synthesising the evidence from This volume has not set out to provide readers
other pieces of work. This can take the form of with a basic introduction to the elementary
either synthesising primary research (e.g. a aspects of topics such as the design of random-
review of studies seeking to compare consent ised controlled trials or cost–benefit analysis
rates for clinical trials by method of invitation) that can be found elsewhere. It does, however,
or synthesising arguments (e.g. arguments over attempt to tackle the frontiers of health tech-
whether a fixed or random effects statistical nology assessment. The book brings together a
model is more appropriate for meta-analysis). In unique collection of contributions that will guide
some cases authors have also sought to collect the reader through current thinking and debates
Introduction 5
about the state and direction of research methods 5. Sackett D, Rosenberg W, Gray J, Haynes R,
in health technology assessment. If the volume Richardson W. Evidence based medicine: what it
stimulates further debate as how to improve our is and what it isn’t. Br. Med. J. 1996; 312: 71–2.
methods of health services research, it will have 6. Canadian Coordinating Office for Health Tech-
achieved its goal. nology Assessment. Guidelines for economic
evaluation of pharmaceuticals: Canada. 2nd ed.
Ottawa: Canadian Coordinating Office for Health
Technology Assessment (CCOHTA); 1997.
REFERENCES 7. Commonwealth Department of Health, Housing
and Community Service. Guidelines for the
pharmaceuticals industry on the submission to the
1. US Congress Office of Technology Assessment.
Pharmaceutical Benefits Advisory Committee.
Identifying Health Technologies That Work: Canberra: Australian Government Publishing
Searching For the Evidence. Washington: US Service, 1992.
Government Printing Office, 1994. 8. National Institute for Clinical Effectiveness:
2. Department of Health NHS Executive. Research Appraisal of new and existing technologies:
and Development In the New NHS: Functions Interim guidelines for manufacturers and sponsors.
and Responsibilities. London: Department of December 1999, London: NICE http://www.nice.
Health, 1994. org.uk/appraisals/appraisals.htm.
3. Cochrane A. Effectiveness and Efficiency Random 9. Antman E, Lau J, Kupelnick B, Mosteller F,
Reflections on the Health Service, Leeds: Nuffield Chalmers T. A comparison of results of meta-
Provincial Hospitals Trust, 1972. analyses of randomised control trials and recom-
4. Wennberg JE, Freeman JL, Culp WJ. Are hospital mendations of experts. JAMA 1992; 268: 240–8.
services rationed in New Haven or over-utilised 10. Chalmers I, Altman D. Systematic Reviews.
in Boston? Lancet 1987; i: 1185–9. London: BMJ Publishing, 1995.
Part I
CLINICAL TRIALS
I N T R O D U C T I O N by R I C H A R D J . L I L F O R D
and A N D R E W S T E V E N S
The concept of control groups is mentioned as moderate gains which collectively contribute to
early as the book of Daniel and was formalised major overall improvements in welfare. These
by Sir Francis Bacon in the 17th century. James modest but worthwhile gains cannot be detected
Lind conducted a famous trial of different meth- easily – unlike the few massive breakthroughs.
ods of prophylaxis against scurvy, and discov- Moderate biases may obscure or exaggerate
ered the benefits of fruit for sailors.1 This was such gains or even create the impression of
not a randomised experiment, and although the benefit in the face of harm. RCTs, blinded when
idea of allocation of therapy by ‘lot’ is an old possible, provide protection against these mod-
one, practical application is really a 20th century erate but relatively important biases.4 Random-
phenomenon. The statistics were worked out by isation provides protection against selection bias
Ronald Fisher in an agricultural context, and a – inaccuracy that occurs if patients in treatment
landmark randomised trial in humans was Sir and control groups have different capacities to
Austin Bradford Hills’ study of streptomycin benefit. Randomisation, provided that the alloca-
treatment for tuberculosis, conducted during the tion is fully concealed, ensures that confounding
Second World War.2 variables are distributed by chance alone. Blind-
The Randomised Controlled Trial (RCT) is ing prevents the possibility of performance bias
the archetypal primary research method of (augmented or impaired outcome resulting from
Health Technology Assessment (HTA). About a administration of different treatments, apart
third of a million RCTs are registered on the from the treatment of interest, to study or con-
Cochrane database of controlled clinical trials, trol patients) and outcome bias (distorted out-
and this list falls far short of covering the entire come, involvement or interpretation in the light
literature base on RCT methodology.3 RCTs of hope and expectation).
have given some very clear pointers to practice: It has been shown empirically that random-
clot busting and antiplatelet drugs are effective ised controlled trials in general produce more
in heart attack, multi-agent chemotherapy delays modest estimates of treatment effects than his-
the onset of AIDS among carriers of HIV virus, torically controlled studies. Good RCTs – studies
and antibiotics are effective in surgery where a which obey the established scientific criteria –
contaminated viscus is breached. Some treat- produce different results from poorly conducted
ments have such spectacular effects that their studies. Well-conducted RCTs are, it is widely
effectiveness can be established without the agreed, the most reliable method for comparing
need for RCTs – volume replacement for most health technologies.5 There may be ques-
massive haemorrhage, anti-D to prevent rhesus tions about their external validity (generalis-
sensitisation in newborns, and use of the supine ability) and about their practicability in various
position to prevent ‘cot death’. However, most contexts, but they remain the bedrock of much
advances in medicine are less spectacular – HTA.
8 The Advanced Handbook of Methods in Evidence Based Healthcare
Randomisation is a simple concept, but RCTs and considers whether the doctrine of informed
have spawned a wide range of complex issues consent – so crucial in the West – may be
that challenge those who commission and con- viewed somewhat differently in cultures where
duct applied research. Large simple trials have an egalitarian relationship between professional
led to some very important discoveries such as and patient is less widely accepted.
the value of magnesium in the treatment for Ross and colleagues then discuss the quality
eclampsia. However, the range of possible inter- of trials. The factors impeding recruitment are
actions (and hence of potential interactions classified as those affecting clinicians’ willing-
between interventions) is rising exponentially, ness to collaborate in a study, their willingness
and it will not be possible to answer all research to offer patients entry and patients’ willingness
questions using the large simple philosophy. to participate, given the offer. Patients seem less
Factorial designed studies (where patients are willing to participate in trials where the compar-
re-randomised to different therapies), families of ator treatments are very different in their nature
trials (where patients are offered different trials and immediate effects – surgery versus radio-
at different stages of their disease according to therapy for prostate cancer is a recent example.
eligibility), and trials nested within prospective The importance of separating the ‘business’ of
frameworks for collection of routine data, will running trials from their conception and instiga-
all be necessary. Some therapies are given to tion is stressed. This ensures that those who may
groups of people, not individuals, and this calls have an emotional or pecuniary interest in the
for cluster trials while at the other end of the results of a trial are kept at arm’s length from
scale, patients may act as their own controls in the processes and analyses which may bias the
n 5 1 trials. Some interventions are complex results – for example, interference in the ran-
‘packages’ made up of many components, where domisation sequence is prevented.
measurement of ‘fidelity’ to the intended pack- However, a strong recommendation from the
age is useful to those who may wish to adopt the authors is that trials should be kept simple so
results. Trials of rapidly evolving technologies that logistic difficulties do not act as a strong
are an increasing concern given the pace of disincentive to participation. Reporting of results,
technological development – the question is so that the link between outcome and allocated
when to start such studies, and one approach is
treatment is clear, is also discussed in this chap-
to offer randomisation at an early stage, accept-
ter and the agreed methodology for reporting
ing that the technology might continue to
studies (CONSORT) is emphasised.7
improve for some time after the start of the
study. Another (often related) problem arises O’Connell and colleagues then consider the
from interventions whose effects are critically problem of relating results from randomised
dependent on practitioner skill or organisational trials to patients in another time and place. Most
factors – one approach is to ensure recruitment authors address these issues by describing the
from a very broad sample, and another is to inclusion and exclusion criteria, subgroup analy-
randomise both patient and practitioner. All sis and (less often) comparison of trial partici-
these refinements have cost implications, and a pants with those not in the trial. However, there
compromise may be necessary between the ideal are strong arguments for more individualised
and the achievable, in this as in other aspects of assessment of benefits and harms. Clinical trials
life. In short, the study of trial design is still can be thought of as a method to produce good
evolving. empirical data with which to populate risk/
This section deals firstly in the ethics of benefit models, in which probabilities of good
RCTs. A general review and discussion of the and bad outcomes, and the values attached to
ethics of RCTs, by Lilford and colleagues con- these outcomes are made explicit. There is also
siders the main philosophical positions and a great need for more empirical work on the
shows that from two points of view – utilitarian transferability of results from one group of
and deontological – the essential requirements patients to another, especially in situations
are that the treatments being compared should where there is no obvious reason why they
be an equal bet in prospect, the situation of might differ. For example, there may be no prior
patient equipoise where the expected utilities of reason to expect results to vary by sex, race or
each treatment on offer are equivalent.6 Clearly age. It is difficult to know what to do when
the competent patient is in the best position to statistically significant subgroup effects arise.
make this decision, given available ‘prior’ An essential principle, however, is that even if
probabilities; hence there is no conflict between relative risks are fixed across different sub-
maximising welfare and respecting autonomy, in groups, absolute risks may vary. This is where
this area of ethics at least. In the subsequent modelling comes into its own, since it enables
chapter, Ashcroft shows how these principles results to be extrapolated to those with a differ-
may be interpreted in different moral traditions ent base-line prognosis.8
Part I: Clinical Trials 9
best interests of current patients who participate often referred to as ‘probabilistic dominance’)
in them and this insensitivity is at odds with the because all values are equal to unity. Under
model doctor–patient relationship.4–6 We will these circumstances, equipoise exists if the
see that, under certain conditions, the societal expected effectiveness of A and B are equal, or,
interest can be promoted without threat to the more formally (and in Bayesian notation) if the
individual’s own best interests. ‘priors’ balance on the same point.
In this chapter, we will analyse the idea that However, we must acknowledge that treat-
the tension between societal and individual ments typically have unequal side-effects a
interests can be resolved by making recourse to priori.9 A, for example, might be mutilating
the epistemological ‘uncertainty’ principle.7 Put surgery, while B is a smaller operation or an
simply, if the clinician does not know which inexpensive drug with few side-effects. In order
treatment is best, then the patient is not dis- for the treatments to be equally desirable under
advantaged in prospect by being randomised and these conditions (i.e. for equipoise to exist), A
society benefits from the knowledge gained. must be better than B, in prospect, on the major
However, it is necessary to clarify what ‘knowl- outcome so that superior effectiveness can be
edge’ and ‘best interests’ mean before they can traded off against worse side-effects. In decision
be properly lodged with ethical theory and then analytic language, the expected utilities of A and
operationalised by applying them to practical B should be the same; the expected advantages
difficulties with trials and their design. of A on the main outcome measure are precisely
offset by the lower side-effects of B. Equipoise
under these circumstances occurs when A is
Knowledge expected to be superior (on the main outcome
measure) to B. If the expected difference in
Dealing first with ‘knowledge’, consider first
outcome equals, but does not exceed the trade-
two treatments (A and B) with equivalent (or
off that individual patients require, then the
no) side-effects. Knowing can have two mean-
patient is in individual equipoise. Of course,
ings: firstly, that there is no preference between
there may be many outcomes where the relative
two treatments, since they have identical side-
effects of the treatment are uncertain. However,
effects and A is as likely to be better than B or
this does not change the fundamental point, that
vice versa. Secondly, that there is some prefer-
a decision analysis can be constructed around
ence in the sense that one of the treatments is
certain effects (such as the mutilation of sur-
more likely to be superior, but this is uncertain;
gery), uncertain effects (for which ‘prior’ prob-
treatment A has not been ‘proven’ to be more
abilities exist, albeit within wider credible inter-
effective than B. Indeed, Gifford observes that
vals), and the values that might be ascribed to
most arguments which use uncertainty as justifi-
these various outcomes. If we leave aside, just
cation for RCTs equivocate on the term ‘know’,
for a moment, any potential negative or positive
sometimes giving the first interpretation above,
effects that the trial itself might have, we may
and sometimes the second.8 Uncertainty as the
argue that the patient is not disadvantaged
opposite of certainty, allows a clinical trial to
relative to routine care, provided the expected
proceed in the face of considerable preference.
utilities of the comparative treatments are
However, if there is ‘absolute’ uncertainty, the
equivalent. The subject of decision analysis – or
decision-maker is ‘agnostic’ or ‘in equipoise’ –
expected utility theory – is fully described
that is, the benefits in prospect for both treat-
elsewhere.10
ments are equal, and the outlook for the patient
is the same, whether or not a trial is conducted.9
For those familiar with Bayesian thinking, this
situation would arise where the point of balance
Equipoise: Collective or Individual?
of the ‘prior’ for the difference A 2 B was zero.
The term equipoise was made popular by Freed-
man in a landmark article in the New England
Best Interests Journal of Medicine,7 and we argue that its
theoretical meaning can be made explicit within
We need to clarify the meaning of ‘best’ in the a Bayesian and decision analytic framework.
context of the clinician’s obligation to give Freedman distinguished between two kinds of
patients the ‘best’ treatment. Two situations equipoise – collective and individual. By col-
apply. Firstly, the above situation where treat- lective equipoise, he meant that a society of
ments A and B (two treatments are discussed experts (say clinicians contemplating a particu-
here for convenience) have equal side-effects in lar trial) do not know which of two or more
prospect, and the major outcomes are binary treatments is best, although clinicians within the
(e.g. live or die). In this situation, A is preferred community may have different ‘hunches’. By
to B, provided it is more effective (a situation individual equipoise, he meant that a particular
14 The Advanced Handbook of Methods in Evidence Based Healthcare
clinician was equipoised, or that he had no beloved family member. Freedman defends his
hunch whatsoever. Using our definition would position on the grounds that doctors are judged
mean that this clinician regards the treatments as by, and should behave according to, collective
having equal expected utilities. Freedman goes norms.
on to argue that ethically collective equipoise However, this does not say how the doctor is
is a sufficient condition for recruiting patients supposed to operationalise the concept of col-
to trials, with the proviso that they consent, lective versus individual equipoise – is the doc-
irrespective of whether the individual clinician tor to tell the patient only that there is collective
in question does or does not have a prefer- uncertainty about the best treatment while keep-
ence. He starts by pointing out that individual ing any personal preference under wraps?
equipoise is a fragile state which might be very Besides, couching consent only in terms of
hard to find, given the exacting requirement collective uncertainty and not individual equi-
for the interaction between ‘prior’ probabilities poise leaves open the very real possibility that
and values to produce precisely equal expected the patient will interpret this to mean that the
utilities across the comparator treatments. clinician is personally equipoised, whereas the
Because individual equipoise is such an elusive clinician might mean only that the usual statis-
state in practice, Freedman argues, it can be tical conventions of ‘certainty’ have not been
discounted as too stringent a moral basis for fulfilled. Equivocation can lead to deception.12
trial entry. When individuals are equipoised, Failure to offer to disclose any such preference
this is simply a bonus, not a requirement. As at a minimum is an economy with the truth, and
scientists, we would like to go along with Freed- tantamount to deception. Deception, intentional
man, but we are conscious of serious flaws in his or otherwise, is a limit to autonomous decision-
argument. making13,14 and so Freedman’s allegiance to
collective equipoise does not sit well with his
position on autonomy: like most modern authors
Problems with Freedman’s Argument he strongly asserts the need to respect individual
choices. Such autonomy is exerted through dif-
The first point that Freedman ignores is that ferent people manifesting their own personal –
collective equipoise is seldom evenly balanced. and often very different – value systems. In
It, too, is a fragile state, only now among a Western cultures, it is a reasonable assumption
group of people. Typically, the numbers of that individuals are likely to have different
clinicians who might have preferences (hunches) values, perhaps as a result of their freedom to
for one treatment or another will be unequal in a choose how to live. People may make very
ratio of, say, 30 to 70. Freedman does not say different trade-offs. Given this, it follows that
how widespread a hunch can be before col- there can be no such thing as collective equi-
lective equipoise collapses into consensus over poise in any decision involving a trade-off –
which treatment is best. Just how many how can it simultaneously be argued that
clinicians need to prefer the trial treatment collective equipoise provides sufficient moral
before a trial of it becomes unethical? This was cover for clinical trials and that autonomy is a
investigated by Johnson and colleagues,11 who fundamental principle, when the latter implies
conducted an ethnometric study to identify the that two different people, confronted with
threshold of collective preference beyond which exactly the same scenario, may make very dif-
a trial would be unacceptable by members of the ferent choices! On the basis of the best ‘prior’
public, and showed that, once the ratio of num- probability estimates, one person may choose
ber of clinicians for the trial treatment to the treatment A, another treatment B, while a third
number of clinicians against it went beyond might be equipoised between A and B and hence
70:30, the great majority of respondents would willing (or eager) to be randomised. Therefore,
regard a trial of a potentially life-saving inter- we think that anybody who holds autonomy as a
vention as unethical. central principle must also regard equipoise as
However, there are yet further problems with the property of the patient. It arises when prior
Freedman’s argument, i.e. that where collective probability estimates and an individual patient’s
equipoise and individual equipoise do not co- values interact in such a way as to produce equal
incide, collective equipoise is a sufficient moral expected utilities.
basis upon which to recruit patients. He argues There is near-universal agreement that the
that the obligation of a doctor to a patient is ‘competent’ patient should be the source of
different from those which, say, a parent has these values.1,15,16 Probabilities are normally the
towards a child – Freedman thus argues that province of the care provider who has a moral
clinicians need not adhere to the ‘gold standard’, obligation to bring as much knowledge and
whereby one should not randomise a patient experience to bear on the problem as possible
unless one was prepared to do the same for a under the constraints of practice. In some cases,
Ethical Issues in Design and Conduct of RCTs 15
patients may wish to make their own judgement or desirable, but are not able to say exactly
on probability, and clinicians, in fully respecting where the most likely or desirable ‘point’ esti-
autonomy, should be prepared to engage in a mate lies. A clinician might say that the most
discussion of probabilities with patients who likely prior effect of a new operation on cancer
wish to do so. It is also worth noting that a null mortality lies between 4 and 6 percentage
‘prior’ does not produce equipoise in a situation points, and a patient might say that a gain
where a priori trade-offs are evident, and seems between 2% and 5% would justify the effects of
to challenge statistical convention of using such surgery. In this case, it could be argued, overlap
an hypothesis as a basis for data analysis. between ‘prior’ probabilities and values is a suf-
Sometimes, a patient who is terminally ill ficient basis for trial entry. Empirical investiga-
may wish to buy time and may accept some tion is needed on this point.
expected short-term losses that otherwise they
would not. A new treatment may give long-term
gains if it turns out to be beneficial and any The Altruistic Patient
losses will be short-lived. So, the probability
distribution of the expected benefits of the two The argument hitherto has presupposed that
trial arms need not always be symmetrical.9 patients wish to maximise their personal expected
utilities, but patients might wish to enter a study
for altruistic reasons.20 The empirical evidence
How Robust is Personal Equipoise? on this point suggests that patients participate
slightly more out of self-interest than out of
Given that patients should not lose out by taking altruism, although it is difficult to draw any
part in a clinical trial, it could be argued that, solid conclusions from the data.21 We would
since equipoise does not normally exist, clinical argue that altruism should not be assumed, but
trials are necessarily unethical.17 This argument any altruism should be voluntarily given, other-
is predicated either on a mathematical or prag- wise the burden of expectation on patients could
matic argument. The mathematical argument erode their trust in the core aims and object-
concerns itself with the extremely low likelihood ives of the profession, especially in life and
that values and ‘prior’ probabilities may interact death situations. Moreover, acknowledging that
in such a way as to provide precisely equal patients may include an element of altruism in
expected utilities. The pragmatic argument is their decisions to enter trials is very different
that in ordinary clinical practice, clinicians and from saying that doctors should request such a
patients do come off the fence and decide one gesture, still less couch their invitation to par-
way or the other – hence, the argument goes – ticipate in such ambiguous terms as to elicit
equipoise is a theoretical concept with no prac- unwitting self-sacrifice. Furthermore, recognis-
tical application. Alderson and colleagues, for ing that patients may wish to forego some
example, reported that 75% of physicians advantage for the good of others, does not
thought they were never in equipoise.18 How- change the fundamental nature of equipoise, as
ever, people may only come off the fence it simply means that the utility the patient gains
because they have to – in which case randomisa- in making such an altruistic gesture, could be
tion may provide a convenient and useful method factored into the expected utility equation.
to unblock an otherwise blocked decision. The It is disturbing to note, however, that many
first argument, concerning the likelihood of patients suspect that their clinicians may enter
finding a situation of exactly equal expected them in trials, even when equipoise is not pres-
utilities, is one that we have analysed in more ent, implying that a degree of altruism might
detail elsewhere.19 In brief, equipoise might not have been assumed by the use of collective
be as rare as the requirement for a mathematical uncertainty.21 The proportion of patients/mem-
dead hit between expected utilities would at first bers of the public, who thought it likely that
imply. It has been argued that it is unrealistic to doctors put people in trials even when physician
regard prior probabilities and value trade-offs as equipoise was not present, ranged from 26% to
exact and precise numerical values, given that 70% across two studies, both of which used
they are constructed in the mind. There may be hypothetical trial scenarios.22,23 Such variation
zones of indifference where the patient cannot might be explained in part by the populations
distinguish between the benefits of one treat- sampled: the general public and an exclusively
ment or another (or over which a clinician could African American population respectively.
not say that one probability is more likely than This fear is not unfounded, as a substantial
another). This argument is predicated in part on proportion of clinicians (in many countries) still
the hypothesis that in assigning probabilities seem to put their scientific duties ahead of their
or values, people are often able to articulate obligations to do their best for each individual
numerical values which are not the most likely patient. Alderson and colleagues, for example,
16 The Advanced Handbook of Methods in Evidence Based Healthcare
reported that 75% of physicians thought they 1 Giving people more information increases
were never in equipoise, at the same time as their knowledge up to a point, after which
53% of the same sample being prepared to enter confusion seems to set in. 33,36,51,56–58
their patients in a placebo-controlled trial of 2 Consent rates to trials are reduced by further
taxomifen.18 Likewise, 73% of responding sur- information, and giving a patient more time
geons, in another study, thought a trial would be to consider his or her response may also
ethical, even though only 28% were ‘uncertain’ reduce willingness to participate.33,51,58,59
about whether HRT could cause a recurrence of
cancer.24 It is possible, however, that they
thought equipoise could exist notwithstanding, Incompetent Patients: Whose Values?
as some patients may be prepared to trade-off
some survival advantage against the relief of When patients are incompetent, either temporar-
symptoms. Taylor explored this issue in more ily or permanently, they are unable to give their
detail, reporting that 36% of respondents would consent and so it may be unclear whose values
be prepared to enter their patients in a trial, even should be used in the decision-making process.
When a situation of incompetence has been
when out of equipoise.25
foreseen and the patient has made voluntary
advance directives for the specific situation in
hand, the doctor may take heed of the values,
Competent Decisions: Patient’s documented in such directives. A recent com-
Understanding prior to Assigning Values parative study suggests that practitioners’ atti-
tudes to advance directives generally are less
Before patients can assign values to an array of favourable than the patients’ own attitudes,
possible outcomes, clinical and psychological, which viewed such directives as a bona fide
they must be able to understand what these expression of their autonomy.60 However, there
outcomes mean. They also need to appreciate is no literature on the validity of advance direc-
that trial entry means that their treatment will be tives when it comes to entering RCTs, and we
allocated by lot, and not selected according to should be cautious about generalising the above
preference. Patients might find it difficult to data, taken in the context of routine practice, to
understand the information prior to consent trials.
because of the nature of the information itself, Sometimes the help of a relative or proxy will
or because of the way in which it is relayed.26 be enlisted. Such people may have further
A great many descriptive studies show that insight into what the patient may have wanted.
the conceptual nature of clinical trials (for ex- In the absence of such a person or in an emer-
ample, the importance of controls or the need gency, some have argued that randomisation
for randomisation) is much less readily under- should be avoided, notwithstanding the large
stood than technical details, such as the side- societal costs of doing so.61 Some Kantians
effects of various treatments.27–51 The true situ- would object to the use of proxies in such
ation might be still worse, since the quantitative circumstances, unless they have been nominated
literature seems to overestimate the extent to in advance by the patients themselves or if the
which patients understood trial details in the patient is over 16 years. Others, including the
above studies, and a qualitative study by Snow- MRC,62 the Royal College of Physicians,63 and
the FDA,64 have argued that, since routine prac-
don and colleagues has shown that many people
tice must proceed on the basis of ‘assumed’
completely failed to grasp that they had been
preferences, i.e. those of the ‘average’ patient, or
assigned to a control group in a study of neo- ‘implied consent’, randomisation may still be
natal extra-corporeal oxygen therapy.52 A further ethical. If a clinician thinks, that on this point at
qualitative study concluded that the concepts of least, his or her own views are ‘typical’, then all
equipoise and of randomisation proved most they have to do is follow the gold standard, and
difficult for patients to understand.53 This sug- treat the patient as they would themselves wish
gests that the amount and quality of com- to be treated. We believe, however, that it is
munication is often inadequate. On the whole necessary to add one further caveat – under
patients, however, seemed satisfied with their these circumstances, no altruism whatever
lot,27,54,55 although this could just mean that they should be assumed – to act otherwise, on such
do not appreciate what they are missing. vulnerable patients (children, people with severe
The findings from studies comparing different mental illness, or the unconscious) would be
methods of obtaining consent on patients’ tantamount to exploiting the vulnerable, and this
understanding and on consent rate are not easy would find little support in the moral literature.
to interpret, but the following results stand A somewhat different situation arises when
out.26 a fully conscious and technically ‘competent’
Ethical Issues in Design and Conduct of RCTs 17
patient seeks to delegate the decision to the by whether or not they agree to participate in a
clinician. This is arguably a valid expression clinical study. We conclude that the relevance of
of autonomy.14 Here, it seems appropriate to any trial effect should not be oversold and that,
respect the patient’s wish not to have to decide while any such effect is broadly reassuring at the
and to accept delegated decision-making respons- societal level, it cannot be used as an induce-
ibility – again assuming that the patient has a ment for an individual patient to participate.
‘typical’ value system. Again, how we deal with
trial situations mirrors how we deal with non-
trial practice. Restrictions on Trial Treatments
So far, we have considered the situation which
Could the Offer of Trial Entry be arises when the treatments in question are nor-
Distressing? mally freely available – the patient may choose
treatment A, treatment B, or randomisation
Trials, the above argument suggests, are ethical between these alternatives. However, the situ-
when a patient has been fully consulted, such ation can arise in which a type of treatment – we
that his or her values are brought into play will call it treatment A – is not available out-
alongside prior probabilities and equipoise has side the trial, say, because a government has
resulted. However, all of this assumes that the restricted its use. Under these circumstances, we
offer of trial entry is itself neutral, that expected argue that the requirement of equipoise between
utilities are not affected (either way) simply by A and B need not apply, it is necessary only that
making the offer to participate. For example, if B is not superior to A. If the patient is indeed
patients’ experience of their care were system- equipoised between treatments A and B, then all
atically and significantly degraded by the offer the above arguments apply. If, however, the
of trial entry, then we would have to think patient prefers A, then the likelihood of receiv-
seriously about continuing advocacy of trials. ing the preferred treatment is maximised by
Comparative studies of different methods of participating in the trial.
obtaining consent have shown that anxiety is Restrictions on access are part of a broader
increased by enhanced levels of information,58 issue of distributive justice, the allocation of
albeit for a short period only.33 There also seems benefits and disbenefits across society. Distrib-
to be an interaction between good overall under- utive justice implies that discrimination without
standing at consent and low anxiety.51 good reason is wrong and, in the medical con-
text, this usually means that individuals who are
in need of treatments should be given an equal
Are Trials on Average a Better Bet than chance of getting them. Doctors should at least
Routine Practice? offer trial entry in such cases. There could still
be an issue, however, over what to do once the
As well as the offer of trial entry, we must trial has come to term. Should randomisation
consider the effect of the whole trial itself on continue in order that some people at least get
participants. Much is often made about putative the preferred treatment? The issue is particularly
beneficial effects of clinical trials on physical pertinent in Third World trials where the local
well-being. We have reviewed this topic and economics do not always permit widespread use
found that, although the evidence is weak, there of the treatments used in the trials after the trial
seems to be a tendency for patients entered in has finished.17 Should all participants be offered
clinical trials to have better clinical outcomes, the beneficial treatment after the trial? And
even when the trial itself produced a null who should pay for the continued use of such
result.65–79 However, further analysis of this treatments?
data, suggests that this effect occurs when there Some authors have questioned whether policy-
is a pre-existing effective treatment whose effec- makers should contrive a situation, where treat-
tiveness could be enhanced in a trial by the use ments are restricted to trial patients, in the first
of protocol or by a Hawthorne effect.1 It there- place. Is it ethical to mount such a trial? Space
fore seems that any trial effect, if true (and precludes detailed consideration of this issue,
of course these results are observational, not but here we simply observe that individual
experimental), may be contingent upon careful clinicians have very different obligations from
use of existing treatments. Since we should policy-makers. The latter are quite properly con-
aspire to apply this standard to all patients, it cerned with the common welfare, while the
seems egregious to argue for more liberal entry former, it seems reasonable to argue, have a
into clinical trials on the basis of a ‘trial effect’. primary duty to maximise the welfare of indi-
Moreover, it is a specific tenet of the Helsinki vidual patients (exceptions arise where patients’
accord, that a patient’s care should be unaffected unfettered actions may seriously threaten others,
18 The Advanced Handbook of Methods in Evidence Based Healthcare
as occurs when people with poorly controlled understandably be tempted by a restricted treat-
epilepsy continue to drive). If we start from ment, but those who land up in the control group
the (reasonable) assumption that resources are will be used solely for the sake of science by
limited, and that policy-makers are charged with drawing the very short straw and receiving sub-
getting the best value for the public purse, and if standard care. However, if, after randomisation,
we make the further assumption that the chances controls were to guess correctly or be told that
of doing so are maximised by obtaining good they were in fact getting worse care than they
trials evidence, then a compelling argument would be getting outside the trial, they could
could be made for a policy whereby new treat- simply withdraw and select a better treatment
ments are restricted to patients in clinical trials, outside the trial. If patients were to refuse trial
even though individual patients acting only in entry, they would just be offered routine care,
self interest might prefer the new method. More- but they would have no chance of getting their
over, if our earlier argument is accepted (that the preferred treatment. For example, some placebo-
requirement for a clinical trial is more exacting controlled trials, which do not include an exist-
than had hitherto been assumed since it must be ing standard treatment, yet are compatible with
based on equipoise rather than just uncertainty), it, are contrived to get patients to forgo the
then it would seem that the argument to restrict standard in exchange for a chance of getting the
access is strengthened – tying the hands of preferred treatment which they can only obtain
clinicians and then patients seems to provide the by accepting the chance of receiving a placebo.84
best way to enhance the common good, while Placebo-controlled trials in the face of standard
enshrining the concept of the sanctity of the treatments which could be, but are not, offered
doctor/patient relationship. as part of the trial would be using individuals
It has been argued that simply restricting solely for the sake of estimating ‘absolute’
preferred treatments to trials does not neces- effectiveness of the treatment under evaluation,
sarily impinge on a patient’s autonomy, since it as opposed to its effectiveness relative to the
merely limits the options open to the patient and current standard.
is not coercive in the sense that it forces patients
to enter trials.80 Other ethicists have argued, APPLYING THE CONSTRAINT OF EQUIPOISE TO
however, that it is unethical to restrict new
treatments to trials when those treatments are RCT DESIGN
potentially life-saving and when there is no
effective standard outside. They argue that, in A clear understanding of equipoise – by which
such cases, individuals are so desperate that they we mean patient equipoise – has many practical
are, to all intents and purposes, coerced into and methodological corollaries.
participating.81,82 The ECMO and AIDS trials
are just two examples. An extreme view is that Underpowered Trials
doctors should not be coerced into accepting the
unethical policy of restricting access to new It has been argued elsewhere that the widely
treatments. What do the patients think? Inter- promulgated notion that ‘underpowered’ clinical
estingly, a large proportion (79%) of responding trials are unethical in terms of their use of
AIDS patients across two studies,41,83 and parents patients, is not as authoritative as it sounds; as
of very ill children in a further one52 thought that long as patients are equipoised, they are not
such medications should be available to suffer- losing out in prospect by taking part, and there-
ers outside the trial, notwithstanding the scien- fore the ethics of ‘underpowered’ studies turns
tific need for proper evaluation and notwith- on the opportunity costs – that is, value for
standing the resource pressures that general money – not on obligations to individuals.85 It
availability would pose. However, it is not easy follows that if the amount of transferable
to interpret these data, for respondents could resource used in carrying out such a study is
have been making a general point about priori- minimal, then such trials are not unethical,
tising health care or rationing rather than the although their scientific value must not be over-
specific point about restricting some new treat- sold so that any altruism is in proportion to their
ments to RCTs. true scientific value. In assessing the latter, it
Sometimes, however, there does seem to be a should be borne in mind that some unbiased
real risk of exploitation by restricting strongly information is better than none,86 and that most
preferred treatments to trials when the control trials are replicated at some time or another, so
arm is expected to be worse than routine or that individually underpowered studies – pro-
standard care, but the net expected utility of the vided they are carried out to a high-quality stand-
trial is greater than that of routine care. This ard – can contribute to a precise overall estimate
could be construed as bribery. Patients may of effects.87–89
Ethical Issues in Design and Conduct of RCTs 19
Sometimes, there is a scientific desire to since the controls get exactly the same form of
preserve blinding by using active placebos care as people elsewhere who are not involved
(placebos, such as sham operations, which in the trial, there is no need to obtain their
mimic the adverse effects of a new interven- consent specifically for participating in research,
tion).102 However, their use is almost always but consent is sought for any routine treatments
unethical and may rely solely on altruism for on offer.110 Others argue that to carry out ran-
recruitment when the intervention is an other- domisation and follow-up without a person’s con-
wise freely available treatment.103 Arguably, this sent amounts to a form of deception.15 There
is implausible when the active placebo is highly seems to be no way to resolve fully these
invasive. There may be less scientific need to arguments, but we make two points here. Firstly,
use active placebos anyway when minor adverse in the case of behavioural interventions, we
effects do not give the intervention away. suggest that ‘assent’ to take part in the study
may be sought from controls, whereby they
understand that they are the controls for an
Cluster RCTs and Behavioural educational intervention, and consent to follow-
Interventions up, but to remain, at least temporarily, ignorant
of further details. Secondly, that surveys have
Cluster trials are of two sorts – those where the shown that some individuals find the Zelen
treatment is indivisible at the individual level method deeply unpopular, when used as a
(such as advertising campaigns on local radio), method to avoid distressing controls in trials of
and those where the intervention in question potentially effective interventions.111 It is not
may be accepted or refused by individuals. In clear, however, from this study, whether these
the first type of study, we cannot consider indi- people are objecting to being kept in the dark
viduals in isolation for the purposes of ‘con- per se, or whether restricted access to a promis-
sent’, and one person’s choice will impinge on ing new treatment is the real problem. Whether
another’s. We have argued elsewhere that to the intervention is individual or not, we have
preclude all such studies on this basis would be argued that individual(s) who ‘deliver’ clusters
unjustifiable; this would be tantamount to saying into a trial should sign a consent form.104 Such
that alterations in service delivery were accept- individuals or ‘guardians’ could look after the
able, as long as they were uncontrolled and best interests of their cluster.
hence not part of a formal prospective evalu-
ation.104 Therefore, we argue that the same con-
ditions should apply as do for conventional trials CONCLUSION
– the more risky or controversial the inter-
vention, the greater the safeguards that should
be required, so that a relatively uncontroversial We have discussed how a patient’s best interests
measure, such as an advertising campaign can be reconciled with the societal need to
advocating a low-fat diet, would require little advance medical knowledge through RCTs. We
consultation, while a more intrusive measure have argued that Freedman is wrong in thinking
(such as fluoridating the water supply) might that collective uncertainty is a sufficient condi-
require extensive local consultation or even a tion for mounting a trial and then recruiting
referendum. patients, and we have clarified what best inter-
Cluster trials at the individual level are ests mean for patients in terms of what we know
usually done in order to promote beneficial herd about a treatment’s effectiveness and of what
effects or to avoid ‘contamination’, where the values might determine best treatment. Compe-
effect of the intervention on one person affects tent patients, provided that they understand all
others in a locality (such as a form of educa- material information, can use their own values
tional therapy to prevent substance misuse). The to decide whether any particular trial would be a
problem with these latter types of intervention good bet. In the case of incompetent patients,
arises from the fact that it is very hard to average values or advance directives, perhaps
separate consent for the intervention from the nominating a proxy or foreseeing a specific
intervention itself, and this poses a particular situation, must be used to protect them. When
difficulty for controls. Zelen randomisation has trial treatments are freely available, patients
been advocated in these circumstances, whereby must be in a state of personal equipoise.
consent is obtained only from the intervention This requirement has a number of implica-
patients, but not from controls.105 This is highly tions for RCT practice and design.
controversial, and has been debated extensively 1 Underpowered trials are not necessarily
elsewhere in the context of individual trials.106–109 unethical in their ‘use’ of patients.
Proponents of this method argue that consent 2 Interim analyses are desirable, although the
would not be sought in non-trial practice, and current practice of keeping preliminary data
Ethical Issues in Design and Conduct of RCTs 21
secret is suspect, especially when trial treat- 11. Johnson N, Lilford RJ, Brazier W. At what level
ments are widely available. of collective equipoise does a clinical trial
3 Adaptive designs do not solve the problem become ethical? J. Med. Ethics 1991; 17: 30–4.
of reducing expected utilities for participants 12. Jackson J. Telling the truth. J. Med. Ethics 1991;
when trial treatments are widely available. 17: 5–9.
4 Placebo-controlled trials generally are not 13. Mill John Stuart. Utilitarianism. London: Long-
usually appropriate when standard treatments mans, Green, Reader and Dyer, 1874.
exist, and those using ‘active’ placebos with 14. Dworkin G. The theory and practice of auton-
adverse effects should be used with caution omy. Cambridge University Press, 1988.
if at all. 15. Ashcroft RE, Chadwick DW, Clark SRL,
5 Cluster RCTs can be categorised into Edwards RHT, Frith L, Hutton JL. Implications
cluster–cluster and individual–cluster trials, of socio-cultural contexts for the ethics of
depending on the type of intervention under clinical trials. Health Technology Assessment
study. Guardians should consent to random- 1997; 1(9).
isation in both types of trial. The best inter- 16. CIOMS and WHO. International guidelines for
ests of the cluster is of paramount concern in biomedical research involving human subjects.
the former type and individuals can consent 1993.
to interventions in the latter type of trial. 17. Lockwood M, Anscombe GEM. Sins of omis-
6 Many behavioural interventions may con- sion? The non-treatment of controls in clinical
taminate controls, either directly through trials. Aristotelian Society 1983; 57: 222.
fully informed consent or indirectly through 18. Alderson P, Madden M, Oakley A, Wilkins R.
close contact of individuals. Cluster trials Women’s knowledge and experience of breast
may be used, but it may be necessary to cancer treatment and research. Social Science
withhold information about the intervention Research Unit, 1994.
from controls. 19. Chard JA, Lilford RJ. The use of equipoise in
clinical trials. Soc. Sci. Med. 1998; 47: 891–8.
20. Nagel T. The possibility of altruism. Princetown:
Princetown University Press, 1970.
REFERENCES 21. Edwards SJL, Lilford RJ, Hewison J. The ethics
of RCTs from the perspectives of patients, the
public and health care professionals. Br. Med. J.
1. Edwards SJL, Lilford RJ, Braunholtz DA, 1998; 317: 1209–12.
Thornton J, Jackson J, Hewison J. Ethical issues 22. Millon-Underwood S. Determinants of partici-
in the design and conduct of Randomised Clinical pation in state-of-the-art cancer prevention, early
Trials. HTA, 1998. detection, screening, and treatment trials. Cancer
2. Gillon R. Recruitment for clinical trials: the Nursing 1993; 16: 25–33.
need for public professional cooperation. J. Med. 23. Cassileth BR, Lusk EJ, Miller DS, Hurtwitz S.
Ethics 1994; 20: 3–4. Attitudes toward clinical trials among patients
3. Fox TF. The ethics of clinical trials. Medico- and the public. JAMA 1982; 248: 968–70.
legal J. 1960; 28: 132–41. 24. Marsden J, Sacks NPD, Baum M, Whitehead
4. Pringle M, Churchill R. Randomised controlled MI, Crook D. A pilot study investigating the
trials in general practice [editorial] [see com- feasibility of conducting a randomised trial of
ments]. Br. Med. J. 1995; 311: 1382–3. HRT use in women with a history of breast
5. Schafer A. The ethics of the randomized clinical cancer – preliminary report of findings. 1999;
trial. N. Eng. J. Med. 1982; 307: 719–24. Unpublished report.
6. Vere DW. Controlled clinical trials – the current 25. Taylor KM. Informed consent: the physicians’
ethical debate. Proc. R. Soc. Med. 1981; 74: perspective. Soc. Sci. Med. 1987; 24: 135–43.
85–7. 26. Edwards SJL, Lilford RJ, Thornton J, Hewison
7. Freedman B. Equipoise and ethics of clinical J. Informed consent for clinical trials: in search
research. N. Eng. J. Med. 1987; 317: 141–5. of the best method. Social Science and Medicine
8. Gifford F. The conflict between randomized 1998; 47(11): 1825–40.
clinical trials and the therapeutic obligation. J. 27. Hassar M, Weintraub M. ‘Uninformed’ consent
Med. Phil. 1986; 11: 347–66. and the wealthy volunteer: an analysis of patient
9. Lilford RJ, Jackson J. Equipoise and the ethics volunteers in a clinical trial of a new anti-
of randomization. J. R. Soc. Med. 1995; 88: flammatory drug. Clin. Pharmacol. Ther. 1976;
552–9. 20: 379–86.
10. French S. Decision theory: an introduction to 28. Marini SB. An evaluation of informed consent
the mathematics of rationality. New York: John with volunteer prisoner subjects. Yale Journal of
Wiley and Sons, 1986. Biology and Medicine 1976; 427: 434.
22 The Advanced Handbook of Methods in Evidence Based Healthcare
29. Bergler JH, Pennington AC, Metcalfe M, Freis 168 patients in a controlled clinical trial. J. Clin.
ED. Informed consent: how much does the Res. Drug Dev. 1994; 8: 237–48.
patient understand? Clin. Pharmacol. Ther. 44. Daugherty C, Ratain, MJ, Grochowski E, et al.
1980; 27: 425–40. Perceptions of cancer patients and their physi-
30. Howard JM, Demets D, BHAT Research Group. cians involved in phase I trials. J. Clin. Oncol.
How informed is informed consent? The BHAT 1995; 13: 1062–72.
experience. Controlled Clinical Trials 1981; 2: 45. DeLuca SA, Korcuska LA, Oberstar BH, et al.
287–304. Are we promoting true informed consent in
31. Penman DT, Demets D, BHAT Research Group. cardiovascular clinical trials? J. Cardiovasc.
Informed consent for investigational chemo- Nursing 1995; 9: 54–61.
therapy: Patients’ and physicians’ perceptions. J. 46. Harth SC, Thong YH. Parental perceptions and
Clin. Oncol. 1984; 2: 849–55. attitudes about informed consent in clinical
32. Rodenhuis S, van den Heuvel WJ, Annyas AA, research involving children. Soc. Sci. Med.
et al. Patient motivation and informed consent in 1995; 40: 1573–7.
a phase I study of anticancer agent. Euro. J. 47. Harrison K, Vlahov D, Jones K, Charron K,
Cancer Clin. Oncol. 1984; 20: 457–62. Clements ML. Medical eligibility, comprehen-
33. Simes RJ, Tattersall MH, Coates AS, Raghavan sion of the consent process, and retention of
D, Solomon HJ, Smartt H. Randomised com- injection drug users recruited for an HIV vac-
parison of procedures for obtaining informed cine trial. J. Acquired Immune Deficiency Syndr.
consent in clinical trials of treatment for cancer. Hum. Retrovirol. 1995; 10: 386–90.
Br. Med. J. (Clin. Res. Ed.) 1986; 293: 1065–8. 48. Olver I, Buchanan L, Laidlaw C, Poulton G. The
34. DCCT Research Group. Implementation of a adequacy of consent forms for informing
multicomponent process to obtain informed con- patients entering oncological clinical trials. Ann.
sent in a diabetes control and complications trial. Oncol. 1995; 6: 867–70.
Controlled Clinical Trials 1998; 10: 96. 49. Postlethwaite RJ, Reynolds JM, Wood AJ,
35. Rennie D, Yank V. Disclosure to the reader of Evans JH, Lewis MA. Recruiting patients to
institutional review board approval and informed
clinical trials: lessons from studies of growth
consent. JAMA 1997; 277: 922–3.
hormone treatment in renal failure. Arch. Dis.
36. Davis S, Seay J, Stone J, et al. Does knowing
Child 1995; 73: 30–5.
about clinical trials increase participation?
50. Tomamichel M, Sessa C, Herzig S, de Jong J,
Unpublished report.
Pagani O, Willems Y, Cavalli F. Informed con-
37. Lynoe N, Sandlund M, Dalqvist G, Jacobsson L.
sent for phase I studies: evaluation of quantity
Informed consent: study of the quality of infor-
and quality of information provided to patients.
mation given to participants in a clinical trial. Br.
Ann. Oncol. 1995; 6: 363–9.
Med. J. 1991; 303: 613.
38. Oddens BJ, Oddens BJ, Algra A, van Gijn J. 51. Aaronson NK, Visserpol E, Leenhouts GHMW,
How much information is retained by partici- Muller MV, Vanderschot ACM, Vandam
pants in clinical trials? Ned. Tijdschr. Geneeskd. FSAM, Keus RB, Koning CCE, Huinink WWT.
1992; 136: 2272–6. Telephone-based nursing and intervention im-
39. Jensen AB, Madsen B, Anderson P, Rose C. proves the effectiveness of the informed consent
Information for cancer patients entering a clin- process in cancer clinical trials. J. Clin. Oncol.
ical trial: evaluation of an information strategy. 1996; 14: 984–96.
Eur. J. Cancer 1999; 29: 2235–8. 52. Snowdon C, Garcia J, Elbourne D. Making
40. Gallet M, Aupetit JF, Servan E, Lestaevel M, sense of randomisation: responses of parents of
Lopez M, Chassoux G. Informed consent, what critically ill babies to random allocation of treat-
information is left after a therapeutic trial? Arch. ment in a clinical trial. Soc. Sci. Med. 1997; 45:
Mal. Coeur Vaiss. 1994; 87: 39–45. 1337–55.
41. Tindall B, Forde S, Ross MW, et al. Effects of 53. Featherstone K, Donovan JL. Random allocation
two formats of informed consent on knowledge or allocation at random? Patients’ perspectives
amongst persons with advanced HIV disease in of participation in a randomised controlled
a clinical trial of didanosine. Patient Education trial [see comments]. Br. Med. J. 1998; 317:
and Counselling 1994; 24: 266. 1177–80.
42. Maslin A. A survey of the opinions on 54. Wilcox M, Schroer S. The perspective of
‘informed consent’ of women currently involved patients with vascular disease on participation in
in clinical trials within a breast unit. Eur. J. clinical trials. J. Vasc. Nurs. 1994; 12: 112–16.
Cancer Care 1994; 3: 153–62. 55. Bevan EG, Chee LC, McGhee SM, McInnes
43. Miller C, Searight HR, Grable D, Schwartz R, GT. Patients’ attitudes to participation in clinical
Sowell C, Barbarash RA. Comprehension and trials. Br. J. Clin. Pharmacol. 1993; 35: 204–7.
recall of the informational content of the 56. Davis S., Seay J., Stone J. et al. Does knowing
informed consent document: An evaluation of about clinical trials increase participation?
Ethical Issues in Design and Conduct of RCTs 23
[Abstract]. Proc. Am. Soc. Clin. Oncol. 1990; 9: Hammar SP. Participants in a prospective, ran-
A232 (Abstract). domized clinical trial for resected non-small cell
57. White DR, Muss HB, Michielutte R, et al. lung cancer have improved survival compared
Informed consent: patient information forms in with nonparticipants in such trials. Cancer 1985;
chemotherapy trials. Am. J. Clin. Oncol. 1984; 7: 56: 1710–18.
183–90. 71. Jha P, Deboer D, Sykora K, Naylor CD. Char-
58. Epstein LC, Lasagna L. Obtaining informed acteristics and mortality outcomes of throm-
consent. Form or substance. Arch. Intern. Med. bolysis trial participants and non-participants: a
1969; 123: 682–8. population-based comparison. J. Am. Coll.
59. Dal-Re R. [Clinical trial of drugs: a study of the Cardiol. 1996; 27: 1335–42.
influence of information on adverse reactions on 72. Karjalainen S, Palva I. Do treatment protocols
obtaining of informed consent] Investigacion improve end results? A study of survival of
clinica con farmacos: estudio de la influencia de patients with multiple myeloma in Finland [see
la informacion sobre reacciones adversas en la comments]. Br. Med. J. 1989; 299: 1069–72.
obtencion del consentimiento informado. Med. 73. Lennox EL, Stiller CA, Jones PH, Wilson LM.
Clin. (Barc.) 1991; 96: 566–9. Nephroblastoma: treatment during 1970–3 and
60. Blondeau D, Valois P, Keyserlingk EW, Hebert the effect on survival of inclusion in the first
M, Lavoie M. Comparison of patients’ and MRC trial. Br. Med. J. 1979; 2: 567–9.
health care professionals’ attitudes towards 74. Duration of survival of children with acute
advance directives. J. Med. Ethics 1998; 24: leukaemia. Report to the Medical Research
328–35. Council from the Committee on Leukaemia and
61. Kapp MB. Proxy decision-making in Alzheimer the Working Party on Leukaemia in Childhood.
disease research. Alzheimer Disease Association Br. Med. J. 1971; 4: 7–9.
1994; 8: 28–37. 75. Schmoor C, Olschewski M, Schumacher M.
62. Working Party on Research on the Mentally Randomised and non-randomised patients in
Incapacitated. The ethical conduct of research clinical trials: experiences with comprehensive
on the mentally incapacitated. London: MRC, cohort studies. Stat. Med. 1996; 15: 263–71.
1991. 76. Stiller CA, Draper GJ. Treatment centre size,
63. The Royal College of Physicians, London. A entry to trials, and survival in acute lympho-
Report to the Royal College of Physicians. blastic leukemia. Arch. Dis. Child. 1989; 64:
Guidelines on the practice of Ethics Committees 657–61.
in medical research involving human subjects. 77. Stiller CA, Eatock EM. Survival from acute
2nd ed., 1990. non-lymphocytic leukaemia, 1971–88: a popula-
64. Hastings Center Report. Symposium on tion based study. Arch. Dis. Child 1994; 70:
Informed Consent. In case of emergency: no 219–23.
need for consent. Hastings Center Report 1997; 78. Ward LC, Fielding JW, Dunn JA, Kelly KA.
27(1): 7–12. The selection of cases for randomised trials: a
65. Mahon J, Laupacis A, Donner A, Wood T. registry survey of concurrent trial and non-trial
Randomised study of n of 1 trials versus stand- patients. The British Stomach Cancer Group. Br.
ard practice [see comments] [published erratum J. Cancer 1992; 66: 943–50.
appears in Br. Med. J. 1996 Jun 1; 312(7043): 79. Williford WO, Krol WF, Buzby GP. Com-
1392]. Br. Med. J. 1996; 312: 1069–74. parison of eligible randomized patients with two
66. Reiser J, Warner JO. The value of participating groups of ineligible patients: can the results of
in an asthma trial. Lancet 1985; 1: 206–7. the VA Total Parenteral Nutrition clinical trial
67. Antman K. A comparison of randomized vs be generalized? J. Clin. Epidemiol. 1993; 46:
concurrent controls. Proc. Am. Assoc. Cancer 1025–34.
Res. 1983; 24: 146. 80. Logue G, Wear S. A desperate solution: individ-
68. Bertelsen K. Protocol allocation and exclusion ual autonomy and the double-blind controlled
in two Danish randomised trials in ovarian can- experiment. J. Med. Philos. 1995; 20: 57–64.
cer. Br. J. Cancer 1991; 64: 1172–6. 81. Minogue BP, Palmer-Fernandez G, Udell L,
69. Anonymous Coronary artery surgery study Waller BN. Individual autonomy and the
(CASS): a randomized trial of coronary artery double-blind controlled experiment: the case of
bypass surgery. Comparability of entry charac- desperate volunteers. J. Med. Philos. 1995; 20:
teristics and survival in randomized patients and 43–55.
nonrandomized patients meeting randomization 82. Truog RD. Randomized control trials: lessons
criteria. J. Am. Coll. Cardiol. 1984; 3: 114–28. from ECMO. Clin. Res. 1992; 40: 519–27.
70. Davis S, Wright PW, Schulman SF, Hill LD, 83. Twomey JG. Investigating pediatric HIV
Pinkham RD, Johnson LP, Jones TW, Kellogg research ethics in the field. West. J. Nurs. Res.
HB, Radke HM, Sikkema WW, Jolly PC, 1994; 16: 413.
24 The Advanced Handbook of Methods in Evidence Based Healthcare
84. Rothman KJ, Michels KB. The continuing 99. Prescott RJ. Feedback of data to participants
unethical use of placebo controls. N. Eng. J. during clinical trials. In: Tagnon HJ, Staquet MJ.
Med. 1994; 331: 394–8. Anonymous Controversies in cancer: Design of
85. Edwards SJ, Lilford RJ, Braunholtz D, Jackson trials and Treatment. New York, 1979; 55–61.
J. Why ‘underpowered’ trials are not necessarily 100. Thornton JG, Lilford RJ. Preterm breech babies
unethical [see comments]. Lancet 1997; 350: and randomised trials of rare conditions [com-
804–7. ment]. [Review] Br. J. Obstet. Gynaecol. 1996;
86. Lilford RJ, Braunholtz D. The statistical basis of 103: 611–13.
public policy: a paradigm shift is overdue [see 101. Simon R. Adaptive treatment assignment meth-
comments]. Br. Med. J. 1996; 313: 603–7. ods and clinical trials. Biometrics 1977; 33:
87. Fayers PM, Machin D. Sample size: how many 743–9.
patients are necessary? [editorial]. Br. J. Cancer 102. Shapiro AK, Shapiro E. The Powerful Placebo.
1995; 72: 1–9. The Johns Hopkins University Press, 1997.
88. Barnard GA. Must clinical trials be large? The 103. Edwards SJL, Braunholtz DA, Stevens AJ,
interpretation of p-values and the combination of Lilford RJ. Are some placebo-controlled trials
test results. Stat. Med. 1990; 9: 601–14. unethical? Unpublished report.
89. Laupacis A. Research by collaboration [see 104. Edwards SJL, Braunholtz DA, Lilford RJ,
comments]. Lancet 1995; 345: 938. Stevens AJ. Ethical issues in the design and
90. Lilford RJ, Johnson N. The alpha and beta errors
conduct of cluster RCTs. Br. Med. J. 1999; 318:
in randomized trials [letter]. N. Eng. J. Med.
1407–9.
1990; 322: 780–1.
105. Donner A. Some aspects of the design and
91. Armitage P. Interim analysis in clinical trials.
analysis of cluster randomized trials. Appl.
Stat. Med. 1991; 10: 925–35.
Statist. 1998; 47: 95–113.
92. Ashby D, Machin D. Stopping rules, interim
106. Horwitz RI, Feinstein AR. Advantages and
analyses and data monitoring committees [edi-
torial]. Br. J. Cancer 1993; 68: 1047–50. drawbacks of the Zelen design for randomized
93. Berry DA. Interim analyses in clinical research. clinical trials. J. Clin. Pharmacol. 1980; 20:
Cancer Invest. 1987; 5: 469–77. 425–7.
94. Berry DA. Interim analyses in clinical trials: 107. Gore SM. Zelen randomisation [letter]. Lancet
classical vs. Bayesian approaches. Stat. Med. 1984; 2: 226–7.
1985; 4: 521–6. 108. Taube A. [n or more in clinical trials? Zelen’s
95. Freedman LS, Spiegelhalter DJ. Comparison of models for randomization] n eller flera vid
Bayesian with group sequential methods for kliniska provningar? Zelens modeller for ran-
monitoring clinical trials [see comments]. domisering. Lakartidningen. 1994; 91: 977–82.
Control. Clin. Trials 1989; 10: 357–67. 109. Torgerson DJ, Roland M. What is Zelen’s
96. Pocock SJ. The role of external evidence in data design? Br. Med. J. 1998; 316: 606.
monitoring of a clinical trial. [Review] Stat. 110. House A, Knapp P. Informed consent. Trials
Med. 1996; 15: 1285–93. that use Zelen’s procedure should be acceptable
97. Clayton D. Ethically optimized designs. Br. J. [letter; comment]. Br. Med. J. 1997; 315: 251.
Clin. Pharmacol. 1982; 13: 469–80. 111. Snowdon C, Elbourne D, Garcia J. Zelen ran-
98. Royall RM. Ethics and statistics in randomized domization: attitudes of parents participating in
clinical trials, with discussion. Stat. Sci. 1991; 6: a neonatal clinical trial. Control. Clin. Trials
52–88. 1999; 20: 149–71.
2
Ethics of Clinical Trials:
Social, Cultural and Economic Factors
RICHARD E. ASHCROFT
human patients. It is now generally accepted that present discrimination against the group in
experimentation on patients involves ethical medical and other contexts.18,19 In other cases
issues. Even where there is an ethical consensus there is concern that some injustice is brought
about what these issues are, and what the par- about by features of trial methodology.20,21 Fur-
ticular ethical requirements are in respect of the ther, there is concern that some groups are over-
proposed experiment, this consensus needs to be or under-represented due to economic factors
underwritten by sound argument. The principal influencing the delivery of healthcare inside or
requirement of any experiment is that it should outside clinical trials.22 It is difficult to dis-
be scientifically useful. This requirement is a entangle the cultural and the economic here,
necessary condition, although it is not sufficient because often the poor happen to be members of
on its own. Most health researchers agree that groups that are discriminated against (or feel
the most scientifically rigorous method for treat- themselves to be so). These considerations once
ment evaluation is the randomised, controlled again lead us to consider whether there could be
clinical trial (RCT).7–10 This is not accepted by methods to recruit more fairly, for instance,
all writers, and there is evidence to suggest that without compromising the scientific integrity of
randomisation and control are concepts not well the trial.
understood by many patients, and that many
patients are unhappy with being assigned ran-
domly to a treatment, or with being assigned to
the control group.11–15 As a consequence, an
METHODS USED IN THE REVIEW
important body of opinion has formed which
argues against particular elements of the RCT The report was commissioned as a systematic
methodology on ethical grounds, and which has review of the literature on ethics, clinical trials,
proposed a number of modifications, or altern- and relevant socio-cultural factors. It was not
atives, to the RCT. intended that the study should undertake any
Our hypothesis was that some of the objec- new research beyond the critical philosophical
tions to the RCT on methodological or ethical analysis of the literature surveyed. The aim of a
grounds might have a systematic basis, not systematic review is to determine what work has
solely attributable to misunderstanding of the been done, and with what quality; to indicate the
aims, premises and methods of the RCT, in current state of knowledge for practical applica-
some particular cases or in general. These objec- tion; and to guide future research and practice.
tions, if they are systematic, may reflect the The survey concentrated on the British context,
beliefs (political, philosophical, religious or cul- using American and other developed countries
tural) of particular groups within society. Should for comparison and, where appropriate, amplifi-
this be the case, these systematic beliefs would cation. The literature survey was restricted, for
require Local Research Ethics Committees or reasons of convenience, to literature published
Institutional Review Boards to pay special atten- or abstracted in English.
tion to the interests of such groups in assessing The questions framed in our hypothesis were
RCTs for their ethical adequacy.16 More than partly philosophical and partly empirical. We
this, these objections might be such that altern- needed to know whether there were in fact any
atives to the RCT should be considered more common objections to the RCT methodology (or
sympathetically, as health assessment tech- to particular trials); and we needed to assemble
nologies more widely acceptable to the ethical the interpretative and philosophical resources to
beliefs of all groups in society. But this would make sense of them. We used a modified sys-
also require showing that alternatives to the tematic reviewing strategy to assemble a data-
RCT did meet the first requirement of human base of articles and book chapters. Systematic
experimentation, which is scientific adequacy searches on Medline, Psychlit and Sociofile CD
and efficiency. Finally, then, objections to the ROM databases, together with hand-searches on
RCT might necessitate a negotiated compromise the most important relevant medical, bioethics
between the need for validated health technolo- and philosophy journals, were used to assemble
gies and the requirements of objecting groups’ as wide a range of empirical and analytic material
ethical principles. as possible and useful. Objections were sought
In the course of the research, we quickly of both direct and indirect kinds. Direct objec-
became aware that there was a second kind of tions comprise explicit statements by patients
problem for ethical research. Groups’ objections and healthcare workers of worries, objections,
to RCTs, and low recruiting of members of and problems to do with recruitment to par-
some groups were sometimes due to issues of ticular trials, compliance with protocols, and
actual or perceived injustice in trial recruitment methodology of trials. Indirect objections were
and process.17 In some cases the perception of limited proactive attempts on the part of the
injustice was due to historical factors: past and project team to extrapolate and infer objections
Ethics of Clinical Trials: Social Factors 27
that particular groups might make to trials, on linked to particular local social circumstances.
the basis of explicit statements about topics Much American evidence does not apply to the
analogous to features of trial methodology, but UK or European health services. However, this
not directly connected to it. For instance, many paints too black a picture. In our study, where
religious groups object on principle to any form we were interested in the quality of the argu-
of gambling, whence it might be inferred that ments used as much as in the evidence used in
they should also object to random assignment. their support (if any), we were able to use
The absence of any objection along these lines evidence from a variety of sources as illustra-
as yet, may perhaps indicate only that this is a tions of more general rational arguments.
pitfall-in-waiting. A wide remit was adopted, on Conversely, we would argue that much evi-
the grounds that careful foresight would benefit dence in ethics research can only be used in this
from moderate speculation. illustrative way. For example, quantitative re-
Searches for analytic material were, of neces- search about methods of obtaining consent may
sity, much broader in scope, because the kinds show that some methods lead to higher recruit-
of objections that might arise could have quite ment rates than others; however it is not clear
broad significance, for example economic, what to conclude from this alone about the
social, political, religious or cultural. Such relative ethical statuses of the methods under
objections might reflect aspects of philosophical trial. Likewise, quantitative methods are good at
arguments which are not normally canvassed in identifying patterns of patients’ preferences, and
contemporary medical ethics, narrowly defined. degrees of preference satisfaction; but satisfac-
Particular care was taken to discover relevant tion of preference and moral rightness are by no
books, as these are often less well represented means always equivalent.
in electronic databases than are abstracted It is, however, true that ethical debate about
articles.1,23 trials does often turn on matters of fact, and
when this is so, it is crucial that the facts be
determined reliably and accurately. Much that
has been published overlooks the role of evi-
EVIDENCE: QUALITY AND DEPTH dence in argument.24 The relative dearth of
useable factual evidence forced us to adopt a
Electronic database searches were conducted theoretical and analytical approach to this topic,
on Medline (1990–September 1996), Psychlit but we also recommend further research to
(1990–September 1996), Sociofile (1974–June improve our knowledge and policy-making in
1996) and Life Sciences (1990–June 1996). this area.1,23
Hand searches were performed on the British
Medical Journal, The Lancet, the New England
Journal of Medicine, the Journal of the American
Medical Association (all 1995–November 1996), KEY ISSUES AND CONTROVERSIES
the Hastings Center Report, the Journal of
Medical Ethics, and Statistics in Medicine (all The following issues can be identified: RCT
complete runs). Some websites of interest were methodology in principle; RCT methods in par-
identified, but they generated little evidence. ticular cases; recruitment and representation;
In the electronic searching, nearly 7000 items consent in principle; and consent in practice.
were identified by use of defined search terms;
however, the relevance of most of these articles
(determined by reading abstracts) was low, and RCT Methodology in Principle
1134 articles were collected. In the final report
just over 300 of these articles were referenced. Much of the debate about clinical trials arises
The searching process uncovered no studies in particular cases, especially when the trial
directly concerned with the attitudes and beliefs involves treatments for high-profile diseases (for
of different cultural groups to clinical trial meth- example cancer and HIV/AIDS), or when the
odology. There is a lot of research, using the trial involves a particular population who are
whole range of experimental methodology, into identified medically by suffering from or sus-
the effectiveness of different methods of obtain- ceptibility to some given condition but who are
ing informed consent, and into methods for also identified socially by some socio-cultural
recruitment generally. Some of this material was marker (for example, gender, ethnic group or
relevant to the issues in hand – particularly those social class). There are usually two levels to
studies which looked at trials with poor recruit- debates framed in this way – a debate about the
ment rates. However, much of the best of this features of RCTs in general (control groups,
research is hard to generalise, either because it randomisation, comparison of therapies, blind-
was linked to particular trials, or because it is ing, risk–benefit calculus and placebo-control),
28 The Advanced Handbook of Methods in Evidence Based Healthcare
and a debate about the appropriateness of apply- The deontological debate is less developed,
ing this methodology to this given group at this save in one particular area – consent – and this
particular time. will be discussed separately. One of our chief
Randomisation is very poorly understood by concerns in this study was to uncover evidence
most people, and generates much controversy. about religious contributions to this deonto-
There is sometimes a demand for people to be logical debate, particularly from the non-
allowed to choose which arm they are assigned Christian religions. The rights-based debate is
to, or – more honestly – a demand that people not well developed, partly because consent gives
should be able to insist on the experimental rights theorists most of what they want, and
therapy.25–28 They can insist on not getting it by partly because few rights theorists in practice
withholding their consent to enter the trial. The use consequentialist reasoning to determine
debate here breaks down into two parts – a what the rights in question involve in the given
debate about the scientific necessity for ran- case. Similarly, duty-based reasoning is very
domisation and controlled comparison, and a familiar (the Declaration of Helsinki specifies a
debate about whether there is something simply
set of duties that bind researchers), but in prac-
wrong and immoral about this methodology,
regardless of its scientific merits. The reasoning tice can be weak, because of ambiguity, excep-
strategy in the first debate is ‘consequentialist’, tions and hard cases for interpreting what the
and in the second is ‘deontological’.29 Conse- principle in question actually requires.33–35
quentialism holds that an act is right when it has
the best consequences. In contrast, deontological
reasoning holds that acts are right in virtue of RCT Methodology in Particular Cases
their conforming to certain fixed principles, and
so an act can be wrong even where it has the In particular situations, it may be argued that
best consequences (for example, lying is just a full RCT is inappropriate or unfair, either
wrong, even where it protects a third party). In because the disease is serious or rare and it
the first debate an RCT is acceptable only if it might be invidious to recruit some but not all of
has the best outcome for the participants and the affected population, or to assign some mem-
society. ‘Best’ is defined with reference to the bers of the population the new therapy, but not
alternatives – not testing at all, obtaining other all. Another argument may be that this trial is
kinds or grades of evidence, allowing patient unjust because it draws on a population which
choice, and so on. However, in the second has already been used as a trial population many
debate, consequences are trumped by considera- more times than average, or is at increased risk
tions about the rights of subjects, the duties of of other discriminatory or risky practices. Argu-
triallists, or theological considerations. For an ments of this type have become familiar from
example, it is conceivable that randomisation AIDS research, research on rare diseases and,
might be thought wrong in the same way that on the world stage, the ‘exporting’ of clinical
gambling is thought wrong by many religions. research to poor or developing countries.36,37
The consequentialist debate is very well Recruitment and randomisation are sometimes
known, and most of the language and techniques seen as unjust because they replicate, or even
we have for discussing RCT ethics are drawn magnify, existing injustices against disadvant-
from this debate, particularly the concept of aged social, cultural or economic groups. Some
proportionality of risk and benefit. This debate cultural groups have succeeded in establishing
has a socio-cultural dimension – perception of the necessity of a group consent prior to individ-
risk and benefit can be linked to social and
ual consent as a mechanism for protecting their
cultural factors; and the definition of risk and
benefit and their chances can be contested. To group’s interests.38–40
some extent this is a technical matter, but One important controversial issue is the use
acceptance of this fact, and of what it indicates of placebo control in trials. It is now generally
in a given trial, rests on a relationship of trust agreed that placebo control is ethical only when
between population, participants and triallists.30 no ‘standard’ (that is, non-experimental) therapy
This relationship is sometimes very strained, for is available, or if there is evidence that this
a variety of reasons. This debate also encom- therapy is actually ineffective or harmful.41–43
passes issues of distributive justice relating to This consensus arises from the conviction that a
recruitment: a fair distribution of the chances of trial is ethical only if it represents a fair choice
access to the trial and of risks and benefit within for the patient. This is the case if equipoise
it, together with wider issues about the chances obtains, but the concept of equipoise is not
of exposure to trials as part of one’s healthcare entirely straightforward.44–46 Also, in extremis,
and about overall access to healthcare of a given patients’ desires can be strong for a treatment
quality.31,32 not yet known to be (in)effective rather than a
Ethics of Clinical Trials: Social Factors 29
treatment known to be ineffective (or placebo). cultural universal is not clear, however, and
It is not at all clear that this desire is irrational or there is evidence that autonomy is less prized in
unethical. some cultures than in the West.55–57 It is argu-
If the intervention being tested is for a rare able that some cultures treat women, for
disease or condition, comparative trials are instance, as incompetent to give full consent, in
somewhat controversial, because of the diffi- the same way that we regard children or the
culty in enrolling ‘sufficiently many’ patients for mentally ill as incompetent. Also, autonomy
a prospective trial. Alternative methods for test- might imply individualism, but it is well known
ing interventions in this case do exist, but they that many cultures regard decision-making as
are subject to some technical controversy. Ethic- hard, or impossible, without consultation with
ally, the main worry is that these difficulties the family or other social grouping.58,59 So there
may lead to therapies not being tested in these is some debate about whether informed consent
areas, or being developed very slowly, thus is not a universal human right at all. In fact, we
harming the interests of the affected patients.47 regard this argument as misleading because
although autonomy may be more or less import-
ant than it is widely believed to be, the role of
Recruitment consent as a protective test remains unchal-
lenged as a principle.
It is sometimes argued that regulation of clinical
trials has unwittingly harmed the interests and
the health of some groups by over-representing Informed Consent in Practice
some groups and under-representing others. For
example, some regulations exclude some or all In practice, however, informed consent con-
women from some trials which would be bene- tinues to be challenged – the number of situ-
ficial to them, on the grounds of possible harms ations where it can be set aside for some reason
to future or actual fetuses.48,49 In the US health- is growing. The old argument that patients can-
care funding context, it is thought by some not consent because they do not know any
authors that the urban poor are unfairly exposed medicine still attracts some attention.60,61 Much
to the risks of trials, because they are enrolled more significant are arguments that patients are
into trials as a cheap form of treatment.50,51 not able to give free consent due to lack of
Some authors argue that not only are there understanding (of language, or of the explana-
injustices in trial management which should be tion given), due to the power relations implicit
addressed through recruitment which is ‘socially in the medical situation, or due to their own
representative’, but that these injustices also bias anxiety or desperation.62,63 The argument that
the studies, making their results ungeneralisable. the requirement to obtain consent harms patients,
For instance, if a drug whose main users will be by creating anxiety, is often raised, as is the
women is not tested on women, then it could argument that patients have a right to know only
be that the results of the study regarding safety as much as they want to know.64,65 In some
or effectiveness may not apply to just the situ- branches of medicine it is claimed that the
ations the drug will be used in most.52,53 The consent requirement holds up valuable research
same argument is applicable to some paediatric or even prevents it (in emergency medicine,
therapies, where recruitment may be difficult or perhaps). Finally, there is the argument that
impossible. consent is at best a pious fiction because it
implies a real choice where in fact the patient
may have – or believe themselves to have – no
Consent in Principle real alternatives.66,67
examined, and we draw out the consequences because it breaches some prima facie duty, right,
for good practice in RCT methodology. or theological imperative. The chief difficulty
which RCT ethics presents in the light of this
question is the apparent conflict between duties
RCT Methodology in Principle to (and rights of) present patients and the needs
and Practice of society. Patients might have a duty to society,
especially if society shares the cost of treat-
There is a literature debating the necessity of ment.73,74 The combined rights of future patients
randomisation and comparison for the testing might deserve consideration.75 Giving one group
and evaluation of therapies. Does randomisation of patients something which no one else can
defeat the influence of known and unknown have, even if blinding means we cannot know
‘confounders’? The importance of this debate is which group this is, is apparently unfair either to
that a trial involving human subjects, and expos- the recipient group because it is in fact less good
ing them to some degree (even if tiny) of risk, than the alternative, or to everyone else because
can only be ethical if it has a good prospect
it is better. This apparent problem collapses
of reaching its goal: knowledge. Unscientific
or unsound experimentation is unethical. The when one considers that we have no reliable
debate about the necessity for randomisation knowledge of harm or benefit unless and until
under a Bayesian philosophy of inference is we have a trial to find out. However, this non-
continuing, but it is worth noting that most problem can be a difficult political issue. It has
practising Bayesian statisticians use randomisa- sometimes been argued that patients should be
tion.68,69 It is safe to say, then, that the weight of able to choose which treatment they receive.
scientific opinion supports randomisation, and Various methods which honour patient prefer-
so that the RCT remains the method of choice. ences have been tried. It is not clear that patient
However what is not clear is the degree to which preferences can be honoured without intro-
alternative methodologies are less reliable, and ducing bias. In addition, whether patients have
the usual rhetoric of ‘gold standard’ methods is a right to specify which treatment they shall
not helpful.70,71 receive is still debated.25–28 Of course they have
A second strand of the controversy concerns a right to refuse treatment, and a right to refuse
the degree to which RCT methodology relies on to enter a trial without incurring any penalty,
a philosophy of inference which makes impos- such as less good treatment than is standard.76
sible early termination of trials when there is This area is difficult, and one strategy for
clear evidence of benefit or harm. This is a responding to patient preferences may be to help
technical issue, beyond the scope of this article, patients to understand why clinical equipoise
but which has clear ethical relevance. In this exists regarding the treatments in the trial, and
area, the controversy is resolved just by con- therefore why there is a trial at all. Another
structing reliable stopping rules.72 strategy may be to help patients define their own
So randomisation as such presents some end points – what they want to get out of the
scientific issues, out of which ethical issues trial.77,78 Either of these techniques may over-
grow. Yet most ethical worries about random- come the initial difference of perception and
isation and controlled comparison arise through interest. Whether it is right to seek to change
failure of understanding of the way the method people’s preferences in the interests of successful
works, and of the initial requirements of cogni- trial management remains a difficult question.
tive equipoise and proportionality of risk and Finally, evidence about the attitudes of par-
benefit within each arm of the trial.
ticular religious or cultural groups to the RCT
We can distinguish consequentialist and
deontological approaches to the ethics of RCTs. methodology as such was considered. We dis-
Consequentialists will usually be satisfied by the covered some evidence that certain groups have
arguments just rehearsed to the effect that the an objection to human experimentation, but that
RCT method is reliable and superior to the the nature of this objection is of lack of trust in
alternatives; once it is granted that drugs and the experimenters, and so if this trust can be
other interventions require testing in the inter- remade, the objection would disappear.19,30 Con-
ests of improving treatment and ensuring safety, trary to our initial hypothesis, we found no
it is clear that one should use the best available evidence that any religion has an objection
method. Consequentialists will then concentrate to randomisation. Some cultural groups have
their attention on particular features of the trial difficulty in accepting equipoise, because they
in question, rather than on the methodology as have difficulty respecting the doctor who claims
such. ignorance on some important medical mat-
The open question which remains is whether ter.55,56 This objection surfaces more visibly in
the superiority of the RCT method is irrelevant the context of consent in particular trials.
Ethics of Clinical Trials: Social Factors 31
more familiar constraints (time, money, and simply written. So the major worry and barrier
complexity), when they define research popula- to true informed consent remains the ‘social
tions. A clear threat to the viability of research distortions’ present in the medical situation.63
is presented by the possibility that a whole A conversation between a patient and a doctor
population may decide that it is being treated may be very unequal, because of the vulnerabil-
as a population of guinea pigs used for others’ ity of the patient due to illness, because of the
benefit. belief that the doctor knows best (or at any rate
more than the patient) about diagnosis and treat-
Consent ment, and – commonly – because there is a
difference in education and social status.62 This
The issues surrounding recruitment to trials bite can make the seeking of consent a complex and
at the point where potential subjects are asked to unreliable process, because of the different
consent to their participation in the trial. There expectations the parties bring to the consent
is an enormous literature on consent. Most of it procedure. This difference in expectation can be
is highly repetitive; we found 340 hits, but could greatly increased when cultural differences in
use only 43 articles as contributing new material. expectation about medicine and the role of the
Much of the literature addressed difficulties in doctor are present. There is good evidence that
obtaining consent; modifications, or exceptions, in many non-Western societies a greater degree
to the consent test; a lot concerns difficulties in of authority or frank paternalism is expected by
obtaining consent; some concerns methods for both patient and doctor. In trials, this can mean
obtaining consent; the psychological mechan- that the requirement of equipoise can be under-
isms and effects of consent; and whether con- stood as an admission of ignorance, and hence
sent is a culturally variable phenomenon are also of incompetence.55,56 In practice, this misunder-
considered. It was found helpful to distinguish standing must be overcome, because consent
among the barriers to consent in practice, before must be insisted on in the interests of protecting
identifying whether there were ‘in principle’ patients.
culturally based objections to the consent test,
because there is much more evidence concern- Consent as a cultural principle
ing the former.
As noted, there is evidence that consent is not
universally accepted as a primary component of
Barriers to consent in practice the medical relationship. But is this a matter of
The principal barrier to consent is – of course – ethical principle, or more a matter of cultural
not asking for it! From time to time studies do convention? In other words, does any culture
come to light where consent was not sought. have explicit objections to consent, either in
Generally this occurs where for some reason it is general, or for particular classes of subjects?
believed that seeking consent would be ‘harmful’ Note that the legal and conceptual identification
to patients, or where seeking consent is prag- of competence and capacity to give consent
matically difficult or impossible.60,82 Occasion- underpins Western ‘commonsense’ exceptions
ally, consent is not sought for research (although to the consent principle of young children, the
it is sought for the intervention), because the mentally ill or disabled, and (sometimes) prison-
difference between research and ordinary prac- ers, and even students and employees of the
tice (especially between research and audit) may researcher. Do other cultures have the same
not be clear, even to the investigators them- exclusion classes? Historically, it would appear
selves.83 However, informed consent remains so – prisoners were once thought ideal ‘volun-
the paramount requirement for ethical research. teers’ for human subjects research.6 It is also
So the main barriers to consent lie in the nature rumoured occasionally that some cultures hold
of ‘informing’ the patient. These barriers to women to be incompetent to give consent to
consent depend on failures of understanding. their own treatment, though we found no evi-
The patient may fail to understand that they dence for this, and in practice this belief – in the
have a choice; and that they have a right to be absence of firm evidence to support it – is flatly
treated well, even if they refuse consent. They racist.
may not understand the concepts in the research, One cultural factor which has an impact on
such as the concept of the trial method, or of the consent is the existence of certain groups (such
therapies under test. And they may not under- as the Maori of New Zealand) which have a
stand the language used to seek consent. Lin- strong collectivist ethos, and which have historic-
guistic barriers (for instance, access to trans- ally been harmed by the dominance of other
lators) are easiest to resolve. Research Ethics groups’ interests. In New Zealand, any research
Committees are now well practised in ensuring which involves Maoris as subjects must be
that patient information sheets are clearly and scrutinised by the Maori community, and the
Ethics of Clinical Trials: Social Factors 33
community as a whole must give its consent wants and attitudes must be encouraged.90 Most
before individuals’ consent can be sought. of all, good practice must recognise that refusals
Philosophically, the force of a ‘group consent’ is and consents are similar in nature: both involve
rather controversial, but politically it is a neces- a similar mixture of reflection on evidence,
sity in that cultural and historical context, and is judgement of interests, cultural background,
widely accepted.84–86 socio-economic situation and past (individual
Another important cultural factor is the or community) experience. Some ‘consents’
family. In many societies, individuals will may be more worrying than some refusals, if
always consult with their family before making what the ‘consent’ rests on is a misunderstand-
a decision, and may in fact be unable to make a ing of the recruitment proposal. The main find-
decision with which they feel happy unless they ing of this study is that it is not possible simply
can so consult. Individuals may even reverse to separate out a ‘cultural factor’ in ethics
their own judgement if their family disagrees of clinical trials which can be dealt with, or
with them. Autonomy-centred cultures (like ignored, simply and apart from addressing edu-
those of Northern Europe) may find this peculiar cational, communication and political ‘barriers
or suspicious. Families can be as coercive as any to participation’.
other institution,87 but most recent research indi- Justice must be the main area of concern.
cates that the role of the family is beneficial, in There is little convincing evidence of any sys-
that it helps the patient to decide what their best tematic cultural objection to the RCT in general,
interests actually are.58 or to particular RCTs in practice. Naturally it is
In sum, it appears from the evidence we have good practice to be sensitive to cultural differ-
on consent that no culture actually disputes or ences when implementing health services and
disregards its importance. But we could dis- providing healthcare, though many of the rele-
tinguish between broader and narrower under- vant differences do not apply specifically to
standings of what consent involves. On the clinical trials, but rather to scientific medicine as
narrow understanding, all that matters is volun- such. Certain cultures have long-established,
tary agreement to the suggested intervention, and significantly different, philosophies of
and this is what is envisaged in the Nuremberg medicine. To the extent that individuals from
Code. On the broad understanding, what matters
this cultural background choose Western meth-
is autonomous, that is, self-determining, choice
ods of care, this is an indication of at least a
of the intervention: a decision that this is what I
provisional acceptance of the methods used for
want, rather than the more passive agreement to
do what you recommend. The broad under- applying and establishing the relevant medical
standing is not culturally universal, and different techniques. Moreover, individuals are increas-
understandings involving family, community, ingly sophisticated in deciding which types of
priest or expert authority take its place. What is medicine they will use and in which circum-
universal is the importance of the narrow under- stances they will use them.91 This applies to
standing of consent. Difficulties exist in estab- members of all cultures who live in contempor-
lishing that the decision is voluntary (that is, that ary multicultural society. Given all of this, the
it does not rest on a misunderstanding or on objections to trial methodology and to particular
implied coercion). However, a voluntary deci- trials, and the ethical sensibilities that such
sion is required and, except in extreme situations objections express, which must receive most
(clearly specified in advance) it is not to be attention in healthcare practice, are those which
negotiated away.88,89 are social and economic in nature.
Issues of justice in healthcare are now at the
top of most agendas, both in academic medical
ethics and in health services practice. A large
IMPLICATIONS FOR HEALTH SERVICE STAFF literature now exists on rationing and resource
AND PRACTITIONERS allocation. However, very little attention has
been focussed on priority setting for research
All of the above remarks, which are aimed at and health technology assessment. It is clearly
triallists and methodologists involved in the important for health services planning and prior-
design of research, apply equally to health service ity setting that fair resource allocation should
staff and practitioners. For instance, ways to keep one eye on the medium term, as well as on
incorporate families into the decision-making this year’s issues. Research should be targeted
process, for example, by allowing the patient to address present and future need trends. This
several days to consider trial entry, while main- would directly address the worry that much
taining that the final decision is the patient’s, clinical research may unjustly focus on highest
must be borne in mind. Sensitivity to cultural visibility conditions, rather than on highest need
variation and its impact on patients’ needs, conditions. The ‘Health of the Nation’ and ‘Our
34 The Advanced Handbook of Methods in Evidence Based Healthcare
Healthier Nation’ initiatives in the United King- SUGGESTIONS FOR FUTURE RESEARCH
dom are examples of how research could be
prioritised. But there is also the complementary
worry that the utilitarian focus on greatest need As we have made clear throughout, the evidence
may lead to ‘rare conditions’ receiving too little in this area is patchy, and all conclusions have
attention. had to be tentative. So there is a clear need for
Related to this issue of justice in targeting good-quality research into all the areas raised in
trials, is the issue of justice to participants after this chapter. The present state of academic
the trial. Many triallists arrange for participants knowledge of non-Western ethical traditions
to continue to receive the treatment free after the regarding medicine is weak. Research into these
completion of the trial, and patients may receive traditions is needed, but in principle little such
treatments on a ‘named patient’ basis if the research will bear on the topic of this chapter,
treatment seems to do them good, even if the simply because of the historical novelty of clin-
trial indicates that in general the treatment is ical trial methodology. A more appropriate form
unsafe or ineffective. These are only fair. Just of research would involve a focussed dialogue
governance of trials may require this in general, with representatives and authorities of these
where at present this is only an occasional traditions, cultures and communities, to deter-
practice. mine their responses to health services research
The main area where justice considerations methodology. A second key area for research is
apply is in the definition of population and in the development of tools for just priority setting
selection of subjects. This is primarily an issue in research and technology assessment. Addi-
for triallists, but health service managers and tionally, health technology assessments typically
referring physicians not directly participating in acknowledge the need for social and ethical
trials should also be aware of the concerns ‘impact assessment’ of the assessed technology,
raised above. but there is little consensus on what this is and
While the main onus of ethical trial practice how it is to be achieved. Finally, existing
falls on the triallists, the burden of ensuring that research into consent should continue: much
ethical standards are established and maintained attention has been given to addressing research-
belongs to Local and Multi-centre Research ers’ problems (poor recruitment rates, lack of
Ethics Committees (or Institutional Review understanding by patients, poor recall of the
Boards). Research Ethics Committees are very detail of consent after the fact, and so on). Less
experienced in considering whether proper attention has been paid to why patients join or
mechanisms exist for obtaining informed con- refuse to join trials, although altruism is often
sent, in obtaining assessments of the scientific hypothesised, and compliance rates are much
validity of the trial, and in scrutinising the risks studied as part of the trial analysis itself.
and benefits of the trial to participants. Two Because there is little attention to the decision-
areas where they are less sure-footed are the making of patients it is hard to interpret refusals;
assessment of inclusion and exclusion criteria and it is refusals which indicate objections, and
for fairness and statistical soundness, and aware- objections which have been – perforce – our
ness of the relevance of socio-cultural factors to main source of relevant data drawn from the
their deliberations. The first issue could usefully studies reviewed in this chapter. It is too early to
be addressed by ensuring that Committees either
say whether there are in fact cultural objections
have a statistician member or ready access to
to trials, but it is clear that the social and
statistical support. Such support may need to be
economic context of a trial has a large impact on
independent of the Research and Development
its success, and is a major component in the
Support Units often used by researchers in
designing their studies. Both issues can be ethical status and public acceptability of the
addressed by appropriate training, and by con- trial.1
tact (perhaps through the Community Health
Councils or their equivalents) with community
groups in their area. It is worth noting that one ACKNOWLEDGEMENTS
of the few reasons a Local Research Ethics
Committee can withhold approval from a study
approved by a Multi-centre Research Ethics This chapter is partly based on work presented
Committee is the existence of particular local in Ashcroft RE, Chadwick DW, Clark SRL,
factors. These local factors will in practice be Edwards RHT, Frith L, Hutton JL, ‘Implications
precisely the social and cultural issues discussed of Socio-cultural contexts for ethics of clinical
in this chapter, because it is these factors (rather trials’. Health Technology Assessment, 1997;
than general ethical issues) which are local. 1(9).1 This work was funded by NHS R&D
Ethics of Clinical Trials: Social Factors 35
Health Technology Assessment programme 17. Roberson NL. Clinical trial participation: view-
grant 93/41/4. Dr Ashcroft thanks Professors points from racial/ethnic groups. Cancer 1994;
Chadwick, Clark, and Edwards, Ms Frith and Dr 74: 2687–91.
Hutton (the lead applicant) for their encourage- 18. Dula A. African American suspicion of the health-
ment and backing. care system is justified: what do we do about
it? Cambridge Q. Healthcare Ethics 1994; 3:
347–57.
19. Gamble VN. A legacy of distrust: African
REFERENCES Americans and medical research. Am. J. Prevent.
Med. 1993; 9(suppl. 3): 35–8.
1. Ashcroft RE, Chadwick DW, Clark SRL, 20. Winn RJ. Obstacles to the accrual of patients to
Edwards RHT, Frith L, Hutton JL. Implications clinical trials in the community setting. Semin.
of socio-cultural contexts for ethics of clinical Oncol. 1994; 21(suppl. 7): 112–17.
trials. Health Technology Assessment 1997; 1(9). 21. Sperber AD, Henkin Y, Shanty S. Methodo-
2. Hill AB. Medical ethics and controlled trials. Br. logical issues in conducting a community-based
Med. J. 1963; 1: 1043–9. clinical drug trial. Family Practice Res. J. 1993;
3. Levine RJ. Ethics and regulation of clinical 13: 311–21.
research. Baltimore: Urban & Schwarzenberg, 22. French GR. When manufacturers are unwilling to
1986. accept the risks or invest the capital: vaccine
4. Beauchamp TL, Childress JF. Principles of bio- production by the Salk Institute to the specifica-
medical ethics, 4th ed. Oxford: Oxford University tion of the US army. AIDS Research and Human
Press, 1994. Retroviruses 1994; 10(suppl. 2): S309.
5. Epstein S. Democratic science? AIDS activism 23. Hutton JL, Ashcroft RE. What does ‘systematic’
and the contested construction of knowledge. mean for reviews of methodology? In: Black N,
Socialist Rev. 1991; 21: 35–64. Brazier J, Fitzpatrick R, Reeves B (eds). Methods
6. Rothman DJ. Ethics and human experimentation: for Health Care Research: A Guide to Best Prac-
Henry Beecher revisited. N. Engl. J. Med. 1987; tice. London: BMJ Books, 1998: 249–54.
317: 1195–9. 24. Mike V. Philosophers assess randomised clinical
7. Cochrane AL. Effectiveness and efficiency: ran- trials: the need for dialogue. Controlled Clinical
dom reflections on health services. London: Trials 1989; 10: 244–53.
Nuffield Provincial Hospitals Trust, 1971. 25. Zelen M. Randomised consent designs for clinical
8. Hutton JL. The ethics of randomised controlled trials: an update. Statist. Med. 1990; 9: 645–56.
trials: a matter of statistical belief? Health Care 26. Altman DG, Whitehead J, Parmar MKB,
Analysis 1996; 4: 95–102. Stenning SP, Fayers PM, Machin D. Randomised
9. Palmer CR. Ethics and statistical methodology in consent designs in cancer clinical trials. Eur. J.
clinical trials. J. Med. Ethics 1993; 19: 219–22. Cancer 1995; 31A: 1934–44.
10. Urbach P. The value of randomisation and control 27. Logue G, Wear S. A desperate solution: indi-
in clinical trials. Statist. Med. 1993; 12: 1421–41. vidual autonomy and the double-blind controlled
11. Papineau D. The virtues of randomisation. Br. J. experiment. J. Med. Philos. 1995; 20: 57–64.
Philos. Sci. 1994; 45: 712–15. 28. Angell M. Patients’ preferences in random-
12. Thornton H. Clinical trials – a brave new partner- ised clinical trials. N. Engl. J. Med. 1984; 310:
ship? J. Med. Ethics 1994; 20: 19–22. 1385–7.
13. Baum M. Clinical trials – a brave new partner- 29. Mackie JL. Ethics: Inventing right and wrong.
ship: a response to Mrs Thornton. J. Med. Ethics Harmondsworth: Penguin, 1977.
1994; 20: 23–5. 30. Alderson P. Trust in informed consent. IME
14. McPherson K. The best and the enemy of the Bulletin 1988: 17–19.
good: randomised controlled trials, uncertainty 31. Lenhard RE Jr. A large private university
and assessing the role of patient choice in medical hospital. Cancer 1993; 72: 2820–3.
decision-making. J. Epidemiol. Community Health 32. Freeman HP. The impact of clinical trial
1994; 48: 6–15. protocols on patient care systems in a large city
15. Llewellyn-Thomas HA, McGreal MJ, Thiel EC, hospital: access for the socially disadvantaged.
Fine S, Erlichman C. Patients’ willingness to Cancer 1993; 72: 2834–8.
enter clinical trials: measuring the association 33. Christakis NA. Ethics are local: engaging cross-
with perceived benefit and preference for decision cultural variation in the ethics of clinical research.
participation. Soc. Sci. Med. 1991; 32: 35–42. Soc. Sci. Med. 1992; 35: 1079–91.
16. McNeill PM. The ethics and politics of human 34. Davis DS. It ain’t necessarily so: clinicians, bio-
experimentation. Cambridge: Cambridge Uni- ethics and religious studies. J. Clin. Ethics 1994;
versity Press, 1993. 5: 315–19.
36 The Advanced Handbook of Methods in Evidence Based Healthcare
35. Engelhardt HT JR. The Foundations of Bioethics, Medicine. Kennedy Institute of Ethics J. 1994; 4:
2nd edn. Oxford: Oxford University Press, 1996. 55–61.
36. Basch TF. Technology transfer to the developing 54. Annas GJ, Grodin MA (eds). The Nazi doctors
world: does new technology have any relevance and the Nuremberg Code: human rights in human
for developing countries? Tubercle and Lung experimentation. Oxford: Oxford University Press,
Disease 1993; 74: 353–8. 1992.
37. Maier-Lenz H. Implementing multicentre, multi- 55. Ishiwata R, Sakai A. The physician–patient
national clinical trials. Drug Information J. 1993; relationship and medical ethics in Japan. Cam-
27: 1077–81. bridge Q. Healthcare Ethics 1994; 3: 60–6.
38. Enquselassie F, Nokes J, Cutts F. Communities’ 56. Sanwal AK, Kumar S, Sahni P, Nundy P.
confidentiality should be maintained and commun- Informed consent in Indian patients. J. R. Soc.
ity consent sought. Br. Med. J. 1996; 312: 54–5. Med. 1996; 89: 196–8.
39. O’Neil JD. The cultural and political context of 57. Levine RJ. Informed consent: some challenges to
patient dissatisfaction in cross-cultural encounters: the universal validity of the Western model. Law,
a Canadian Inuit study. Medical Anthropology Medicine and Healthcare 1991; 19: 207–13.
Quarterly 1989; 3: 325–44. 58. Nelson HL, Nelson JL. The patient in the family:
40. Carrese JA, Rhodes LA. Western bioethics on the an ethics of medicine and families. London:
Navajo reservation: benefit or harm? JAMA 1996; Routledge, 1995.
274: 826–9. 59. Zion D. Can communities protect autonomy?
41. Lockwood M, Anscombe GEM. Sins of Omis- Ethical dilemmas in HIV preventative drug
sion? The non-treatment of controls in clinical trials. Cambridge Q. Healthcare Ethics 1995; 4:
trials Aristotelian Society 1983; Suppl. LVII: 516–23.
207–27. 60. Tobias JS, Souhami RL. Fully informed consent
42. DeHeyn PP, d’Hooge R. Placebos in clinical can be needlessly cruel. Br. Med. J. 1993; 307:
practice and research. J. Med. Ethics 1996; 22: 1199–201.
140–6. 61. Thornton H. Many subjects in trial were not
43. Rothman KJ, Michels KB. The continuing un- asked for consent. Br. Med. J. 1996; 312: 509.
ethical use of placebo controls. N. Engl. J. Med. 62. Alderson P. Consent and the social context.
1994; 331: 394–8. Nursing Ethics 1995; 2: 347–50.
44. Freedman B. Equipoise and the ethics of clinical 63. Habermas J. Communication and the evolution of
research. N. Engl. J. Med. 1987; 317: 141–5. society. London: Heinemann Educational, 1979.
45. Johnson N, Lilford RJ, Brazier W. At what level 64. Silverman WA. The myth of informed consent in
of collective equipoise does a clinical trial daily practice and in clinical trials. J. Med. Ethics
become ethical? J. Med. Ethics 1991; 17: 30–4. 1989; 15: 6–11.
46. Alderson P. Equipoise as a means of managing 65. Lilleyman JS. Informed consent: how informed
uncertainty: personal, community and proxy. J. and consent to what? Pediat. Hematol. Oncol.
Med. Ethics 1996; 22: 135–9. 1995; 12(6): xiii–xvi.
47. Lilford RJ, Thornton JG, Braunholtz D. Clinical 66. Burchell HB. Vicissitudes in clinical trial research:
trials and rare diseases: a way out of a conun- subjects, participants, patients. Controlled Clinical
drum. Br. Med. J. 1995; 311: 1621–5. Trials 1992; 13: 185–9.
48. Merton V. The exclusion of the pregnant, preg- 67. Brownlea A. Participation: myths, realities and
nable and once-pregnable people (a.k.a. women) prognosis. Soc. Sci. Med. 1987; 25: 605–14.
from biomedical research. Am. J. Law & Medicine 68. Berry DA. A case for Bayesianism in clinical trials.
1993; XIX: 369–451. Statist. Med. 1993; 12: 1377–404.
49. Merkatz RB, Junod SW. Historical background of 69. Whitehead J. The case for frequentism in clinical
changes in FDA policy on the study and evalu- trials. Statist. Med. 1993; 12: 1405–19.
ation of drugs in women. Academic Medicine 70. Schulz KF, Chalmers I, Hayes RJ, Altman DG.
1994; 69: 703–7. Empirical evidence of bias: dimensions of meth-
50. McCarthy CR. Historical background of clinical odological quality associated with estimates of
trials involving women and minorities. Academic treatment effects in controlled trials. JAMA 1995;
Medicine 1994; 69: 695–8. 273: 408–12.
51. Levinsky NG. Social, institutional and economic 71. D’Agostino RB, Kwan H. Measuring effective-
barriers to the exercise of patients’ rights. N. ness: what to expect without a randomised control
Engl. J. Med. 1996; 334: 532–4. group. Medical Care 1995; 33: AS95–AS105.
52. Dresser R. Wanted: single, white male for 72. Berry DA. Interim analyses in clinical trials:
medical research. Hastings Center Report 1992; classical versus Bayesian approaches. Statist. Med.
22: 24–9. 1985; 4; 521–6.
53. Mastroianni AC, Faden R, Federman D. Women 73. Caplan AL. Is there a duty to serve as a subject in
and health research: a report from the Institute of biomedical research? In: If I were a rich man,
Ethics of Clinical Trials: Social Factors 37
could I buy a Pancreas? And other essays on the 83. Reiser SJ. Criteria for standard versus experi-
ethics of healthcare. Bloomington: Indiana Uni- mental care. Health Affairs 1994; 13: 127–36.
versity Press, 1992. 84. Campbell AV, Charlesworth M, Gillett G, Jones
74. Emson HE. Rights, duties and responsibilities in G. Medical Ethics, 2nd edn. Oxford: Oxford
healthcare. J. Appl. Philos. 1992; 9: 3–11. University Press, 1997.
75. Parfit D. Reasons and Persons. Oxford: Oxford 85. Verdun-Jones S, Weisstub DN. Consent to human
University Press, 1984. experimentation in Quebec: the application of the
76. Nuremberg Code and Helsinki Declaration, both civil law principle of personal inviolability to
reprinted in Annas and Grodin, op. cit. (54). protect special populations. Int. J. Law Psychiatry
77. Fox R. What do patients want from medical 1995; 18: 163–82.
research? J. R. Soc. Med. 1996; 89: 301–2. 86. Kymlicka W. Liberalism, Community and Culture.
78. Kee F. Patients’ prerogatives and perceptions of Oxford: Oxford University Press, 1989.
benefit. Br. Med. J. 1996; 312: 958–60. 87. Anonymous. The case of the coercive family.
79. Ashcroft RE. Human research subjects, selection Cambridge Q. Healthcare Ethics 1995; 5:
of. In: Chadwick R (ed.), Encyclopedia of Applied 135–42.
Ethics. San Diego: Academic Press, 1997, vol. 2: 88. Jonas H. Philosophical reflections on experi-
627–39. menting with human subjects. Daedalus 1969;
80. Antman K. Reimbursement issues facing patients, 98: 219–47.
providers and payers. Cancer 1993; 72: 2842–5. 89. Fethe C. Beyond voluntary consent: Hans Jonas
81. Royal College of Physicians. Research involving on the moral requirements of human experimenta-
patients: summary and recommendations. J. R. tion. J. Med. Ethics 1993; 19: 99–103.
Coll. Physicians 1990; 24: 10–14. 90. Gostin LO. Informed consent, cultural sensitivity
82. Karlawish JHT, Hall JB. The controversy over and respect for persons JAMA 1995; 274: 844–5.
emergency research: a review of the issues and 91. Bhopal RS. The interrelationship of folk, tra-
suggestions for a resolution. Am. J. Resp. Crit. ditional and western medicine within an Asian
Care Med. 1996; 153: 499–506. community. Soc. Sci. Med. 1986; 22: 99–105.
3
Factors that Limit the Number, Progress
and Quality of Randomised Controlled
Trials: A Systematic Review
clinically important difference. There is there- sections depending on the type of material iden-
fore a high risk of drawing a false conclusion – tified for that part of the review.
that there is no difference when one does exist.
Furthermore, reported differences may distort
the truth as only large effects reach ‘statistical THE EVIDENCE AND IMPLICATIONS FOR
significance’ in small trials. The pooling of data METHODOLOGISTS
through meta-analysis sometimes resolves this
problem. Nevertheless, it needs great care and
considerable effort, and the findings may be The literature identified for this methodological
distorted by publication bias, inadequate report- review differed in a number of important ways
ing of trial methods, insecure random allocation, from reviews of effectiveness or efficacy.
losses to follow-up, and failure to base analyses Firstly, theory needed to be given relatively
on the randomised groups. more weight than empirical evidence, largely
These concerns about the design, conduct, because there were many areas of the review
analysis and reporting of RCTs led us to carry where it was not possible to carry out empirical
out a systematic review of the literature on trials research. In other areas where evidence would
with two main aims: have been expected, it was rare, and rigorous
evidence even rarer. The ‘best’ method for
1 To assemble and classify a comprehensive recruiting patients to randomised trials provides
bibliography of factors limiting the number, an apt example. Whether we applied the cri-
progress and quality of RCTs. terion of effectiveness (p. 43) or that of cost
2 To collate and report the findings, identify- effectiveness (p. 48), we were unable to make
ing areas where firm conclusions can be any clear recommendations based on com-
drawn and areas where further research is parative studies. This can be attributed both to
necessary. the general lack of rigour in the published com-
parisons, none of which was experimental, and
to the wide range of topics and contexts in
METHODS which these comparisons were set. The variety
of published material ensured that the character
We undertook a systematic review of the lit- of each section of the review differed enor-
erature on trials, based on three bibliographic mously: an introductory paragraph at the start of
databases, Medline, Embase and CINAHL, each section highlights the main features.
covering the period 1986 to 1996. The scope of
the review was too broad to be comprehensive Design Issues
in all of the areas covered. Rather, the review
attempted to cover the diversity of factors limit- The evidence for this section tended not to be
ing the quality, number and progress of RCTs. the outcome of empirical research, but rather
The issues were considered under the follow- the result of reviews of research methods and
ing categories: design issues; barriers to par- theoretical discussion. In addition, text books on
ticipation; conduct and structure; analysis; trial design were considered, for example that by
reporting; and costs. After initial screening of Pocock.1 Using the evidence, various recom-
references identified electronically, all those mendations for trial design are proposed.
English language references considered ‘pos-
sibly relevant’ were obtained. From each full Defining participants, interventions and
text article, data were extracted on category
of reference (as above), study methodology
outcomes
(primary research, reviews, other articles), sub- We recommend that, following a systematic
ject area, intervention type, clinical activity (pre- review of existing evidence, a relevant well-
vention, treatment, screening or diagnosis), and formulated question should be developed, speci-
setting (hospital, general practice, or com- fying participants, interventions and outcomes.
munity). A scoring system was used to classify Wide patient eligibility criteria are generally
each article from 3, ‘highly relevant’, to 0, preferred to give representativeness and good
‘irrelevant’, for each category. There was con- recruitment rates.2,3 However, a more homo-
siderable overlap in relevant articles between geneous group may be preferable when evaluat-
categories, for example relevant articles for ing expensive or hazardous interventions. The
‘design’ were often relevant for ‘conduct’. choice of treatment and comparator should take
All references identified of relevance to indi- into account the results of previous research and
vidual subject areas were passed to the author the objectives of the trial: the use of a placebo or
responsible for that section of the review for active comparison treatment, and the need for
synthesis. Methods of synthesis varied between a run-in phase should be considered. Several
Factors Limiting Number, Progress and Quality of RCTs 41
settings and geographical areas may not be Tognoni et al. for a failed trial was that the
ideal. Despite the limitations of the papers, questions to be tested were not of sufficient
many barriers to participation were indeed interest to participating clinicians.28
found. Synthesis of the material was into three
major categories: barriers to clinician recruit- Barriers to patient participation
ment; barriers to patient recruitment; and the
clinician as a barrier to patient recruitment. The review confirmed that the additional
demands of a study may cause concern for some
patients, influence their decision to participate,
Barriers to clinician participation and lead to later ‘drop-out’. The main causes of
Lack of time was considered a major barrier in concern were additional procedures and appoint-
several studies, including the time pressures ments which may cause discomfort, incon-
from usual clinical practice and management venience or additional expense.34–39 Travel and
duties: such pressures may be increasing for travel costs were unpopular with participants
clinicians.19,20 The time demands of recruitment, and were reasons for refusing to take part, for
the consent process, and follow-up in trials may missing appointments and for ‘dropping-out’.
be a further barrier.19,21–24 Clinicians participat- A strong preference for, or against, a particu-
ing in clinical trials may be ill-prepared for a lar treatment (offered in the trial or outside it)
research role because of inadequate research can limit recruitment. Such preferences included
experience,20,25 training,26 or clinical skills.27 a wish not to change medication,40,41 not to
Lack of support staff, for example clinical trial take placebo,35,42 not to take an ‘experimental’
nurses, has been blamed for poor clinician medication, or not to take any medication.41
recruitment.19,26–28 Some patients requested a specific interven-
Participation in RCTs may alter the doctor– tion.35,38,42–44 Patients may choose not to take part
patient relationship and concern about this may in research at all, 29,37,44,45 or have an aversion to
act as a barrier. The main issues highlighted by treatment choice by random allocation.36,38,46
a number of authors were the difficulty for Patients may find the issues of uncertainty
clinicians of admitting that they do not know difficult to cope with, particularly if they feel
which treatment is best21,23,24 and the perceived that the efficacy of the treatment on offer is
conflict between the clinician role and research unproven, and may have a distrust of hospital or
(or scientific) role.27,29,30 As a result, some clini- medicine and fear the unknown.41,47 Under such
cians feel that their rapport with patients can be circumstances, they find it unpleasant to decide
damaged by participation in a trial.21,23,31 about taking part in a trial48 and may even prefer
Concern of the clinician for the patient may the doctor to make the decision.49
affect their decision to take part in a trial or Information and consent emerged as major
not, including worry about treatment toxicity or concerns. Evidence about the provision of infor-
side effects,19,25 the burden of the trial for mation on trials was conflicting, because patients
patients,19,22,24,28,29 or a reluctance to recruit – both participating and non-participating –
severely ill patients.32 Benson and colleagues wanted more information on trials,39,40,48,50
found that a fear of feeling responsible if the although providing information may reduce
patient did not receive the treatment which recruitment and alter the response to treat-
turned out to be best deterred surgeons from ment.45,46,51 Many authors have recommended
taking part in a trial.21 that the information should be simply presented,
Loss of clinical autonomy, including loss including Harth and Thong,38 Bevan et al.,42
of decision-making power and independence, Antman et al.,47 Roberson51 and Jensen et al.52
being accountable to a third party, and restric- For a more detailed discussion of trials of strat-
tion of the ability to individualise patient care, egies for patient recruitment, see Chapter 1.
were also causes for concern.21,23,30,33 Obtaining
consent can also be an important barrier to Clinician as barrier to patient
clinician participation.21,23,24,31
It is unclear whether lack of financial rewards
participation
is a deterrent to clinician participation.19,28 Atwood and colleagues35 and Sutherland and
Schaeffer and colleagues stressed the need for colleagues53 found that an ‘important person’
personal encouragement and support to achieve (e.g. spouse, family member, close friend) had a
successful participation.33 considerable influence on the decision to partici-
Two studies mentioned the importance of the pate, with patients unlikely to participate if those
research question. Scientifically uninteresting close to them were against it. However, the
trials were felt to be an impediment to recruit- doctor recruiting to the trial has been reported
ment by 17% of respondents in a survey by by many authors to have the greatest influence
Ashery and McAuliffe.19 One reason given by on the decision to enter a trial.34,49,52 Clinicians
Factors Limiting Number, Progress and Quality of RCTs 43
themselves may therefore act as a barrier to solutions experienced in individual trials, and
patient recruitment, even in trials where they most review articles were based on anecdotal
have agreed to take part. The protocol itself was evidence. Some more rigorous studies, including
blamed for restricting recruitment in several trials were also identified. From the articles, a
studies,19,24,37,43,54,55 including incompatibil- number of problems were identified which can
ities of the protocol with usual clinical prac- arise during the conduct of a trial, to prevent it
tice21,23,29,56,57 or excessive and poorly designed starting, or limit its progress and quality, along
data collection.24,54 with possible solutions.
Clinicians may find giving information diffi-
cult, when describing both clinical trials in gen- Failure to start
eral and specific trials,29,57 with particular con-
cerns about assessing the level of information There are no good data on the number of ran-
required by patients.23,27 Even amongst clini- domised trials that fail to start, however one
cians who have agreed to collaborate in a trial, survey by Easterbrook and Matthews of clinical
the level of information and the process of studies which had received ethics approval
obtaining consent vary.58 found that 21% never started largely due to lack
of funding, the principal investigator leaving the
Recommendations for clinician and patient institution, logistic problems or expected recruit-
ment problems.20
recruitment
The review set out to identify barriers to partici- Poor recruitment
pation in trials, rather than strategies to over-
come such barriers. The findings confirm results Recruitment is often a major problem in RCTs.
of previous reviews of the subject (for example Easterbrook and Matthews estimated that half of
by Hunninghake et al.59 and Swanson and all clinical studies fail to reach their planned
Ward60) suggesting that even though evidence is size.20 A survey by Coombs and colleagues of
available on the topic, solutions have not been recruitment amongst a cohort of 41 trials
implemented. Despite the limitations of our showed only 34% achieved or surpassed their
review, it is possible to make recommendations planned size, whilst a further 34% recruited less
from the findings. than 75% of their planned size.6 In addition to
To overcome problems with clinician recruit- problems in the design (such as narrow inclusion
ment, the RCT should address a question that is criteria) and participation (discussed earlier),
seen as important. The protocol should be clear overestimating the pool of eligible patients or the
and simple, with the minimum required of recruitment rate, using too few recruitment strat-
clinicians, and straightforward data collection. egies, and lack of active involvement of the staff
Where possible, dedicated research staff should in recruitment, have also been blamed.59,61
be available to provide support and encourage- Poor recruitment can increase the cost of the
ment for the clinical staff. Pragmatic trial study, or lead to its abandonment or reduced
designs are likely to be more acceptable to size and hence loss of statistical power.19,20,59 In
clinicians, since this type of design permits more one trial, Neaton and co-workers found that poor
clinical freedom. recruitment led to falsification of recruitment.62
To overcome barriers to patient recruitment, Possible solutions suggested by various authors
the demands of the study should be kept to a (for example by Ashery and McAuliffe19 and
minimum consistent with the scientific purpose Hunninghake et al.59) include careful planning
of the study, with the extent and purpose of the and piloting of recruitment, use of multiple
study and investigations clearly explained. research strategies, close monitoring of recruit-
Patients should be supported in what may be a ment (using recruitment logs), the use of recruit-
difficult decision to take part in a RCT, and ment co-ordinators, setting recruitment goals,
should not be placed under pressure to do so. having contingency plans in case of poor recruit-
Dedicated research staff may help with pro- ment, and providing training in recruitment.
viding information and obtaining patient con-
sent, and monitoring the recruitment process. Early stopping
The recruitment aspects of a RCT should be Trials may stop early because of logistical prob-
carefully planned and piloted to avoid poor lems, because of the results of other similar
recruitment. trials, because the intervention being tested is
superseded, or because interim analyses demon-
Conduct and Structure strate unequivocal evidence of harm or benefit,
or that the trial has little chance of showing a
Many of the original articles identified for this clinically important benefit.20,63–65 Interim analy-
section simply reported problems and possible ses should be planned, be based on good-quality
44 The Advanced Handbook of Methods in Evidence Based Healthcare
data and use appropriate statistical tests and Drop-outs and loss to follow-up
stopping guidelines to avoid inappropriate con-
clusions.63–65 Any decision to stop a trial early The results of a trial may be invalid if a large
should be based on clinical and ethical con- proportion of those randomised are excluded
siderations, as well as statistical ones.63–66 from analysis because of protocol violation, or
are lost to follow-up. Surveys of the reporting of
trial results have shown that trials often exclude
Compliance a significant number of randomised patients
Poor compliance with a trial protocol either by from the analysis (for example, surveys by
the investigator or the patient can have import- Rossetti et al.,10 Liberati et al.,80 Kleijnen
ant effects on the overall result, although the et al.,81 Solomon et al.,82 Sonis and Joines,83
effects will vary depending on the intervention Lionetto et al.,84 Schulz et al.85). It is important
under study.67 Investigators may recruit ineligible to collect as many details as possible about the
patients, prescribe prohibited medication, fail to patient at the time of recruitment to help with
prescribe or carry out the allocated treatment follow-up, and tagging records in national regis-
adequately, or transfer the patient from one arm ters may help.86 Patients who withdraw from
of the trial to another (so-called crossovers).68,69 treatment, or who violate the protocol, should
still be followed-up to allow an intention-to-treat
Participants may take too much or too little
analysis (see p. 45).
medication, may take it at inappropriate intervals,
may fail to attend for follow-up appointments,
or fail to undergo certain tests.68,70–72 Blinding
Poor compliance reduces the power of a trial During a trial any of the following individuals
to identify a treatment effect, resulting in the may be blinded to treatment allocation to reduce
need for increased sample size or an increased the risk of bias: the patient/participant, the
risk of a false-negative or false-positive error.73–75 healthcare professional, or the outcome assessor.
It is therefore important to measure compliance It is not always possible to blind the patient or
to assess the risk of false-positive or -negative health professional, for example in surgical
results and to evaluate acceptability of the treat- trials, but it is usually possible to blind the
ment to patients. outcome assessor. In several ‘double-blind’
However, compliance appears to be rarely trials (where doctors and patients are supposed
assessed in trials,67 perhaps due to difficulties in to have been unaware of treatment allocation),
defining and measuring it.72,71 Vander Stichele doctors and patients became unblinded during
suggested defining patients as compliers, partial the trial because of treatment effects, side-
compliers, overusers, erratic users, partial drop- effects, or the lack of them. Unblinding may
outs and drop-outs, but did not give clear defini- lead to bias because the doctor may monitor or
tions.74 Measuring compliance with pharmaco- treat one group differently, or the patient’s
logical treatments is possible, although all knowledge may affect the outcome:87,88 unblind-
methods are flawed.70–72,74,76 Defining and meas- ing the outcome assessor may also lead to
uring compliance with non-pharmacological bias.89,90 We therefore suggest that outcome
treatments is even more difficult.77,78 assessment should be blind wherever possible.
The effort expended on improving compli-
ance depends on the type of trial: in explanatory
or efficacy trials, compliance is essential; in Quality assurance
pragmatic or clinical management trials, inter- It is extremely important that data collected and
ested in assessing whether the treatment is effec- processed during a trial should be accurate.91,92
tive for the average patient under circumstances Good clinical practice guidelines emphasise the
that reflect real life, compliance should also need for quality assurance and audit throughout
reflect real life.72,74 Several solutions have been the study to identify fraudulent or careless
proposed to improve compliance in trials: edu- practice.91–93 However, high levels of quality
cation about the trial; assessing compliance dur- assurance increase the administrative cost and
ing a run-in period to exclude obvious non- workload of trials, and may not always be
compliers; simplifying treatment and follow- feasible.94
up regimens; and maintaining communication
with participants by newsletter or telephone
calls.35,56,61 (For further discussion of trials-
Collaboration and communication
assessing methods to improve patient compli- Problems can arise in trials that require collab-
ance with non-trial medications, see the recent oration amongst a number of different health
review by Cochrane.79) professionals and agencies, although these can
Factors Limiting Number, Progress and Quality of RCTs 45
usually be solved through good communica- and Rockhold and Enas have argued that it
tion.19 Inexperienced trialists should be sup- is not feasible for all trials to have a Data
ported by experienced trialists,6 and the princi- Monitoring Committee due to the cost and lack
pal investigator should provide clear leadership of experienced statisticians and independent
through to the completion of the trial.95 Multi- experts.64,99 They recommend formal Data Mon-
centre trials provide additional problems with itoring Committees for large ‘pivotal’ trials, that
requirement for collaborators’ meetings, site is those measuring life-threatening end-points
visits and possibly translation of trial material and those where there is the potential for serious
into appropriate languages.95,96 Staff turnover at side-effects. In other trials, interim analyses
sites can hinder the progress of trials,97 and the could be carried out by the trial statistician and
loss of the principal investigator can be particu- principal investigator with the option of forming a
larly disruptive.20 Data Monitoring Committee if necessary.66,99,100
Interim analyses should not be released to other
Conflict of interest investigators unless there are good reasons: evi-
dence from a study by Green and colleagues
Van Gijn and Bogousslavsky discussed the
suggested that open reporting of interim results
problems of conflict of interest.93,98 Investigators
was associated with declining recruitment, stop-
can be involved in several trials simultaneously,
ping trials without meeting defined objectives
and this could lead to conflict of interest,
and early publication of results found later to be
although it may be minimised if the inclusion
inconsistent with final results.101
criteria are clearly different. Commercially
The concern that involvement of a com-
sponsored trials where investigators receive a
mercial sponsor in the design, data collection,
large honorarium for each patient entered are
analysis and reporting of a trial may lead to bias
likely to be given preference over unsponsored
in favour of their product is difficult to sub-
trials, and this may not be in the public interest.
stantiate. However, there are examples where
Bogousslavsky and van Gijn suggested that
bias has occurred, such as: the reporting of data
investigators should receive reimbursement only
from only favourable centres; the use of dosage
for the extra work involved in entering patients
regimens that favour the drug manufactured by
into the trial, and that no additional honoraria
the sponsor; and the reporting of conclusions
should be given. However, calculating the cost
that do not match the data presented.94,102 This is
of this additional work may be difficult.
particularly worrying since a survey from one
region of the UK showed that 65% of trials are
Organisational structure sponsored by the pharmaceutical industry.103
Randomised controlled trials probably need to Van Gijn recommended that pharmaceutical
have a good organisational and administrative industry sponsored trials also required independ-
base to succeed, but little has been published ent Steering Committees and Data Monitoring
about this. Large multi-centre trials need a Committees.93
Steering Committee with overall responsibility
for the trial, a co-ordinating centre to handle Analysis
materials, data collection and communication
with trialists, a trial co-ordinator to run the co- Because of the variety of trial designs and types
ordinating centre, and a data monitoring com- of data, discussion of analysis is limited to those
mittee.95 Smaller trials probably need a scaled- areas which have attracted criticism in the lit-
down, but similar, structure. Some individuals erature, where empirical evidence leads us to
may be on committees for several trials which make recommendations for practice. Hence im-
could lead to conflicts of interest:98 we did not portant areas, such as discussion of the Bayesian
identify research relating to this issue, or how and frequentist philosophies, are excluded. (For
members of Steering Committees or Data a full discussion of Bayesian and frequentist
Monitoring Committees are selected. philosophies, see Chapter 16.)
The main role of the Data Monitoring Com-
mittee is to review planned interim analyses to
see if there is accumulating evidence to suggest
Intention-to-treat analysis
that the trial should be stopped early, or in some The principle of ‘intention-to-treat’ analysis
instances prolonged.66 Data Monitoring Com- arises from randomised trials where, for a num-
mittees should be: independent from the trial ber of reasons, the treatment to which a patient
and the sponsor, include experts from all neces- is randomised might not be received, in full or
sary disciplines, have explicit terms of reference, part, by that patient. In these circumstances,
have access to the necessary data, collaborate there is a concern that those patients who do not
with other similar trials, and have appropriate receive the allocated treatment may well differ
resources.64–66 However, Fleming and DeMets, in some way from those who do receive it, and
46 The Advanced Handbook of Methods in Evidence Based Healthcare
that the selection processes may differ between normally taken in randomised trials). A separate
treatment groups. An analysis based on only estimate of the treatment difference will also be
those patients who received the treatment as obtained from those patients rejecting random-
randomised would therefore give a biased esti- isation but selecting one of the treatments under
mate of treatment differences. This bias can be investigation. Analysis of the results focuses on
mitigated by analysing according to the groups inferences about the difference between the size
into which patients were randomised. If out- of the treatment effects in the randomised and
come observations are available on all patients, non-randomised subgroups, with formal sig-
treatment comparisons are potentially unbiased nificance testing being based on a ‘treatment
(though of course bias may arise from some by randomisation’ interaction. The main dis-
other aspect of the trial’s design or conduct). advantage of such an approach is an increased
Another advantage of intention-to-treat is that it workload for trialists, and its routine use is not
mimics ‘real life’ with respect to errors in treat- recommended.112,113
ment and poor adherence to treatment, and gives
a more realistic assessment of treatment policy. Multiple end-points
The majority of papers dealing with general
issues in the analysis of RCTs emphasise the Multiple end-points are one of several ways in
importance of intention-to-treat analysis.65,104 which the interpretation of significance tests can
Studies by Peduzzi and colleagues, and Lee and be affected by a multiplicity of tests being
colleagues which compared the use of intention- performed. A systematic review by Pocock and
to-treat analysis to alternative strategies, con- colleagues of 45 RCTs in prestigious journals
cluded that only intention-to-treat is recom- found a median of six end-points being reported,
mended.105,106 However, in systematic reviews with a median of four significance tests being
reporting the proportion of trials presenting reported.114 They highlighted the particular mul-
intention-to-treat analysis, adherence to the prin- tiple testing problem of analysis of repeated
ciple is not widely demonstrated.9,10,83,107,108 measures over time, seriously increasing the risk
Concerns about intention-to-treat analysis of finding a difference where one does not
include the possible need for a larger sample exist.
size than the ‘true’ treatment difference would
suggest, concerns about trials where large num- Subgroup analysis
bers of patients have crossed over to the altern- An understandable objective of trialists can be
ative management option, concern about ignor- to identify subgroups of patients who respond
ing compliance, reduction in power, and loss particularly well to a particular treatment. Statis-
of analysis of a treatment’s efficacy (see for tical considerations highlight the dangers of
example, Rabeneck et al.109 and Sheiner and multiple testing arising from this approach, such
Rubin110). An alternative approach to intention- that chance alone will be responsible for
to-treat analysis is the ‘per protocol’ analysis, in enhanced treatment differences within some
which only subjects conforming to the protocol subgroups. The review cited above found that
are subject to analysis. In addition to the con- 51% of trials reported subgroup analyses.114
cerns about selection bias outlined above, Lee Another systematic review by Gøtzsche et al.
and colleagues highlight the danger inherent in found that results from subgroup analyses were
this approach of its susceptibility to manipula- stressed in four trials where the main analysis
tion and data dredging.106 had not favoured the new drug.115 Freedman and
colleagues recommended that the goals of analy-
Comprehensive cohort follow-up study sis should be to test the primary question posed
design by the trial over the total set of subjects, with the
conduct of secondary analyses to identify ques-
One criticism which is made of RCTs is that tions with sufficient scientific basis to be tested
they often lack generalisability. This may arise as primary questions in future trials.116
because only a small proportion of the available
population may be successfully randomised into
a trial. An alternative approach is the ‘compre-
Significance levels
hensive cohort follow-up approach’, described Criticisms of the use of significance levels iden-
by Olschewski et al. and Schmoor et al., in tified by this review have been focussed on the
which everybody eligible for entry into a trial interrelationship between statistical significance
should be followed-up in a similar manner, and clinical significance. As such, it is as much
irrespective of whether they accepted random- a design issue as an issue for analysis. Analyses
isation.111,112 The treatment difference between often focus on statistical significance rather than
the randomised treatments will be estimated in clinical significance,117,118 although they may
those accepting randomisation (the approach not necessarily be the same.104 In a systematic
Factors Limiting Number, Progress and Quality of RCTs 47
compared to those with non-significant results, are those suggested by Drummond and Davies129
and those with external funding were also more endorsed by others, as follows:
likely to be published.103,125 However, trials
sponsored by pharmaceutical companies are less (a) Treatments to be compared should be those
likely to be published, often because the spon- likely to be considered by future decision
sors had control of the data. The main reason makers.130,131
identified by Dickersin and Min for trials being (b) Because resource use may be more or less
unpublished was that the authors did not write variable than patient outcomes, sample size
them up: no trial remained unpublished because calculations should consider economically
of rejection by a journal, although this does important effect sizes, notably in costs and
occur.125 Another survey by Davidson of trials patient utilities, as well as clinically import-
published in major journals found that the ant effect sizes.131–133
majority of trials supported by pharmaceutical (c) The estimation of treatment costs should
companies reported results that favoured the exclude all trial costs, though this is rarely
company’s product.126 By comparison, fewer easy.131,134
trials which did not acknowledge commercial (d) The reporting of treatment costs should
support reported results in favour of a new separate the quantities of resources used
therapy over traditional therapy. from the prices paid.135,136
Many studies including RCTs are only pub- (e) Patient outcome measures should be rigor-
lished in abstract form and therefore may be ously chosen and should generally include
difficult to identify. A systematic review by clinical measures, disease-specific meas-
Scherer et al. showed that about 35% of RCTs ures of quality of life, and generic meas-
initially presented as abstracts were never pub- ures of patient utility.137,138
lished in full, and this was more likely if they (f) If the economic value of different technol-
were small and had non-significant results.127 As ogies is to be compared across trials, find-
a result of publication bias, world-wide pros- ings about resources and outcomes should
pective registration of randomised trials has be reported in standard form, for example
been called for, since retrospective identification net direct costs (that is, direct costs minus
of unpublished trials is unsatisfactory.128 direct financial benefits) and net gain in
‘quality-adjusted life years’.137,139
Multiple publication of the same trial (g) Extrapolating the findings of the trial to
other settings should generally be based on
Failure to publish trial results causes problems economic modelling that takes account of
but, paradoxically, so too does multiple publica- differences between trial centres and uses
tion of the same trial. Again, this is more data from other centres, for example on
common for trials showing statistically signifi- treatment costs.23,136,137
cant results.103 Multiple reports can cause confu-
sion, since it may be unclear that they all relate In addition, we recommended that clinicians
to the same trial and there may be discrepancies considering the inclusion of an economic
in the results and conclusions of different evaluation should work closely with health
reports. economists.131,133,135,139
Trial costs
Costs
The absence of resources to undertake trials is
Consideration of costs related to two issues: the certain to reduce their number, and shortage of
need for economic evaluation as part of a health resources is likely to limit their progress or
technology assessment; and the costs of under- general quality, or both.23 Additional costs to
taking the research itself. The papers reviewed patients of taking part in trials (particularly of
for this section were based on methodological or consideration in the US medical system) can act
theoretical papers, reviews, and case studies or as a deterrent to recruitment,60,140,141 particularly
surveys. if the additional treatment costs will be passed
on to the patient,22,142 or the insurer who may
refuse to reimburse such costs.140,142
Economic evaluation Many authors advocate that patients should be
Few randomised trials have reported an econ- fully reimbursed for all their resulting costs and
omic evaluation, possibly because of difficulties some even propose fees for participants.60,143
in conducting such evaluations and the lack of The costs could be either covered by research
generalisability from one healthcare context to funds60,141 or national initiatives.22,142 It has been
another. The recommendations from this section suggested that unproven treatments should be
Factors Limiting Number, Progress and Quality of RCTs 49
funded only for patients in peer-reviewed most general need is for prospective studies to
trials.23 identify features that cause trials to be success-
The cost of conducting the research itself was ful or unsuccessful. Which trials fail to start or
also considered in a number of American studies. to make progress? It is possible that it is the
For example, several studies (including those by weaker trials that fail, through weak designs or
Silagy et al.,144 BjornsonBenson et al.145 and lack of clinical support for the question. If so,
King et al.146), compared the effectiveness or they need to be strengthened or stopped. How-
cost effectiveness of different methods of ever, many specific issues also require more
recruiting patients, without identifying consis- research to be carried out.
tent results. Little work has been carried out on
other aspects of research conduct, for example,
the costs of tests and investigations, clinician Design
time or quality assurance processes.
Research is required to identify the most appro-
priate and cost-effective designs of trials: for
IMPLICATIONS FOR THE HEALTH SERVICE example, when would it be most appropriate to
use the comprehensive cohort design? It would
appear sensible to measure core outcome vari-
The main implication of this review – and one
ables in trials within specific clinical areas, such
which will be difficult to implement – is to try
that comparisons could be made between trials,
and change the culture of the health service. The
but further research is needed to identify the
value of evidence provided by RCTs is increas-
most valuable sets of measures. Further work is
ingly understood. However, the need to carry
needed to distinguish the best types of analysis,
out such research within the health service is
in particular for the use of repeated measures
less widely accepted: it is considered an addi-
and subgroup analyses.
tional function over and above clinical and
administrative efforts. There needs to be recog-
nition that for results of trials to be of most
relevance, they should be carried out under
Structure and Administration
usual clinical conditions within the health service.
The structure and organisational features of a
We therefore believe that research should be
trial may be the key to its success or failure.
encouraged as a core activity of clinicians.
However, little is known of the impact of such
The 1998 changes in funding arrangements
organisational structures such as: the setting
within the UK National Health Service, imple-
of the research; the role of individuals within
menting the recommendations of the Research
the research team, for example the principal
and Development Task Force (the ‘Culyer
investigator, the clinical collaborators and the
Report’147), should go some way towards chang-
research co-ordinator; the appropriate levels of
ing the culture, and should overcome some of
quality control; the measurement of compliance;
the previous financial barriers to trials. These
the function of Steering Committees and Data
changes ensure that both treatment costs and
Monitoring Committees. In addition, the ideal
‘service support costs’ of peer-reviewed trials
structure for small trials and large trials may
are met by the NHS. We believe that the new
differ significantly, as may the structures of
arrangements may also help in eliminating
commercially sponsored trials. A particularly
research that is not worth supporting in the
under-researched area is that of the costs of
UK.
conducting trials.
To ensure that new technologies are not
adopted into widespread use before they are
fully evaluated, we suggest that guidelines
should be developed for the evaluation of new
Recruitment
technologies in the UK. The health service
Recruitment was identified by the review as
should be actively involved in developing these
probably the greatest problem affecting all trials.
guidelines and in setting the wider UK research
Further research is needed to identify the types
agenda, not only for the NHS Research and
of research designs which least interfere with
Development Programme.
clinical practice and cause least disruption to
patients. We need a better understanding of
clinician and patient motivation for taking
SUGGESTIONS FOR FURTHER RESEARCH part in trials, so that we can identify the best
methods to approach potential participants in
In common with much research, this review has order to obtain their fully informed participa-
identified more questions than answers. The tion, whilst minimising anxiety. The role of
50 The Advanced Handbook of Methods in Evidence Based Healthcare
rewards for clinicians and patients requires fur- 9. Silagy CA, Jewell D. Review of 39 years of
ther investigation. randomized controlled trials in the British
Journal of General Practice. Br. J. Gen. Pract.
1994; 44(385): 359–63.
Reporting 10. Rossetti L, Marchetti I, Orzalesi N, Scorpiglione
N, Torri V, Liberati A. Randomized clinical
The introduction of the CONSORT guidelines trials on medical treatment of glaucoma. Are
for reporting trials was believed to herald an they appropriate to guide clinical practice? Arch.
improvement in the standards of reporting, but Ophthalmol. 1993; 111: 96–103.
further work is needed to find out if this was the 11. Balas EA, Austin SM, Ewigman BG, Brown
outcome. There is a particular need to invest- GD, Mitchell JA. Methods of randomized con-
igate the reporting of commercially sponsored trolled clinical trials in health services research.
research. Medical Care 1995; 33: 687–99.
12. Taves RR. Minimization: a new method of
assigning patients to treatment and control
CONCLUSIONS groups. Clin. Pharmacol. Ther. 1974; 15:
443–53.
This review was designed to identify factors that 13. Altman DG, Dore CJ. Randomisation and base-
limit the number, progress and quality of RCTs. line comparisons in clinical trials. Lancet 1990;
Indeed, we have identified so many factors that 335: 149–53.
our admiration for the trials that are successful is 14. Schulz KF, Chalmers I, Hayes RJ, Altman DG.
even greater. These successful trials, and the Empirical evidence of bias. Dimensions of
evidence of this chapter, provide an excellent methodological quality associated with estimates
yardstick for all trialists, both new and existing. of treatment effects in controlled trials. JAMA
We exhort them all to continue to enhance the 1995; 273: 408–12.
number and quality of randomised trials. Only 15. Chalmers TC, Celano P, Sacks HS, Smith H. Bias
thus can we all improve the quality of the evi- in treatment assignment in controlled clinical
dence base for healthcare, and thus the quality of trials. N. Engl. J. Med. 1983; 309: 1358–61.
healthcare itself. 16. Peto R, Collins R, Gray R. Large-scale random-
ized evidence: large, simple trials and overviews
of trials. J. Clin. Epidemiol. 1995; 48: 23–40.
17. Fayers PM, Machin D. Sample size: how many
REFERENCES patients are necessary? Br. J. Cancer 1995; 72:
1–9.
1. Pocock SJ. Clinical Trials – A Practical 18. Edwards SJL, Lilford RJ, Braunholtz D, Jackson
Approach. John Wiley: Chichester, 1983. J. Why ‘underpowered’ trials are not necessarily
2. Yusuf S, Held P, Teo KK, Toretsky ER. Selec- unethical. Lancet 1997; 350: 804–7.
tion of patients for randomized controlled trials: 19. Ashery RS, McAuliffe WE. Implementation
implications of wide or narrow eligibility cri- issues and techniques in randomized trials of
teria. Statist. Med. 1990; 9: 73–83. outpatient psychosocial treatments for drug
3. Collins R, Peto R, Gray R, Parish S. Large-scale abusers: recruitment of subjects. Am. J. Drug
randomized evidence: trials and overviews. Alcohol Abuse 1992; 18: 305–29.
Oxford Textbook of Medicine, 3rd edn, 1996; 1: 20. Easterbrook PJ, Matthews DR. Fate of research
21–32. studies. J. R. Soc. Med. 1992; 85: 71–6.
4. Nixon DW. Special aspects of cancer prevention 21. Benson AB, Pregler JP, Bean JA, Rademaker
trials. Cancer 1994; 74(9 Suppl.): 2683–6. AW, Eshler B, Anderson, K. Oncologists’ reluc-
5. Blair PS, Silman AJ. Can clinical trials in rheu- tance to accrue patients onto clinical trials: an
matology be improved? Curr. Opin. Rheumatol. Illinois Cancer Center study. J. Clin. Oncol.
1991; 3(2): 272–9. 1991; 9: 2067–75.
6. Coombs DW, Dimick A, Bronstein JM, Potts 22. Foley JF, Moertel CG. Improving accrual into
LH, Bowens B. Conceptual and methodologic cancer clinical trials. J. Cancer Education 1991;
problems in the evaluation of a new burn treat- 6: 165–73.
ment modality. J. Burn Care Rehabilitation 23. Smyth JF, Mossman J, Hall R, Hepburn S,
1993; 14: 568–71. Pinkerton R, Richards M, Thatcher N, Box J.
7. Fleming TR, DeMets DL. Surrogate end points Conducting clinical research in the new NHS:
in clinical trials: are we being misled? Ann. the model of cancer. United Kingdom Co-
Intern. Med. 1996; 125: 605–13. ordinating Committee on Cancer Research. Br.
8. Talley NJ. A critique of therapeutic trials in Med. J. 1994; 309: 457–61.
Helicobacter pylori-positive functional dys- 24. Aaronson NK, VisserPol E, Leenhouts GHMW
pepsia. Gastroenterology 1994; 106: 1174–83. et al. Telephone-based nursing intervention
Factors Limiting Number, Progress and Quality of RCTs 51
improves the effectiveness of the informed con- 36. Bowen J, Hirsch S. Recruitment rates and fac-
sent process in cancer clinical trials. J. Clin. tors affecting recruitment for a clinical trial of a
Oncol. 1996; 14: 984–96. putative anti-psychotic agent in the treatment of
25. Taylor KM, Kelner M. Interpreting physician acute schizophrenia. Human Psychopharmacol.
participation in randomized clinical trials: the 1992; 7: 337–41.
physician orientation profile. J. Health Soc. 37. Autret E, Dutertre JP, Barbier P, Jonville AP,
Behav. 1987; 28: 389–400. Pierre F, Berger C. Parental opinions about
26. Taylor KM, Feldstein ML, Skeel RT, Pandya biomedical research in children in Tours,
KJ, Ng P, Carbone PP. Fundamental dilemmas France. Dev. Pharmacol. Ther. 1993; 20:
of the randomized clinical trial process: results 64–71.
of a survey of the 1,737 Eastern Cooperative 38. Harth SC, Thong YH. Parental perceptions and
Oncology Group investigators. J. Clin. Oncol. attitudes about informed consent in clinical
1994; 12(9): 1796–805. research involving children. Soc. Sci. Med.
27. Winn RJ, Miransky J, Kerner JF, Kennelly L, 1995; 40: 1573–7.
Michaelson RA, Sturgeon, SR. An evaluation of 39. Schwartz CE, Fox BH. Who says yes? Identify-
physician determinants in the referral of patients ing selection biases in a psychosocial inter-
for cancer clinical trials in the community set- vention study of multiple sclerosis. Soc. Sci.
ting. Prog. Clin. Biol. Res. 1984; 156: 63–73. Med. 1995; 40: 359–70.
28. Tognoni G, Alli C, Avanzini F, Bettelli G, 40. Newburg SM, Holland AE, Pearce LA. Motiva-
Colombo F, Corso R, Marchioli R, Zussino A. tion of subjects to participate in a research trial.
Randomized clinical trials in general practice: Applied Nursing Research 1992; 5: 89–93.
lessons from a failure. Br. Med. J. 1991; 303: 41. Plaisier PW, Berger MY, van der Hul RL, Nijs
969–71. HG, den Toom R, Terpstra, OT, Bruining HA.
29. Penn ZJ, Steer PJ. Reasons for declining partici- Unexpected difficulties in randomizing patients
pation in a prospective randomized trial to deter- in a surgical trial: a prospective study comparing
mine the optimum mode of delivery of the extracorporeal shock wave lithotripsy with open
preterm breech. Controlled Clinical Trials 1990; cholecystectomy. World J. Surg. 1994; 18:
11: 226–31. 769–72.
30. Shea S, Bigger JT, Campion J, Fleiss JL, 42. Bevan EG, Chee LC, McInnes GT. Patients’
Rolnitzky LM, Schron E, Gorkin L, Handshaw attitudes to participation in clinical trials. Br. J.
K, Kinney MR, Branyon M. Enrolment in clin- Clin. Pharmacol. 1992; 34: 156P–7P.
ical trials: institutional factors affecting enroll- 43. Langley GR, Sutherland HJ, Wong S, Minkin S,
ment in the cardiac arrhythmia suppression trial Llewellyn-Thomas HA, Till JE. Why are (or are
(CAST). Controlled Clinical Trials 1992; 13: not) patients given the option to enter clinical
466–86. trials? Controlled Clinical Trials 1987; 8:
31. Henzlova MJ, Blackburn BH, Bradley EJ, 49–59.
Rogers WJ. Patient perception of a long-term 44. Slevin M, Mossman J, Bowling A, Leonard R,
clinical trial: Experience using a close-out ques- Steward W, Harper P, McIllmurray M, Thatcher
tionnaire in the Studies of Left Ventricular Dys- N. Volunteers or victims: patients’ views of
function (SOLVD) trial. Controlled Clinical randomized cancer clinical trials. Br. J. Cancer
Trials 1994; 15: 284–93. 1995; 71: 1270–4.
32. Siminoff LA, Fetting JH, Abeloff MD. Doctor- 45. Bergmann JF, Chassany O, Gandiol J, Deblois
patient communication about breast cancer P, Kanis JA, Segrestaa JM, Caulin C, Dahan R.
adjuvant therapy. J. Clin. Oncol. 1989; 7: A randomized clinical trial of the effect of
1192–200. informed consent on the analgesic activity of
33. Schaeffer MH, Krantz DS, Wichman A, Masur placebo and naproxen in cancer pain. Clinical
H, Reed E, Vinicky JK. The impact of disease Trials and Meta-Analysis 1994; 29: 41–7.
severity on the informed consent process in 46. Dahan R, Caulin C, Figea L, Kanis JA, Caulin F,
clinical research. Am. J. Med. 1996; 100: Segrestaa JM. Does informed consent influence
261–8. therapeutic outcome? A clinical trial of the hyp-
34. Wadland WC, Hughes JR, SeckerWalker RH, notic activity of placebo in patients admitted to
Bronson DL Fenwick J. Recruitment in a pri- hospital. Br. Med. J. 1986; 293: 363–4.
mary care trial on smoking cessation. Family 47. Antman K, Amato D, Wood W, Carson J, Suit
Medicine 1990; 22: 201–4. H, Proppe K, Carey R, Greenberger J, Wilson R,
35. Atwood JR, Haase J, Rees-McGee S et al. Frei E. Selection bias in clinical trials. J. Clin.
Reasons related to adherence in community- Oncol. 1985; 3: 1142–7.
based field studies. Patient Education & Coun- 48. Stone JM, Page FJ, Laidlaw CR, Cooper I.
selling 1992; 19: 251–9. Selection of patients for randomized trials: a
52 The Advanced Handbook of Methods in Evidence Based Healthcare
study based on the MACOP-B vs CHOP in NHL research. West. J. Nursing Res. 1989; 11:
study. Aust. N. Z. J. Med. 1994; 24: 536–40. 418–30.
49. Lynoe N, Sandlund M, Dahlqvist G, Jacobsson 62. Neaton JD, Grimm RH, Jr., Cutler JA. Recruit-
L. Informed consent: study of quality of infor- ment of participants for the multiple risk factor
mation given to participants in a clinical trial. Br. intervention trial (MRFIT). Controlled Clinical
Med. J. 1991; 303: 610–13. Trials 1987; 8(4 Suppl.): 41S–53S.
50. Llewellyn-Thomas HA, McGreal MJ, Thiel EC, 63. Pocock SJ. When to stop a clinical trial. Br.
Fine S, Erlichman C. Patients’ willingness to Med. J. 1992; 305: 235–40.
enter clinical trials: measuring the association 64. Fleming TR, DeMets DL. Monitoring of clinical
with perceived benefit and preference for deci- trials: issues and recommendations. Controlled
sion participation. Soc. Sci. Med. 1991; 32: Clinical Trials 1993; 14: 183–97.
35–42. 65. Anonymous. The early termination of clinical
51. Roberson NL. Clinical trial participation. View- trials: causes, consequences, and control. With
points from racial/ethnic groups. Cancer 1994; special reference to trials in the field of arrhyth-
74(9 Suppl.): 2687–91. mias and sudden death. Task Force of the Work-
52. Jensen AB, Madsen B, Andersen P, Rose C. ing Group on Arrhythmias of the European
Information for cancer patients entering a clin- Society of Cardiology. Eur. Heart J. 1994; 15:
ical trial – an evaluation of an information 721–38.
strategy. Eur. J. Cancer 1993; 29A: 2235–8. 66. Buyse M. Interim analyses, stopping rules and
53. Sutherland HJ, Carlin K, Harper W, Martin LJ, data monitoring in clinical trials in Europe.
Greenberg CV, Till JE, Boyd NF. A study of Statist. Med. 1993; 12: 509–20.
diet and breast cancer prevention in Canada: 67. Urquhart J. Role of patient compliance in clinical
why healthy women participate in controlled pharmacokinetics. A review of recent research.
trials. Cancer Causes Control 1993; 4: 521–8. Clin. Pharmacokinet. 1994; 27: 202–15.
54. Miller C, Searight HR, Grable D, Schwartz R, 68. Besch CL. Compliance in clinical trials. AIDS
Sowell C, Barbarash RA. Comprehension and 1995; 9: 1–10.
recall of the informational content of the 69. Klein MC, Kaczorowski J, Robbins JM, Gauthier
informed consent document: An evaluation of RJ, Jorgensen SH, Joshi AK. Physicians’ beliefs
168 patients in a controlled clinical trial. J. Clin. and behaviour during a radomized controlled
Res. Drug Dev. 1994; 8: 237–48. trial of episiotomy: Consequences for women
55. Gallo C, Perrone F, De Placido S, Giusti C. in their care. Can. Med. Assoc. J. 1995; 153:
Informed versus randomized consent to clinical 769–79.
trials. Lancet 1995; 346: 1060–4. 70. Rudd P, Ahmed S, Zachary V, Barton C,
56. Dunbar J, Cleary PA, Siebert C, Baker L, Brink Bonduelle D. Improved compliance measures:
S, Nathan DM. Implementation of a multi- applications in an ambulatory hypertensive drug
component process to obtain informed consent trial. Clin. Pharmacol. Ther. 1990; 48: 676–85.
in the diabetes control and complications trial. 71. Urquhart J. Ascertaining how much compliance
Controlled Clinical Trials 1989; 10: 83–96. is enough with outpatient antibiotic regimens.
57. Dickinson CJ. Clinical research in the NHS Postgrad. Med. J. 1992; 68(Suppl. 3): S49–58.
today. J. R. Coll. Phys. (Lond.) 1994; 28: 72. Melnikow J, Kiefe C. Patient compliance and
460–3. medical research: Issues in methodology. J. Gen.
58. Chang RW, Falconer J, Stulberg SD, Arnold Intern. Med. 1994; 9: 96–105.
WJ, Dyer AR. Prerandomization: an alternative 73. Pledger GW. Compliance in clinical trials: im-
to classic randomization. The effects on recruit- pact on design, analysis and interpretation. Epi-
ment in a controlled trial of arthroscopy for lepsy Research 1988; 1(Supplement): 125–33.
osteoarthrosis of the knee. J. Bone Joint Surg. 74. Vander Stichele R. Measurement of patient
1990; 72: 1451–5. compliance and the interpretation of randomized
59. Hunninghake DB, Darby CA, Probstfield JL. clinical trials. Eur. J. Clin. Pharmacol. 1991; 41:
Recruitment experience in clinical trials: lit- 27–35.
erature summary and annotated bibliography. 75. Schechtman KB, Gordon ME. A comprehensive
Controlled Clinical Trials 1987; 8(4 Suppl.): algorithm for determining whether a run-in strat-
6S–30S. egy will be a cost-effective design modification
60. Swanson GM, Ward AJ. Recruiting minorities in a randomized clinical trial. Statist. Med. 1993;
into clinical trials: toward a participant-friendly 12: 111–28.
system. J. Nat. Cancer Inst. 1995; 87: 76. Stephenson BJ, Rowe BH, Haynes RB,
1747–59. Macharia WM, Leon G. Is this patient taking
61. Diekmann JM, Smith JM. Strategies for access- the treatment as prescribed? JAMA 1993; 269:
ment and recruitment of subjects for nursing 2779–81.
Factors Limiting Number, Progress and Quality of RCTs 53
77. Lawrence W. Some problems with clinical trials. 91. Poy E. Objectives of QC systems and QA func-
(James Ewing lecture) Arch. Surg. 1991; 126: tion in clinical research. Quality Assurance
370–8. 1993; 2: 326–31.
78. Richmond RL, Anderson P. Research in general 92. Wood DJ. Quality assurance in phase II and III
practice for smokers and excessive drinkers in studies. Quality Assurance 1993; 2: 353–8.
Australia and the UK. II: representativeness of 93. van Gijn J. Who is in Charge? Investigator and
the results. Addiction 1994; 89: 41–7. sponsor alliances. Cerebrovasc. Dis. 1995; 5:
79. Haynes RB, McKibbon KA, Kanani R. System- 22–6.
atic review of randomized controlled trials of the 94. Julian D. Trials and tribulations. Cardiovasc.
effects of patient adherence and outcomes of Res. 1994; 28: 598–603.
interventions to assist patients to follow pre- 95. Warlow C. How to do it: organise a multicentre
scriptions for medications. Collaboration on trial. Br. Med. J. 1990; 300: 180–3.
Effective Professional Practice Module of The 96. Hogg RJ. Trials and tribulations of multicenter
Cochrane Database on Systematic Reviews studies. Lessons learned from the experiences
1996; 1. of the Southwest Pediatric Nephrology Study
80. Liberati A, Himel HN, Chalmers TC. A quality Group (SPNSG). Pediatr. Nephrol. 1991; 5:
assessment of randomized control trials of pri- 348–51.
mary treatment of breast cancer. J. Clin. Oncol. 97. McCandlish R, Renfrew M. Trial and tribulation
1986; 4: 942–51. . . . treatments for inverted nipples . . . difficul-
81. Kleijnen J, ter Riet G, Knipschild P. Acupunc- ties involved in research. Nursing Times 1991;
ture and asthma: a review of controlled trials. 87: 40–1.
Thorax 1991; 46: 799–802. 98. Bogousslavsky J. The jungle of acute stroke
82. Solomon MJ, Laxamana A, Devore L, McLeod trials, trialists and sponsors. Cerebrovasc. Dis.
RS. Randomized controlled trials in surgery. 1995; 5: 1–2.
Surgery 1994; 115: 707–12. 99. Rockhold FW, Enas GG. Data monitoring and
83. Sonis J, Joines J. The quality of clinical trials interim analyses in the pharmaceutical industry:
published in The Journal of Family Practice,
ethical and logistical considerations. Statist.
1974–1991. J. Family Practice 1994; 39:
Med. 1993; 12: 471–9.
225–35.
100. Parmar MK, Machin D. Monitoring clinical
84. Lionetto R, Pugliese V, Bruzzi P, Rosso R. No
trials: experience of, and proposals under con-
standard treatment is available for advanced
sideration by, the Cancer Therapy Committee of
pancreatic cancer. Eur. J. Cancer 1995; 31A:
the British Medical Research Council. Statist.
882–7.
Med. 1993; 12: 497–504.
85. Schulz KF, Grimes DA, Altman DG, Hayes RJ.
101. Green SJ, Fleming TR, O’Fallon JR. Policies
Blinding and exclusions after allocation in ran-
domized controlled trials: survey of published for study monitoring and interim reporting of
parallel group trials in obstetrics and gynaecol- results. J. Clin. Oncol. 1987; 5: 1477–84.
ogy. Br. Med. J. 1996; 312: 742–4. 102. Rochon PA, Gurwitz JH, Simms RW, Fortin PR,
86. Stewart LA, Clarke MJ. Practical methodology Felson DT, Minaker KL, Chalmers TC. A study
of meta-analyses (overviews) using updated of manufacturer-supported trials of nonsteroidal
individual patient data. Cochrane Working anti-inflammatory drugs in the treatment of
Group. Statist. Med. 1995; 14: 2057–79. arthritis. Arch. Intern. Med. 1994; 154: 157–63.
87. Fisher S, Greenberg RP. How sound is the 103. Easterbrook PJ, Berlin JA, Gopalan R,
double-blind design for evaluating psychotropic Matthews DR. Publication bias in clinical
drugs? J. Nervous Mental Dis. 1993; 181: research. Lancet 1991; 337: 867–72.
345–50. 104. Bigby M, Gadenne AS. Understanding and
88. Kirsch I, Rosadino MJ. Do double-blind studies evaluating clinical trials. J. Am. Acad. Dermatol.
with informed consent yield externally valid 1996; 34: 555–94.
results? An empirical test. Psychopharmacology 105. Peduzzi P, Detre K, Wittes J, Holford T. Intent-
1993; 110: 437–42. to-treat analysis and the problem of crossovers.
89. Carroll KM, Rounsaville BJ, Nich C. Blind An example from the Veterans Administra-
man’s bluff: effectiveness and significance of tion coronary bypass surgery study. J. Thorac.
psychotherapy and pharmacotherapy blinding Cardiovasc. Surg. 1991; 101: 481–7.
procedures in a clinical trial. J. Consult. Clin. 106. Lee YJ, Ellenberg JH, Hirtz G, Nelson KB.
Psychol. 1994; 62: 276–80. Analysis of clinical trials by treatment actually
90. Noseworthy JH, Ebers GC, Vandervoort MK, received: is it really an option? Statist. Med.
Farquhar RE, Yetisir E, Roberts R. The impact 1991; 10: 1595–605.
of blinding on the results of a randomized, 107. Nicolucci A, Grilli R, Alexanian AA, Apolone
placebo-controlled multiple sclerosis clinical G, Torri V, Liberati A. Quality, evolution, and
trial. Neurology 1994; 44: 16–20. clinical implications of randomized, controlled
54 The Advanced Handbook of Methods in Evidence Based Healthcare
trials on the treatment of lung cancer. A lost 121. Ravnskov U. Cholesterol lowering trials in
opportunity for meta-analysis. JAMA 1989; 262: coronary heart disease: frequency of citation and
2101–7. outcome. Br. Med. J. 1992; 305: 15–19.
108. Koes BW, Bouter LM, van der Heijden GJ. 122. Begg C, Cho M, Eastwood S, Horton R, Moher
Methodological quality of randomized clinical D, Olkin I, Pitkin RX, Rennie D, Schulz KF,
trials on treatment efficacy in low back pain. Simel D, Stroup DF. Improving the quality of
Spine 1995; 20: 228–35. reporting of randomized controlled trials: the
109. Rabeneck L, Viscoli CM, Horwitz RI. Problems CONSORT statement. JAMA 1996; 276: 637–9.
in the conduct and analysis of randomized clin- 123. Altman DG. Better reporting of randomized
ical trials: are we getting the right answers to the controlled trials: the CONSORT statement. Br.
wrong questions? Arch. Intern. Med. 1992; 152: Med. J. 1996; 313: 570–1.
507–12. 124. Chalmers TC, Frank CS, Reitman D. Minimiz-
110. Sheiner LB, Rubin DB. Intention-to-treat analysis ing the three stages of publication bias. JAMA
and the goals of clinical trials. Clin. Pharmacol. 1990; 263: 1392–5.
Ther. 1995; 57: 6–15. 125. Dickersin K, Min YI. Publication bias: the prob-
111. Olschewski M, Schumacher M, Davis KB. lem that won’t go away. Ann. N. Y. Acad. Sci.
Analysis of randomized and nonrandomized 1993; 703: 135–46.
patients in clinical trials using the comprehen- 126. Davidson RA. Source of funding and outcome
sive cohort follow-up study design. Controlled of clinical trials. J. Gen. Intern. Med. 1986; 1:
Clinical Trials 1992; 13: 226–39. 155–8.
112. Schmoor C, Olschewski M, Schumacher M. 127. Scherer RW, Dickersin K, Langenberg P. Full
Randomized and non-randomized patients in publication of results initially presented in
clinical trials: experiences with comprehensive abstract: a meta-analysis. JAMA 1994; 272:
cohort studies. Statist. Med. 1996; 15: 263–71. 158–62.
113. Silverman WA, Altman DG. Patients’ prefer- 128. Hetherington J, Dickersin K, Chalmers I,
ences and randomized. Lancet 1996; 347: Meinert CL. Retrospective and prospective
171–4. identification of unpublished controlled trials:
114. Pocock SJ, Hughes MD, Lee RJ. Statistical lessons from a survey of obstetricians and pedia-
problems in the reporting of clinical trials. A tricians. Pediatrics 1989; 84: 374–80.
survey of three medical journals. N. Engl. J. 129. Drummond MF, Davies L. Economic analysis
Med. 1985; 317: 426–32. alongside clinical trials: revisiting the methodo-
115. Gøtzsche PC. Methodology and overt and hid- logical issues. Int. J. Technol. Assess. Health
den bias in reports of 196 double-blind trials of Care 1991; 7: 561–73.
nonsteroidal antiinflammatory drugs in rheuma- 130. Hansson L, Hedner T, Dahlof B. Prospective
toid arthritis. Controlled Clinical Trials 1989; randomized open blinded end-point (PROBE)
10: 31–56. study. A novel design for intervention trials.
116. Freedman LS, Simon R, Foulkes MA, Friedman Prospective Randomized Open Blinded End-
L, Geller NL, Gordon DJ, Mowery R. Inclusion Point. Blood Pressure 1992; 1: 113–19.
of women and minorities in clinical trials and 131. Schwartz JS. Prostate Cancer Intervention Versus
the NIH revitalization act of 1993 – the per- Observation Trial: economic analysis in study
spective of NIH clinical trialists. Controlled design and conditions of uncertainty. Mono-
Clinical Trials 1995; 16: 277–85. graphs – Nat. Cancer Inst. 1995; 19: 73–5.
117. Klein DF. Improvement of phase III psycho- 132. Drummond M, O’Brien B. Clinical importance,
tropic drug trials by intensive phase II work. statistical significance and the assessment of
Neuropsychopharmacology 1991; 4: 251–8. economic and quality-of-life outcomes. Health
118. Lindgren BR, Wielinski CL, Finkelstein SM, Economics 1993; 2: 205–12.
Warwick WJ. Contrasting clinical and statistical 133. Torgerson DJ, Ryan M, Ratcliffe J. Economics
significance within the research setting. Pediatr. in sample size determination for clinical trials.
Pulmonol. 1993; 16: 336–40. Q. J. Med. 1995; 88: 517–21.
119. Marsoni S, Torri W, Taiana A et al. Critical 134. Oddone E, Weinberger M, Hurder A, Henderson
review of the quality and development of ran- W, Simel D. Measuring activities in clinical
domized clinical trials (RCTs) and their influ- trials using random work sampling: implications
ence on the treatment of advanced epithelial for cost-effectiveness analysis and measurement
ovarian cancer. Ann. Oncol. 1990; 1: 343–50. of the intervention. J. Clin. Epidemiol. 1995; 48:
120. Clarke M, Chalmers I. Discussion sections in 1011–18.
reports of controlled trials published in five 135. Bennett CL, Armitage JL, Buchner D, Gulati S.
general medical journals: islands in search of Economic analysis in phase III clinical cancer
continents. JAMA 1998; 280: 280–2. trials. Cancer Invest. 1994; 12: 336–42.
Factors Limiting Number, Progress and Quality of RCTs 55
136. Drummond M. Economic analysis alongside clin- Reimbursement problems. Cancer 1994; 74(9
ical trials: problems and potential. J. Rheumatol. Suppl.): 2662–5.
1995; 22: 1403–7. 143. Bigorra J, Banos JE. Weight of financial reward
137. Bonsel GJ, Rutten FF, Uyl-de Groot CA. in the decision by medical students and experi-
Economic evaluation alongside cancer trials: enced healthy volunteers to participate in clin-
methodological and practical aspects. Eur. J. ical trials. Eur. J. Clin. Pharmacol. 1990; 38:
Cancer 1993; 29A(Suppl. 7): S10–4. 443–6.
138. Tannock IF. New perspectives in combined 144. Silagy CA, Campion K, McNeil JJ, Worsam B,
radiotherapy and chemotherapy treatment. Lung Donnan GA, Tonkin AM. Comparison of
Cancer 1994; 10(Suppl. 1): S29–51. recruitment strategies for a large-scale clinical
139. Drummond MF. Economic analysis alongside trial in the elderly. J. Clin. Epidemiol. 1991; 44:
clinical trials: an introduction for clinical re- 1105–14.
searchers. London: Department of Health, 1994. 145. BjornsonBenson WM, Stibolt TB, Manske KA,
140. Tilley BC, Peterson EL, Kleerekoper M, Phillips Zavela KJ, Youtsey DJ, Buist AS. Monitor-
E, Nelson DA, Shorck, MA. Designing clinical ing recruitment effectiveness and cost in a clin-
trials of treatment for osteoporosis: recruitment ical trial. Controlled Clinical Trials 1993; 14:
and follow-up. Calcified Tissue International 52S–67S.
1990; 47: 327–31. 146. King AC, Harris RB, Haskell WL. Effect of
141. Welsh KA, Ballard E, Nash F, Raiford K, recruitment strategy on types of subjects entered
Harrell L. Issues affecting minority participa- into a primary prevention clinical trial. Ann.
tion in research studies of Alzheimer disease. Epidemiol. 1994; 4: 312–20.
Alzheimer Disease & Associated Disorders 147. Research and Development Task Force. Sup-
1994; 8(Suppl. 4): 38–48. porting Research and Development in the NHS.
142. Fleming ID. Barriers to clinical trials. Part I: London: HMSO, 1994.
4
Results of Clinical Trials and Systematic
Reviews: To Whom Do They Apply?
Generalisability (or applicability) involves Potentially relevant articles were obtained and a
applying the results from the study population to structured review was carried out by members of
the treatment target population, but as Collet and the multi-disciplinary research group, including
Boissel indicate this will be different to those a classification according to subject areas.
who actually get the treatment.1 Annotated bibliographies of relevant papers
We have carried out a qualitative, but system- were compiled for each subject area.
atic review of the literature on the applicability After reviewing the methods described in the
of results of RCTs and systematic reviews of literature and a consideration of their strengths
RCTs. The specific aims were to: and weaknesses, we developed a ‘five-step’
(a) define the methods that are currently sug- process based on a risk–benefit model proposed
gested for applying clinical trial results in by Lubsen and Tijssen for applying the results
clinical practice; of RCTs and systematic reviews to individual
(b) assess the strengths and weaknesses of patient management decisions.2
currently suggested approaches;
(c) develop an appropriate method based on a
consideration of benefit and risk to apply
clinical trial results to individual patients; EVIDENCE
and
(d) consider the implications for design and A total of 568 papers, published to the end of
reporting of clinical trials and the develop- 1996, were identified as being of potential rele-
ment of clinical guidelines. vance, and a further 1185 papers were excluded
as not relevant. After review, completion of the
cover sheet and discussion, 419 papers were
METHODS considered to be of direct relevance to the topic
and were grouped by topic area for the anno-
Our goal was to identify and review the lit- tated bibliography. The main methods for
erature about methods for generalising (or assessing generalisability described in the lit-
applying) the results from RCTs and systematic erature fall into five areas: (1) consideration of
reviews of RCTs. We used a combination of patient selection and inclusion/exclusion cri-
searching electronic databases, hand-searching teria; (2) comparison of trial subjects with those
key medical journals and bibliographic searches not in the trial; (3) subgroup analysis of treat-
of relevant papers to identify and retrieve papers ment effect; (4) risk-based individualisation of
that discussed methods for applying the results benefits and harms; and (5) a mix of other ap-
of clinical trials. An iterative approach was used proaches. Each of these approaches is described
to develop a search strategy for Medline based in more detail in the next section.
on keywords and MeSH headings in relevant The literature on this topic is diffuse and was
papers, and names of authors known to have difficult to identify. It is our view that the
published on the topic. The search was very predominance of papers concerned with the
broad, and included papers that were reports influences of patient eligibility, recruitment and
of RCTs or systematic reviews in which the participation in clinical trials represents an
applicability of the results was considered, dis- important misunderstanding of the relationships
cussion papers, and papers focussing on meth- which determine the magnitude of the benefits
odological or statistical issues relating to the and harms that can be expected with treatment.
application of results of RCTs. This preoccupation with issues of patient selec-
We hand-searched a number of key journals tion, inclusion and exclusion criteria and the
including: Controlled Clinical Trials; Statistics
(un)representativeness of trial participants has
in Medicine; Journal of the American Medical
Association; Biometrics; Journal of Clinical Epi- inhibited the development of methods for the
demiology; The Lancet; New England Journal of accurate application of trial results.
Medicine; Medical Decision Making; British
Medical Journal; Annals of Internal Medicine;
and International Journal of Technology Assess- KEY ISSUES: CURRENT METHODS
ment in Health Care. We also searched the
Cochrane Review Methodology Database in the
Cochrane Library for relevant references. The methods described in the literature concen-
Papers identified as being of potential rele- trated primarily on consideration of the charac-
vance from the title and/or abstract were down- teristics of the settings and populations in which
loaded from Medline into a Reference Manager trials are conducted, and to whom the average
database; some citations were entered manually. treatment effect might apply.
Applying the Results of Clinical Trials and Reviews 59
Consideration of Subject Selection outside of, the relevant studies and then adjust-
ing the estimated average treatment effect for
An examination of the representativeness of the any observed differences.35–43 Sophisticated ap-
subjects who ultimately participate in random- proaches include the comprehensive cohort
ised controlled trials and other evaluations of design and cross-design synthesis.
interventions, through consideration of the The comprehensive cohort design is a pros-
source population, the inclusion and exclusion pective cohort study with a randomised sub-
criteria and the effects of obtaining informed cohort.36–39 The idea is to compare the ran-
consent has received most attention in the lit- domised patients with those who were eligible,
erature.3–28 This line of reasoning is based on but who were not randomised, on both base-
sampling theory and the corresponding need for line characteristics and outcomes (the non-
a representative sample of patients.6,12,14 Such a randomised patients having been followed-up in
consideration of ‘external validity’ is important the same way as the trial participants after
when the aim is to estimate the prevalence of a receiving one of the interventions being com-
condition in a population, or to estimate certain pared). For example, a comprehensive cohort
characteristics such as levels of knowledge, of patients with coronary heart disease was
patterns of behaviour or the range in attitudes. assembled to compare surgical and medical
When this is extended to clinical trials, the management.36–38 The survival of patients
conclusion that must be drawn from this ap- receiving each of the treatment modalities was
proach is that the results are applicable only to compared both within the randomised and the
subjects who resemble those included in the non-randomised patients. However, the analysis
trial.6,12,13,29,30 of the non-randomised patients is plagued by the
However, the major purposes of inclusion and usual sources of bias associated with observa-
exclusion criteria are to improve study power
tional studies, including confounding. Also, an
(through inclusion of high-risk groups who are
more likely to have the outcomes of interest, to interaction between the treatment and one or
minimise losses of subjects through death from more prognostic factors (which may be dis-
other causes and to ensure good compliance to tributed unevenly between the randomised and
the intervention of interest) and to maximise non-randomised groups) may produce an appar-
safety of subjects in the trial (by minimising the ent difference or an apparent similarity in the
occurrences of adverse effects).31 However, results from the randomised and non-randomised
there may be subjects excluded from the study components of the study.39
who would experience an overall net benefit The underlying principle of cross design
from the intervention, and there would certainly synthesis is to capture the strengths, and to
be subjects included who (despite a statistically minimise the weaknesses, of the randomised and
significant overall treatment effect) have experi- non-randomised study designs.40,41 The process,
enced net harm.2,31,32 which combines the unbiased comparison of
This approach is further limited by the fact treatments from a randomised trial with the
that a representative sample of eligible subjects wider representativeness of a clinical database
is rarely achieved. Also, there are predictable (or large cohort of patients with the condition of
differences between the eligible subjects and interest), consists of four steps:
those who ultimately participate in the trial, and
between those who meet the eligibility criteria 1 Assess existing randomised studies for gen-
and those who would be candidates for the inter- eralisability across the full range of relevant
vention but are not eligible.15,17–19,21,28,30,33,34 patients.
Finally, the average treatment effect is an 2 Assess databases of (non-randomised) pa-
overall summary averaged over all subjects in tients for ‘unbalanced comparison groups’.
the trial. The argument has been that this then 3 Adjust the results of each randomised study
applies to ‘average’ subjects. But, how does one and each database analysis, compensating
define an ‘average subject’? How different does for biases as needed.
a subject need to be from those in the study 4 Synthesise the studies’ adjusted results
before one concludes that the results cannot be within, and across, the design categories.
applied?
The main limitation of this method is the arbi-
trary nature of the judgements and standardisa-
Comparison of Subjects Within and tions required to make the necessary adjust-
Outside of the Relevant Studies ments.41 Also, the aim is to produce a single
(unbiased) estimate of the average treatment
Several approaches were described in the lit- effect ignoring possible variability in treatment
erature for comparing subjects within, and effect and potential effect modifiers. As argued
60 The Advanced Handbook of Methods in Evidence Based Healthcare
later, an exploration of the variability in esti- fall into many subgroups, and the results from
mates of treatment effect and possible sources of individual subgroup analyses may conflict. For
this variability leads to a better understanding of example, how does one treat an elderly, African-
the underlying mechanisms through which the American, male patient if different treatment
treatment works. effects are suggested in each of the subgroups
Some authors have proposed that the results defined by age, ethnicity and gender?
of a randomised trial can be readily applied to
other groups who have similar characteristics to
those included in the trial(s).35 For example, Risk-Based Individualisation of Benefit
groups with a similar age range, gender mix, co- and Harm
morbidities and disease severity. This represents
a more formal approach to the consideration of A few authors have proposed methods for
applying the ‘average’ treatment effect to ‘aver- applying results to individual patients based on a
age’ subjects. consideration of the likely benefits and harms
from treatment.2,31,32 Based on the premise that it
is inappropriate to judge the applicability of a
Subgroup Analysis of Treatment Effects trial’s results by consideration of the subject
entry criteria, and that the average treatment
Another common approach has been to examine effect estimated in a trial is not applicable to all
the effect of the intervention in subgroups of subjects (neither within, nor outside of the
subjects defined by characteristics such as gen- study), the idea is to identify in which individ-
der, ethnicity, severity of disease, etc.25,44–52 The uals the treatment will do more good than harm.
idea is to compare the effect of the intervention A model proposed by Lubsen and Tijssen sug-
in various subgroups, e.g. men and women, the gests that patient benefit (as measured by the
elderly, or different ethnic groups. A common, prevention of an adverse event such as stroke or
but incorrect method is to perform a statistical death) increases with (untreated) risk of the
test of treatment effect within each subgroup. outcome, but harm or rates of adverse events
The interpretation of the results is that the caused by the treatment will remain relatively
intervention should be offered only to those fixed (Figure 1).2 As the absolute risk reduction
subgroups of patients in which a (statistically) is related to the untreated (or baseline) risk of
significant treatment effect was detected. In
the event, high-risk patients are more likely to
addition, some have suggested that treatment
experience net benefit. Also for some low-risk
should not be offered to those subgroups in
patients, the likely harms may outweigh the
which a significant treatment effect was not
expected benefits.
detected. Performing statistical tests for each
Glasziou and Irwig adapted this risk–benefit
individual subgroup can result in both false-
approach to identify which patients with non-
positive (a spurious statistically significant treat-
rheumatic atrial fibrillation may benefit from
ment effect in some groups) and false-negative
treatment with warfarin to prevent a stroke.31
(an apparent lack of a significant treatment
effect due to small numbers of subjects in some This approach has been further developed into a
subgroups) results.53 Such analyses have led to
the interesting observation that a treatment bene-
fit of intravenous atenolol in patients with sus-
pected myocardial infarction only occurs in
those born under the astrological birth sign
Scorpio,54 and aspirin had an effect in those
patients with suspected myocardial infarction
born under Capricorn, but not Gemini or
Libra.55
The correct approach is to perform a test for
treatment by subgroup interaction.6,56–60 This
results in a single statistical test, rather than
multiple tests within subgroups. If a significant
interaction is found, which indicates hetero-
geneity of treatment effect across groups, then
further work can be carried out to identify which
groups are different. Figure 1 Risk–benefit approach to applying
However, there is a difficulty in the inter- the evidence to indivdiual patients (based on
pretation of these analyses. Individual subjects the model of Lubsen and Tijssen2).
Applying the Results of Clinical Trials and Reviews 61
‘five-step’ process and is described in greater of warfarin to prevent stroke in patients with
detail in the next section. atrial fibrillation, the infrastructure for monitor-
ing blood levels has to be in place and acces-
sible to patients. This may not be the case for
Other Approaches patients living in rural areas.
Davis suggested that to determine the extent to
which the results of a clinical trial can be Can the results be applied in other
generalised, several pieces of information (from settings and to other patients?
various sources) are required.61 He used the ex-
ample of applying the results of the cholesterol- It is important to consider whether there are
lowering trials in the prevention of coronary pathophysiological differences which may alter
heart disease to illustrate the process. The ques- the biological mechanism through which the
tion was whether the results could be extra- treatment has its effect.64–66 For example, could
polated to women and to older men. The addi- the treatment effect be modified by gender, age
tional pieces of information required came from or ethnicity, or by spectrum of the disease or
laboratory studies in which the pathogenic timing of the intervention? A well-known ex-
mechanism of the cholesterol-lowering drugs ample where this is true is the treatment of
was examined to determine whether the mech- hypertension in blacks. The majority of trials of
anism is the same across age and genders, and lipid-lowering drugs were carried out in middle-
from animal, genetic, observational, clinical and aged men, and many have debated whether the
epidemiological studies as necessary. results apply to women and to older people.
Variations in the nature of the disease may
also change the effect of treatment. For example,
geographical patterns of antibiotic-resistant bac-
PROPOSED METHODS FOR APPLYING RESULTS
teria and drug-resistant malaria may alter the
OF TRIALS treatment’s effectiveness. Also, the treatment
may not be effective in severe disease where the
General Issues to Consider patient is very sick and the disease process is
now irreversible.
Firstly, applicability of the evidence to other The timing of the intervention should be con-
settings and to different patient groups should be sidered. Some treatments may be less effective
considered in its broadest context. Important if administration is delayed (e.g. thrombolytics
issues are whether the intervention can be repro- after myocardial infarction), and so it may not
duced in the setting of interest and whether there be appropriate to prescribe the treatment beyond
are biological factors that may make the treat- a certain time after the onset of symptoms or
ment less effective. disease.
These biological effect modifiers are probably
Can the intervention be reproduced in the not all that common, but they may be important
setting of interest? in a few instances.65 However, the assumption
should be made that they are unlikely, and the
This needs careful consideration. Generally, if burden of proof really rests on demonstrating
the intervention is a drug then it is readily
that they do modify the treatment effect, rather
transferable to other settings. However, the drug
than assuming that they may.
has to be licensed for use and it must be
affordable to the patients and to the healthcare A situation in which biological factors are
system. If the intervention is a surgical pro- important is the generalisation of results about a
cedure, then surgeons with appropriate training treatment for one form of disease to another
and support staff are required, together with the variant. For example, the Diabetes Control and
necessary equipment.62 The transferability of Complications Trial (DCCT) Research Group
complex interventions requires a more careful showed that intensive therapy and tight control
consideration. For example, each component of in patients with insulin-dependent diabetes sig-
a community health intervention programme nificantly reduced the occurrence of retinopathy,
must be taken into account and be applicable in neuropathy, proteinuria and microalbuminuria.67
the new setting. This includes the availability However, there was also a three-fold increase in
of appropriately trained personnel, the socio- the incidence of severe hypoglycaemic reactions
cultural acceptability of the programme content, and an overall weight gain among those receiv-
and so on.63 Also, the availability of the required ing intensive therapy. As patients with non-
infrastructure for any type of intervention should insulin-dependent diabetes mellitus (NIDDM)
be considered. For example, for the prescription tend to be middle-aged or elderly, they could not
62 The Advanced Handbook of Methods in Evidence Based Healthcare
tolerate this frequency of hypoglycaemic epi- The risk–benefit approach of Lubsen and
sodes and their hyperinsulinaemia may be aggra- Tijssen2 was adapted to identify which patients
vated by weight gain.68 Therefore intensive ther- with non-rheumatic atrial fibrillation may bene-
apy with insulin for NIDDM could aggravate, fit from treatment with warfarin to prevent a
rather than ameliorate, diabetic complications. stroke.31 A modified version of this approach is
recommended for use and is described in detail
in this section.
Using evidence in primary care Briefly, the principle underlying this approach
A common argument against applying the results is to apply the relative risk reduction due to the
of clinical trials to clinical practice is that trials intervention (which is assumed to be constant,
are typically carried out in tertiary care settings but this should be verified), to an individual’s
and so are not applicable in community hospi- predicted (baseline) risk if left untreated, to
tals, general practice and other primary care obtain an estimate of their expected absolute
settings.69,70 The belief is that because patients in risk reduction. This is then weighed up against
primary care are likely to have a lower probabil- the potential harms or adverse effects of the
ity of the outcome being prevented (i.e. have a intervention to determine whether the benefits
lower baseline risk), and milder disease, that the are likely to outweigh the risks.
biological response to treatment may be differ- This approach requires the consideration of
ent (and in particular that the treatment will be five questions, each of which should be ad-
less effective). This implies that the relative risk dressed when thinking about applicability (Table
reduction varies with baseline (or underlying) 3). The first three questions relate directly to the
risk of the event being prevented by the treat- overall effect of treatment – the expected out-
ment. That is, the relative treatment effect is comes (both benefits and harms) and the robust-
lower (or it may be higher) in patients with ness of the estimate of (relative) treatment effect
lower baseline risk. This is an assumption that (i.e. is it constant across a range of potential
needs to be checked. However, the absolute risk effect modifiers and levels of baseline risk?).
reduction (which is of more relevance to the These address the issue of transferability of the
individual) is likely to be smaller for lower average treatment effect. The last two questions
levels of baseline risk. Therefore an important cover aspects of individualising or particularis-
step, which is described later in this section, is a ing the treatment decision through estimating
consideration of the patient’s baseline risk and a the expected absolute risk reduction based on an
prediction of the absolute risk reduction. It may individual’s baseline risk and then taking into
be the case that for a larger proportion of account patient’s preferences in the weighing up
patients in primary care, the expected benefits of benefits and harms.
will not justify the potential harms and costs. When considering each of these five issues,
However, there are likely to be some patients often there will be insufficient data to address
who will benefit from treatment and therefore each of them in detail. However, it is still a
the treatment should not be excluded from pri- helpful exercise to go through these steps and to
mary care entirely. make explicit what additional information is
A consideration of relative versus absolute required to be able to apply the evidence about
risk reductions and their relationships with base- an intervention’s effectiveness in an accurate
line risk is discussed in greater detail later in this manner. With improvements in trial design, and
section. particularly in the analysis and reporting of
trials, it will become possible to use this
approach in an increasing number of situations.
Suggested Process for Applying
Evidence Transferability
As discussed in the previous section, traditional What are the beneficial and harmful effects
approaches for considering the applicability of
results of trials and systematic reviews are very of the intervention?
limiting, and are based on the flawed logic that The answer to the question, ‘to whom should the
if patients in the trial received an average benefit treatment be offered so that more good than
then patients like those in the trial will also, on harm is achieved?’ requires a careful considera-
average, benefit. However, this ‘average’ in the tion of all the potential benefits of the treatment
trial may include patients with greater than aver- as well as the harms (risks or side-effects) that
age benefit, some with the average benefit, and could be attributed to the treatment. Of particu-
some in whom there was net harm. lar concern are patients who are at low risk of
Applying the Results of Clinical Trials and Reviews 63
the main outcome which the intervention is Typically, a study or systematic review evalu-
supposed to prevent (death, stroke, myocardial ating the effectiveness of the intervention focuses
infarction, etc.). Generally, in such patients the on a single primary end-point and possibly a
intervention can be expected to achieve only a small number of secondary end-points. Some-
small absolute risk reduction, so the harms are times side-effects are reported but often the
much more important as they are likely to out- number of participants with side-effects is small
weigh the benefits. because the trial was not designed to detect the
64 The Advanced Handbook of Methods in Evidence Based Healthcare
Table 4 Potential benefits and harms associated with two preventive treatments
Benefits Harms
effects of different treatments on these outcomes Are there variations in the relative
(particularly long-term effects). treatment effect?
All patient-relevant end-points should be con-
sidered that are potentially influenced by the To better understand how the intervention
intervention, including, in particular, adverse works, and how to apply the results of evalu-
effects. For example, anti-arrhythmic drugs have ations, it is useful to determine whether treat-
pro-arrhythmic effects; anticoagulants increase ment effect is constant or whether it varies
the risk of bleeding. Particularly for groups at according to various characteristics relating to
low risk of the primary outcome such adverse patients (e.g. age, gender, biochemical markers);
effects may be crucial. It is helpful to begin by the disease (e.g. hormone receptor status); the
tabulating all that is known about possible intervention (e.g. the timing, compliance, or
positive and negative effects of the interven- intensity of the intervention); and the measure of
tion, even if data for some outcomes are not effect used (relative risk versus risk difference).
available. In working through this, an understanding of
Examples of the potential benefits and harms the terms ‘effect modification’, ‘heterogeneity’
associated with antihypertensive treatment in the and ‘interaction’ is required (Table 5).
elderly and screening of high-risk people for When there are several studies evaluating
colorectal cancer are displayed in Table 4.71,72 the intervention, the treatment effect can be
Trial design setting, subjects, co-administered therapy, length of follow-up, outcomes measured
and how, completeness of follow-up, study quality
Characteristics of:
Patient age, gender, race, co-morbidity, biochemical markers, genetic markers
Disease method and accuracy of diagnosis, severity, stage, responsiveness
Intervention form, mode of administration, dose, intensity, timing, duration, compliance
explored further by firstly conducting a test for across different groups is the level of (untreated)
heterogeneity. If the treatment effect appears to risk. Generally, low-risk groups will have less to
be heterogeneous (or non-constant) then the gain from an intervention than high-risk groups,
sources of heterogeneity (or important effect and hence they may not gain sufficient benefit to
modifiers) should be examined via testing for outweigh harms from treatment (adverse effects,
interactions between each of the variables and costs, etc.).
treatment. In fact, since the test of heterogeneity The benefit for an individual (as measured by
typically has low power and sometimes hetero- the absolute risk reduction or Number Needed to
geneity may not be apparent even in the pres- Treat; NNT) is likely to depend strongly on
ence of effect modification (because the studies the baseline level of risk. Table 7 shows the
were homogeneous on the effect modifier), relationship between baseline risk and absolute
interactions should be explored regardless of the risk reduction or NNT for a constant relative
result of the test of heterogeneity. risk reduction.
Potential effect modifiers will vary according Which patients will obtain a net benefit will
to the treatment and disease under investigation. depend on the harmful effects of the inter-
For example, a number of biological/disease vention, such as adverse drug reactions, side-
factors and patient features may be effect modi- effects, the burden of compliance and monitor-
fiers in relation to treatment of breast cancer ing, etc. The greater the potential harm, the
including age/menopausal status, nodal status, greater the benefit required to make it worth-
receptor status, tumour size, grade, ploidy and while. This trade-off, and how it relates to
type, vascular invasion, labelling index, S-phase baseline risk is illustrated in Figure 1, which
fraction, body mass and family history.57 Hetero- shows benefit increasing but harm being con-
geneity may occur by chance (but this can be
stant across different levels of baseline risk.
checked using a statistical test) or by choosing
This extrapolation from the estimated relative
the wrong measure of treatment effect (relative
risk versus risk difference). If these causes have risk to the expected absolute risk reduction for
been dismissed, then possible sources of hetero- different risk groups assumes that the relative
geneity can be grouped into four areas. Some risk reduction remains constant – an assumption
examples are shown in Table 6. that needs to be checked. For many interventions
the relative risk appears to be reasonably con-
How does the treatment effect vary with stant. An analysis of individual study results
included in 112 meta-analyses found a statis-
baseline risk level? tically significant relationship between the meas-
A very important factor to consider when look- ure of treatment effect and baseline risk in 15
ing at whether the magnitude of benefit varies (13%) meta-analyses with relative risk as the
10% 5% 20 2.5% 40
1% 0.5% 200 0.25% 400
1 in 1000 0.05% 2000 1 in 4000 4000
1 in 10 000 1 in 20 000 20 000 1 in 40 000 40 000
66 The Advanced Handbook of Methods in Evidence Based Healthcare
measure of treatment effect, in 16 (14%) with surrogate for the underlying baseline or un-
the odds ratio, and 35 (31%) with the risk treated risk).82 In their example – tocolysis for
difference.74 Thus, the relative risk was constant pre-term birth – there appeared to be a decreas-
for 87% of the systematic reviews. However, ing relative risk reduction with decreasing
there are clearly cases that deviate considerably control group risk, i.e. a non-constant relative
from this assumption. For example, Rothwell risk. While such an analysis is potentially very
has shown that in patients with carotid artery informative, a number of authors have discussed
stenosis treated with carotid endarterectomy, the the accompanying methodological pitfalls, and
relative risk is very different for different risk alternative methods based on Bayesian and fre-
groups.32 In the high-risk groups there is a high quentist approaches have been proposed.78–81
relative risk reduction, but in the low-risk group The other assumption being made here is that
there is no apparent benefit. By contrast, he the harms are constant across levels of baseline
showed that aspirin has a similar relative benefit risk for the event being prevented. In some
across both groups. A second example of a non- cases, higher-risk patients may be more suscep-
constant relative risk reduction is the Class I tible to harm (e.g. surgical mortality). However,
anti-arrhythmic drugs. Boissel et al. found in in many cases the harms may be directly asso-
trials of Class I anti-arrhythmics in patients after ciated with exposure to the intervention itself
myocardial infarction, that because of different (e.g. adverse effects of drugs). However, again
inclusion criteria, there was a large variation this assumption should be checked.
in the degree of risk in these trials.75 There
appeared to be a beneficial effect in the very
high risk, but in lower-risk groups they appeared Application to Individuals
to do net harm. This result was confirmed by
the CAST study which showed a doubling of What are the predicted absolute risk
mortality from flecainide.76
Since most patient groups to which trial
reductions for individuals?
results will be applied will not have a risk The first three steps involved the identification
identical to the average seen in the trial, one of of all potential patient-relevant benefits and
the crucial analyses in assessing the applicability harms, identification of treatment effect modi-
of trial results will be a systematic study of how fiers and exploration of the relationship between
relative benefit varies with baseline risk. This the magnitude of the treatment effect and base-
may be done either within trials, by looking at line (or untreated) risk. These all address the
different prognostic groups, or between trials, by degree to which the estimates of treatment effect
making use of the variation in average baseline from a single study, or a systematic review of
risk due to the different selection processes of several studies, can be transferred to other set-
trials. tings and to other patients.
For example, in a systematic review of 39 The next two issues to be considered relate
RCTs examining the effects of angiotensin- directly to applying the results to an individual
converting enzyme (ACE) inhibitors in the treat- (or a group of similar individuals).
ment of congestive heart failure, the studies While the relative risk is useful for assessing
were stratified according to the annual mortality the biological strength of response to the treat-
rate – the low-risk group with mortality of up to ment, to judge whether therapy is worthwhile
15% (28 studies) and the high-risk group with for an individual, the absolute magnitude of
mortality greater than 15% (11 studies).77 The benefit should be estimated. This might be
relative risks for mortality in the group treated expressed as the absolute risk reduction, or as
with ACE inhibitors compared with those receiv- the NNT. However it is expressed, it varies with
ing placebo were 0.88 (95% confidence interval the patient’s baseline risk (which is sometimes
0.80–0.97) for the low-risk group and 0.64 referred to as the Patient’s Expected Event Rate,
(0.51–0.81) for the high-risk group. Therefore, PEER): for low-risk patients absolute benefit
both the absolute and relative risk reductions may not outweigh the absolute harm. If the
due to use of ACE inhibitors in patients with relative risk reduction is constant across levels
impaired left ventricular dysfunction and symp- of baseline risk and there is no effect modifica-
toms of heart failure appear to increase with tion, then the average relative risk reduction can
mortality risk. be applied to a particular patient. If this is not
When examining this relationship between the case, then the relationships need to be
treatment effect and baseline risk, there are a described so that the appropriate relative risk
number of methodological problems that can can be applied for a particular patient.
produce misleading results.78–81 For example, To estimate the absolute risk reduction for an
Brand and Kragt suggested plotting the relative individual, the expected relative risk reduction
risk against the risk in the control group (as a can be applied to the PEER. This then requires
Applying the Results of Clinical Trials and Reviews 67
an estimate of the PEER which can be obtained choose to, make a decision about the inter-
from a previously developed (and preferably vention based on their preferences.
validated) prognostic model linking values of Often, there will be multiple effects asso-
various (baseline/pre-randomisation) character- ciated with the intervention – both beneficial
istics of the patient to the probability of the and harmful. These need to be weighed up and
disease of interest. trade-offs made in formulating recommendations
The following example illustrates this point. and in making individual treatment decisions.
In considering whether to treat two patients with The trade-offs between different outcomes
aspirin to prevent a further cardiovascular event, can involve either the quality of the outcome,
Boissel estimated their one-year mortality risk such as different dimensions of quality of life, or
using a previously established risk score.83 they may involve the timing of occurrence of the
Patient 1: A 45-year-old man who has normal outcome. A particularly frequent and important
blood cholesterol and glucose levels, blood trade-off between outcomes is that between
pressure and weight and who is a non-smoker. quality and quantity of life. For example, many
Presented with an uncomplicated inferior surgical procedures, such as hip replacement or
myocardial infarction. His one-year mortality cholecystectomy, involve a mortality risk that
risk is 2%. may be acceptable because of the reduction in
symptoms or improvement of function. This
Patient 2: A 65-year-old woman who has had balancing of benefits and harms requires a com-
hypertension for 30 years, has diabetes, had a mon scale. For example, we may need to
previous myocardial infarction, smokes two compare having a stroke with a death from
packs of cigarettes a day and reports breath- intracranial haemorrhage. This is not easy, but
lessness with exercise. Presented with her such a valuation is implicit in clinical decisions.
second anterior myocardial infarction together If it is made explicit, discussion about indi-
with signs of heart failure. Her one-year mor- vidualisation of treatment will be easier. Quality
tality risk is 30%. of life scales, ranking outcomes on a scale
Applying a (constant) relative risk reduction of between death (0) and normal good health (1)
15% in all-cause mortality, the two patients can attempt to do this.87 Where there are trade-offs
expect absolute risk reductions of 0.3% and between quality and quantity of life, the Quality
4.5% respectively (NNTs of 300 and 22). These Adjusted Life Year (QALY) attempts to com-
can be weighed against an excess risk of bine both into a single measure.88 Another sum-
cerebral bleeding with aspirin of 0.12% for both mary measure that has been used in cancer trials
patients. is TWiST, the Time Without Symptoms and
There are many examples of prognostic models Toxicity, which deals simultaneously with sur-
in the literature. They tend to be based on cohort vival gains and the changes in quality of life due
(prospective) studies in which a sample of to toxicity (loss) and delays or avoidance of
patients are recruited, various baseline charac- relapse (gain).89 This measure was refined to
teristics are recorded, and then the patients are include differential weightings for quality of
followed over time to see whether they develop life while experiencing toxicity and relapse
the outcome of interest. Statistical modelling (Q-TWiST).90,91
(such as logistic regression or proportional haz- Appropriate reporting and analysis of clinical
ards models) is then used to examine which trials is needed to make the trade-offs explicit,
combination of variables best predict which and to allow an informed decision by healthcare
patients experience the outcome of interest. provider and patient. However, if the previous
Examples from the literature include prediction step of estimating an individual’s potential bene-
of 10-year survival after definitive surgical ther- fits and harms is done well, and presented in a
apy for primary cutaneous melanoma,84 predic- clear way, often the trade-offs will be clear. The
tion of 5-year risk of a cardiovascular event85 central issue is whether, for the individual
and expected rates of thromboembolic stroke in patient, the predicted absolute benefit has
patients with atrial fibrillation.86 greater value than the harm and cost of treat-
ment. For example, when does the reduction in
strokes outweigh the risk of bleeding from anti-
Do the benefits outweigh the harms? coagulation, or when does the benefit of surgery
The final consideration in formulating recom- outweigh its risk?
mendations about to whom the intervention If no adverse events have been identified, then
should be applied is how the potential benefits it may seem that only benefits need to be con-
and harms compare. Information on the benefits sidered. However, serious adverse events may
and harms by levels of baseline risk should be be rare and hence the concern is how reliably
presented so that individuals (patients) together they have been excluded. This ‘proof of safety’
with their healthcare providers can, if they is usually more difficult, and rare adverse effects
68 The Advanced Handbook of Methods in Evidence Based Healthcare
(say of drugs) may only emerge during post- Development of Clinical Practice
marketing monitoring. Several authors have dis- Guidelines
cussed the statistical difficulties of excluding
such serious rare events. For example, if 300 Guidelines-writing groups should consider broad
patients have been treated in controlled trials issues regarding the applicability of the evidence
and no serious adverse effects detected, the including the reproducibility of the intervention
confidence interval for the rate of such events is
in different settings and biological factors that
from 0% to 1% – that is, we can be 95% certain
may alter the treatment’s effectiveness (although
that the risk of an adverse effect is less than
these are likely to be rare). They also need to
1%.92
develop more explicit processes to consider trial
applicability, and the weighing of harms and
benefits for individual patients. Therefore, guide-
IMPLICATIONS AND FURTHER RESEARCH lines for developing clinical practice guidelines
should include methods for assessing trial ap-
The results of this systematic review have impli- plicability and should recommend more formal
cations for the design and reporting of clinical processes to encourage accurate application of
trials, methods for applying the results of trials the evidence.
and systematic reviews in clinical practice and Traditional methods for assessing applicabil-
how groups developing clinical practice guide- ity based on subject selection for the trials,
lines should consider the evidence when formu- inclusion/exclusion criteria, and subgroup analy-
lating treatment recommendations. Each of these ses are flawed. Rather, guidelines-writing groups,
will be discussed in turn. using the process we have described, should iden-
tify in which individuals or homogeneous groups
the treatment is more likely to do good than harm
Conduct and Reporting of Clinical as the basis for treatment recommendations.
Trials and Systematic Reviews Guidelines should be designed to enable the
clinician and patient to assess the individual’s
The principal need in designing and reporting risk or prognosis and the consequent potential
clinical trials is that all patient-relevant out- benefit and harm from treatment (the New Zeal-
comes (both benefits and harms) are measured and guidelines on the treatment of hypertension
and reported. Also, effect modification and the are an excellent example85). Ultimately, they
variation in treatment effect across levels of should not adopt a prescriptive algorithmic
baseline risk should be explored and discussed. approach.
The role of, and need for, inclusion and exclu-
sion criteria should be reconsidered. There is a
need for large pragmatic trials with hetero- Areas for Further Research
geneous patient populations to facilitate an
exploration of effect modification. Further methodological and empirical research
People undertaking systematic reviews should is required to examine relationships between
examine heterogeneity, effect modification and relative treatment effect and baseline risk, and to
the importance of baseline risk rather than con- develop appropriate statistical methods to do
centrate on producing a pooled (average) esti- this. These methods should take account of the
mate of treatment effect. A balance sheet of negative correlation between a measure of treat-
benefits and harms should be included in the ment effect and the event rate in the control
report. group, the error associated with the estimate of
the event rate in the control group, regression–
dilution bias and problems of misclassification
Clinical Practice regarding presence/absence of the outcome of
interest due to limitations in the measurements.
Practitioners, when attempting to apply the Methods for presenting the magnitude of the
results of trials in treatment decisions, need to treatment effect for both benefits and harms that
go beyond a consideration of the characteristics differ in levels of severity and timing of occur-
of the trial patients, settings and available clin- rence require further development.
ical care and how they compare with their Further methodological development is needed
situation. to create prognostic models from studies to
Rather, decision aids and clinical practice enable an independent estimation of predicted
guidelines should be made available to assist in risk. There is a need for further investigation of
the identification of those patients in whom the the use of comprehensive cohorts for develop-
treatment is more likely to do good than harm. ing and testing prognostic models using both
Applying the Results of Clinical Trials and Reviews 69
patients who entered the trial and those who did 11. Feagan BG, McDonald JWD, Koval JJ. Thera-
not get randomised. peutics and inflammatory bowel disease: a guide
Finally, research is needed into the question to the interpretation of randomized controlled
of whether the adoption of ‘formal’ approaches trials. Gastroenterology 1996; 110: 275–83.
to the application of trial results leads to better 12. Schooler NR. How generalizable are the results
outcomes for patients. of clinical trials? Psychopharm. Bull. 1980; 16:
29–31.
13. Remington RD. Potential impact of exclusion
criteria on results of hypertension trials. Hyper-
ACKNOWLEDGEMENTS tension 1989; 13(Suppl. I): I66–8.
14. Rapaport MH, Frevert T, Babior S, Zisook S,
This work was supported by a grant from the NHS Judd LL. A comparison of demographic vari-
Health and Technology Assessment Research ables, symptom profiles, and measurement of
and Development Programme. functioning in symptomatic volunteers and an
outpatient clinical population. Psychopharm. Bull.
1995; 31: 111–14.
15. Antman K, Amato D, Wood W, et al. Selection
REFERENCES bias in clinical trials. J. Clin. Oncol. 1985; 3:
1142–7.
16. Kramer MS, Shapiro SH. Scientific challenges in
1. Collet JP, Boissel JP. The sick population –
the application of randomized trials. JAMA 1984;
treated population: the need for a better definition.
252: 2739–45.
The VALIDATA Group. Eur. J. Clin. Pharmacol.
17. Juster HR, Heimberg RG, Engelberg B. Self
1991; 41: 267–71.
selection and sample selection in a treatment
2. Lubsen J, Tijssen JG. Large trials with simple
study of social phobia. Behav. Res. Ther. 1995;
protocols: indications and contraindications. 33: 321–4.
Controlled Clinical Trials 1989; 10(Suppl.): 18. Llewellyn-Thomas HA, McGreal MJ, Thiel EC,
151S–60S. Fine S, Erlichmen C. Patients’ willingness to
3. Yusuf S, Held P, Teo KK, Toretsky ER. Selection enter clinical trials: measuring the association
of patients for randomized controlled trials: im- with perceived benefit and preference for decision
plications of wide or narrow eligibility criteria. participation. Soc. Sci. Med. 1991; 32: 35–42.
Statist. Med. 1990; 9: 73–83. 19. Wilhelmsen L, Ljungberg S, Wedel H, Werko L.
4. Sackett DL. The fourth prerequisite: the inclusion/ A comparison between participants and non-
exclusion criteria strike a balance between effi- participants in a primary preventive trial. J.
ciency and generalisability. In: Shapiro SH, Louis Chron. Dis. 1976; 29: 331–9.
TA (eds) Clinical Trials: Issues and Approaches. 20. Charlson ME, Horwitz RI. Applying results of
New York: Marcel Dekker. 1983: Vol. 46, randomised trials to clinical practice: impact of
pp. 72–4. losses before randomisation. Br. Med. J. Clin.
5. George SL. Reducing patient eligibility criteria in Res. Ed. 1984; 289: 1281–4.
cancer trials. J. Clin. Oncol. 1996; 14: 1364–70. 21. Smith P, Arnesen H. Mortality in non-consenters
6. Simon R. Patient subset and variation in thera- in a post-myocardial infarction trial. J. Intern.
peutic efficacy. In: Chaput de Saintonge DM, Med. 1990; 228: 253–6.
Vere DW (eds) Current Problems in Clinical 22. Biener L, DePue JD, Emmors KM, Linnan L,
Trials. 1st edn. London: Blackwell Scientific Abrams DB. Recruitment of work sites to a health
Publications, 1984. promotion research trial. J. Occup. Med. 1994;
7. King AC, Harris RB, Haskell WL. Effect of 36: 631–6.
recruitment strategy on types of subjects entered 23. Pincus T. Limitations to standard randomized
into a primary prevention clinical trial. Ann. Epi- controlled trials to evaluate combination therapies
demiol. 1994; 4: 312–20. in rheumatic diseases. Agents-Actions 1993;
8. Simpson FO. Fallacies in the interpretation of the 44(Suppl.): 83–91.
large-scale trials of treatment of mild to moderate 24. Brundage MD, Mackillop WJ. Locally advanced
hypertension. J. Cardiovasc. Pharmacol. 1990; non-small cell lung cancer: do we know the ques-
16(Suppl. 7): S92–5. tions? A survey of randomized trials from
9. Wittes RE. Problems in the medical interpretation 1966–1993. J. Clin. Epidemiol. 1996; 49: 183–92.
of overviews. Statist. Med. 1987; 6: 269–80. 25. Ziegelstein RC, Chandra NC. Selecting patients
10. Richmond RL, Anderson P. Research in general who benefit from thrombolysis. Coronary Artery
practice for smokers and excessive drinkers in Disease 1994; 5: 282–6.
Australia and the UK. II. Representativeness of 26. Mitchell JR. ‘But will it help my patients with
the results. Addiction 1994; 89: 41–7. myocardial infarction?’ The implications of
70 The Advanced Handbook of Methods in Evidence Based Healthcare
recent trials for everyday country folk. Br. Med. J. 41. Droitcour J, Silberman G, Chelimsky E. A new
1982; 285: 1140–8. form of meta-analysis for combining results from
27. Ausman JI, Diaz FG. Critique of the randomized clinical trials and medical practice
Extracranial–Intracranial Bypass Study. Surg. databases. Int. J. Technol. Assess. Health Care
Neurol. 1986; 26: 218–21. 1993; 9: 440–9.
28. Kober L, Torp-Pedersen C. On behalf of the 42. Hlatky MA, Califf RM, Harrell FE Jr, Lee KL,
TRACE Study Group. Clinical characteristics and Mark DB, Pryor DB. Comparison of predictions
mortality of patients screened for entry into the based on observational data with the results of
Trandolapril Cardiac Evaluation (TRACE) Study. randomized controlled clinical trials of coronary
Am. J. Cardiol. 1995; 76: 1–5. artery bypass surgery. J. Am. Coll. Cardiol. 1988;
29. Pocock SJ. Organisation and planning. In: 11: 237–45.
Clinical Trials: a practical approach. London: A 43. Hlatky MA, Califf RM, Harrell FE, et al. Clinical
Wiley Medical Publication, 1984, Chapter 3, judgement and therapeutic decision making. J.
pp. 35–8. Am. Coll. Cardiol. 1990; 15: 1–14.
30. Williford WO, Krol WF, Buzby GP. Comparison 44. The Cardiac Arrhythmia Suppression Trial
of eligible randomized patients with two groups (CAST) Investigators. Preliminary report: effect
of ineligible patients: can the results of the VA of encainide and flecainide on mortality in a
total parenteral nutrition clinical trial be gener- randomized trial of arrhythmia suppression after
alized? J. Clin. Epidemiol. 1993; 46: 1025–34. myocardial infarction. N. Engl. J. Med. 1989;
31. Glasziou PP, Irwig LM. An evidence based 321: 406–12.
approach to individualising treatment. Br. Med. J. 45. Early Breast Cancer Trialists’ Collaborative
1995; 311: 1356–9. Group. Systemic treatment of early breast cancer
32. Rothwell PM. Can overall results of clinical trials by hormonal, cytotoxic, or immune therapy. Part
be applied to all patients? Lancet 1995; 345: I. Lancet 1992; 339: 1–15. Part II. Lancet 1992;
1616–19. 339: 71–85.
33. Day AL, Rhoton AL, Little JR. The Extracranial- 46. Kazempour K, Kammerman LA, Farr SS. Sur-
Intracranial Bypass Study. Surg. Neurol. 1986; vival effects of ZDV, ddI and ddC in patients
26: 222–6. with CD4 # 50 cells/mm3. J. Acquir. Immune
34. Sundt TM. Was the international randomized trial Defic. Syndr. Hum. Retrovirol. 1995; 10(Suppl. 2):
of extracranial-intracranial arterial bypass rep- S97–106.
resentative of the population at risk? N. Engl. J. 47. Fine MJ, Smith MA, Carson CA, et al. Efficacy
Med. 1987; 316: 814–16. of pneumococcal vaccination in adults. A meta-
35. Walsh JT, Gray D, Keating NA, Cowley AJ, analysis of randomized controlled trials. Arch.
Hampton JR. ACE for whom? Implications for Intern. Med. 1994; 154: 2666–77.
clinical practice of post-infarct trials. Br. Heart J. 48. Easton JD, Wilterdink JL. Carotid endarterect-
1995; 73: 470–4. omy: trials and tribulations. Ann. Neurol. 1994;
36. Schumacher M, Davis K. Combining randomized 35: 5–17.
and nonrandomized patients in the statistical analy- 49. Gruppo Italiano per lo studio della Sopavvivenza
sis of clinical trials. Rec. Results Cancer Res. 1988; nell’Infarto Miocardico. GISSI–2: a factorial ran-
111: 130–7. domised trial of alteplase versus streptokinase and
37. Davis K. Comprehensive cohort study: the use of heparin versus no heparin among 12,490 patients
registry data to confirm and extend a random- with acute myocardial infarction. Lancet 1990;
ized trial. Rec. Results Cancer Res. 1988; 111: 336: 65–71.
138–48. 50. Multicentre Acute Stroke Trial-Italy (MAST-I)
38. Olschewski M, Schumacher M, Davis KB. Analy- Group. Randomised controlled trial of strepto-
sis of randomized and nonrandomized patients in kinase, aspirin, and combination of both in treat-
clinical trials using comprehensive cohort follow- ment of acute ischemic stroke. Lancet 1995; 346:
up study design. Controlled Clinical Trials 1992; 1509–14.
13: 226–39. 51. Gheorghiade M, Schultz L, Tilley B, Kao W,
39. Schmoor C, Olschewski M, Schumacher M. Goldstein S. Effects of propranolol in non-Q-
Randomized and non-randomized patients in clin- wave acute myocardial infarction in the Beta
ical trials: experiences with comprehensive cohort Blocker Heart Attack Trial (BHAT). Am. J.
studies. Statist. Med. 1996; 15: 263–71. Cardiol. 1990; 66: 129–33.
40. Mitchell GJ, Glenn J, Pryor DB, Cohen WS. 52. The EC/IC Bypass Study Group. Failure of
Cross design synthesis. A new strategy for extracranial-intracranial arterial bypass to reduce
medical effectiveness research. Washington DC. the risk of ischemic stroke. Results of an inter-
United States General Accounting Office, 1992, national randomized trial. N. Engl. J. Med. 1985;
Doc. No.B-244808. 313: 1191–200.
Applying the Results of Clinical Trials and Reviews 71
53. Peto R. Why do we need systematic overviews 69. Sonis J, Doukas D, Klinkman M, Reed B, Ruffin
of randomized trials? Statist. Med. 1987; 6: MT. Applicability of clinical trial results to pri-
233–44. mary care. JAMA 1998; 280: 1746.
54. Collins R, Gray R, Godwin J, Peto R. Avoidance 70. Katon W, Von Korf M, Lin E, et al. Methodo-
of large biases and large random errors in the logic issues in randomized trials of liaison psy-
assessment of moderate treatment effects: the chiatry in primary care. Psychosom. Med. 1994;
need for systematic overviews. Statist. Med. 56: 97–103.
1987; 6: 245–54. 71. Thijs L, Fagard R, Lijnen P, et al. Why is
55. Peto R, Collins R, Gray R. Large-scale random- antihypertensive drug therapy needed in elderly
ized evidence: large, simple trials and overviews patients with systolodiastolic hypertension? J.
of trials. J. Clin. Epidemiol. 1995; 48: 23–40. Hypertens. Suppl. 1994; 12: S25–34.
56. Pocock SJ, Hughes MD. Estimation issues in 72. Eddy DM. Clinical Decision Making: from theory
clinical trials and overviews. Statist. Med. 1990; to practice. Comparing benefits and harms: the
9: 657–71. balance sheet. JAMA 1990; 263: 2498–501.
57. Gelber RD, Goldhirsch A. Interpretation of 73. Guyatt GH, Sackett DL, Sinclair JC, et al. Users’
results from subset analyses within overviews of guides to the medical literature. IX. A method for
randomized clinical trials. Statist. Med. 1987; 6: grading health care recommendations. JAMA
371–8. 1995; 274: 1800–4.
58. Gail M, Simon R. Testing for qualitative inter- 74. Schmid CH, Lau J, McIntosh MW, Cappelleri JC.
actions betweeen treatment effects and patient An empirical study of the effect of the control
subsets. Biometrics 1985; 41: 361–72. rate as a predictor of treatment efficacy in meta-
59. Pocock SJ. Further aspects of data analysis. analysis of clinical trials. Statist. Med. 1998; 17:
In: Clinical trials: a practical approach. London: 1923–42.
A Wiley Medical Publication, 1984, Chapter 14, 75. Boissel JP, Collet JP, Lievre M, Girard P. An effect
pp. 213–16. model for the assessment of drug benefit: example
of antiarrthythmic drugs in postmyocardial infarc-
60. Shuster J, van Eys J. Interaction between prog-
tion patients. J. Cardiovasc. Pharmacol. 1993; 22:
nostic factors and treatment. Controlled Clinical
356–63.
Trials 1983; 4: 209–14.
76. The Cardiac Arrhythmia Suppression Trial
61. Davis CE. Generalizing from clinical trials. Con-
(CAST) Investigators. Preliminary report: effect
trolled Clinical Trials 1994; 15: 11–14.
of encainide and flecainide on mortality in a
62. Tannock IF. Assessment of study design in clin-
randomized trial of arrhythmia suppression after
ical trials for bladder cancer. Urol. Clin. North.
myocardial infarction. N. Engl. J. Med. 1989;
Am. 1992; 19: 655–62.
321: 406–12.
63. Gyorkos TW, Tannenbaum TN, Abrahamowicz 77. North of England Evidence-based Guideline
M, et al. An approach to the development of Development Project. Evidence based clinical
practice guidelines for community health inter- practice guideline. ACE inhibitors in the primary
ventions. Can. J. Public Health (Suppl. 1) 1994; care management of adults with symptomatic
Jul–Aug: S8–13. heart failure. Newcastle upon Tyne: Centre for
64. Nunnelee JD. The inclusion of women in clinical Health Services Research, 1997.
trials of antihypertensive medications: a review of 78. McIntosh M. The population risk as an explana-
twenty-four trials in one pharmacology journal. J. tory variable in research synthesis of clinical
Vasc. Nurs. 1995; 13: 41–9. trials. Statist. Med. 1996; 15: 1713–28.
65. Piantadosi S, Wittes J. Politically correct clinical 79. Sharp S, Thompson SG, Douglas GA. The rela-
trials. Controlled Clinical Trials 1993; 14: tion between treatment benefit and underlying
562–7. risk in meta-analysis. Br. Med. J. 1996; 313:
66. Dans AL, Dans LF, Guyatt GH, et al. Users’ 735–8.
guides to the medical literature: XIV. How to 80. Walter SD. Variation in baseline risk as an
decide on the applicability of clinical trial results explanation of heterogeneity in meta-analysis.
to your patient. JAMA 1998; 279: 545–9. Statist. Med. 1997; 16: 2883–900.
67. Diabetes Control and Complications Trial 81. Thompson SG, Smith TC, Sharp SJ. Investigating
Research Group (DCCT). The effect of intensive underlying risk as a source of heterogeneity in
treatment of diabetes on the development and meta-analysis. Statist. Med. 1997; 16: 2741–58.
progression of long term complications in insulin 82. Brand R, Kragt H. Importance of trends in the
dependent diabetes mellitus. N. Engl. J. Med. interpretation of an overall odds ratio in the meta-
1993; 329: 977–86. analysis of clinical trials. Statist. Med. 1992; 11:
68. Marshall KG. Prevention. How much harm? How 2077–82.
much benefit? 2. Ten potential pitfalls in deter- 83. Boissel JP. Individualizing aspirin therapy for
mining the clinical significance of benefits. Can. prevention of cardiovascular events. JAMA 1998;
Med. Assoc. J. 1996; 154: 1837–43. 280: 1949–50.
72 The Advanced Handbook of Methods in Evidence Based Healthcare
84. Schuchter L, Schultz DJ, Synnestvedt M, et al. A adjusted life years. Int. J. Technol. Assess. Health
prognostic model for predicting 10-year survival Care 1989; 5: 559–75.
in patients with primary melanoma. Ann. Intern. 89. Price K, Gelber R, Isley M, Goldhirsch A, Coates
Med. 1996; 125: 369–75. A, Castiglione M. Time without symptoms and
85. Guidelines for the management of mildly raised toxicity (TWiST): a quality-of-life-oriented end-
blood pressure in New Zealand. The Core Ser- point to evaluate adjuvant therapy. Adjuvant
vices Committee, Wellington, New Zealand. Therapy of Cancer 1987; V: 455–65.
86. Stroke Prevention in Atrial Fibrillation Invest- 90. Goldhirsch A, Gelber RD, Simes RJ, Glasziou P,
igators. Warfarin versus aspirin for prevention of Coates AS. Costs and benefits of adjuvant therapy
thromboembolism in atrial fibrillation: Stroke in breast cancer. A quality-adjusted survival analy-
Prevention in Atrial Fibrillation II Study. Lancet sis. J. Clin. Oncol. 1989; 7: 36–44.
1994; 343: 687–91. 91. Glasziou PP, Simes RJ, Gelber RD. Quality
87. Ware JE Jr, Sherbourne CD. The MOS 36-item adjusted survival analysis. Statist. Med. 1990; 9:
short-form health survey (SF–36). I. Conceptual 1259–76.
framework and item selection. Med. Care 1992; 92. Hanley JA, Lippman-Hand A. If nothing goes
30: 473–83. wrong is everything all right? Interpreting zero
88. Torrance G, Feeny D. Utilities and quality- numerators. JAMA 1983; 249: 1743–4.
5
The Placebo Effect:
Methodological Process and
Implications of a Structured Review
interpretation, which emphasises patient auton- are unspecified rather than non-specific.34 The
omy and involvement rather than professional challenge of gaining a better understanding of
paternalism and control.24 the placebo effect remains, and awaits con-
Probably the most widely quoted definitions ceptual developments that isolate parameters
of placebo and placebo effect are those of and scientific enquiry that tests for their precise
Shapiro.25–29 Shapiro extends the definition clinical significance. The end product of this
of the placebo to ‘any therapy (or component of process will be the transformation of the non-
therapy) deliberately used for non-specific psy- specific (placebo) effect into named specific
chological or psychophysiological effect . . . and therapeutic activities.2,3,8,27,31,34–36 At that point,
without specific activity for the condition being the term ‘placebo’ effect could be dispensed
treated’. The placebo effect, accordingly, is with since there would no longer be any mysteri-
defined as ‘the non-specific psychological ous, atheoretical component to therapy. Thus,
or psychophysiological effect produced by there is an inherent paradox in investigating
placebos’. The focus on ‘non-specific’ in these placebo effects. Once they are understood they
definitions suggests that the magnitude of the are no longer defined as placebo effects. It was
placebo effect can be deduced by excluding one purpose of the review to develop a frame-
known specific effects of the therapy on the work for the study of non-specific effects, as a
condition in question.30 first step towards explicating the mechanisms
Grünbaum advanced on Shapiro’s definition by which they can enhance beneficial health
by observing that a treatment is composed of outcomes.
two components: characteristic factors and inci- It has been argued that eliciting the placebo
dental factors.4,5 The characteristic factors are effect does not require a placebo in the tra-
those that are known or believed to affect the ditional sense of the term, and that placebos and
disease as a result of the theoretical rationale for the placebo effect should therefore be separately
the therapy. The incidental factors are those that defined.8,15,37,38 According to this interpretation,
may affect disease but cannot be derived from the placebo effect derives from the symbolic
the theoretical rationale for the therapy. Therapies effect of treatment as determined by the total
can affect the target conditions for which they are context in which healthcare is delivered. The
intended, and/or other aspects of patients’ causes of the placebo effect are located any-
health. Given this framework, Grünbaum argued where in the care delivery process, except in
that the terms ‘specific’ and ‘non-specific’ to the inert placebo itself. Consistent with this
distinguish between treatment and placebo approach is the suggestion that placebo effects
effects are not helpful. A placebo may have a are generic in applicability.37,39 It also permits
highly specific effect on the target disease, for the extension of the definition of a placebo
example it can make a headache go away. The beyond that of an inert medication to encapsu-
specificity metaphor is borrowed from medicine, late all aspects of the treatment environment.
where specific therapies are developed to Many things, including practitioners themselves,
address specific diseases, and a general panacea have been described as placebos.40 Indeed, a
is looked upon with scepticism. However, as continuum of placebos has been suggested,
Shepherd31 observed, the concept of specificity ranging from tangible items such as scars, pills,
in biology and medicine has always been at injections, white coats and procedures, to in-
issue: that is, whether or not any given disease tangible features of healthcare delivery such as
has a specific versus non-specific cause and touch, gesture, ambience and support.41,42
cure. The issue of whether or not a placebo is
The term ‘incidental’ is intended to signal that required to produce the placebo effect reduces
such effects are not expected on the basis of the therefore to one of semantics, and hinges on the
underlying rationale behind the treatment. This definitions of placebo and placebo effect that are
point is similar to that made by Critelli and adopted. If a placebo is narrowly defined as an
Neumann32 when they observed that placebos inert substance it is only one aspect of the total
may be theoretically inert, but not therapeutic- treatment context, and one of many possible
ally inert. In other words, placebos may have means of eliciting a broadly defined placebo
beneficial effects for no apparent reason. How- effect.
ever, that should not be taken to imply that the For the purposes of the review, an inclusive
theoretical basis for incidental effects is inevit- definition of placebo was adopted. The relation-
ably mysterious. ship between any aspects of the healthcare deliv-
Many commentators have noted that the ery encounter were open to investigation for
placebo effect is equated to non-specific because their effects on health outcomes. This included
of ignorance about its component parts. Reflect- the effects of placebo medications, but extended
ing this, it has been referred to as ‘Factor X’.33 beyond that to consider the significance of the
A distinction can be made between effects that features associated with the environment in
The Placebo Effect: Methodological Process and Implications 75
which care is delivered and the practitioner– fied: expectancy, anxiety reduction, classical
patient interaction. This approach acknowledges conditioning, and social support. The expect-
that an opportunity exists for the placebo effect ancy mechanism occupied the central place in
to be activated in some form or another in the psychological literature on the placebo
virtually all encounters between healthcare prac- effect44–47 and was therefore chosen as the focus
titioners and their patients. It suggests that with for the review. The decision was influenced by
a proper understanding of the placebo effect resource considerations, and by the view held
practitioners can, through their relationships by clinicians that expectancy is the probable
with their patients, and in conjunction with explanation of why treatments, once thought
appropriate medical technologies, use a multi- to be efficacious by their proponents but
tude of non-deceptive means to promote positive found more recently not to be, produce positive
placebo responses. Ethical dilemmas associated outcomes.48
with prescribing placebo medications are there- Expectancy can be said to subsume anxiety
by avoided.43 The advantage of this approach reduction and classical conditioning, and so
was that the results of the review could have
these mechanisms were not removed from the
potentially wide practical significance. A variety
of ethically acceptable healthcare delivery im- conceptual framework.7,36,49–55 For example,
provements could subsequently be considered. anxiety reduction as a placebo mechanism, may
be a consequence of positive expectancy.
Similarly, in so far as past experience sets up
ESTABLISHING THE CONCEPTUAL FRAMEWORK learned expectancies, these possible classical
conditioning effects may be more usefully
understood in terms of expectancy mechanisms
The challenge faced by the review team was a since there is some doubt as to whether humans
potentially large literature (searching for placebo can be classically conditioned.56 So that, for
as a keyword alone, produced approximately example, positive responses to repeated asso-
20 000 references) and a topic that was impre- ciation of medical care with symptom relief,
cisely defined. To handle both issues, a con- even when the therapy is non-active, would be
ceptual framework was developed based on the
attributed to the creation of positive expecta-
assumption that certain indicators of the process
tions. Similarly, negative conditioned responses
of healthcare delivery act as placebos to produce
placebo responses. This assumption enabled us (‘placebo sag’) to new therapies following a
to distinguish the mechanisms of the placebo series of ineffective therapies36,57 would be
effect from the placebo responses. Features of attributed to the creation of negative expectan-
healthcare delivery were identified as ‘determin- cies. Both mechanisms were therefore addressed
ants’, and the placebo responses as ‘outcomes’. in the review, but only in so far as they could be
Together with the mechanisms by which the interpreted in the context of expectancy.
placebo effects operate, these two related classes Social support, on the other hand, is deter-
of factors formed the model of the placebo mined by family members and outside organisa-
effect. In an initial scoping exercise, it became tions as well as healthcare practitioners. Since
apparent that placebo effects had been demon- the review was confined to the healthcare set-
strated in numerous studies across a range of ting, social support was addressed only in so far
treatments for a range of disorders. To review as the role of the patient–practitioner inter-
the literature on all aspects of the model would actions were considered as determinants.
have produced a very large literature, and would
have been well beyond the scale of the project.
Therefore it was decided to focus the review on Determinants
the mechanisms by which the placebo effects
operate, since this would increase the possibility Having chosen expectancy as the mechanism by
of harnessing these effects as therapies. The which placebo effects operate, determinants
background literature was used to identify the could then be specified. Determinants include all
mechanisms of the placebo effect and to estab- aspects of the treatment environment since the
lish which determinants might act on them to causes of the placebo effect are located through-
produce the health outcomes. out the care delivery process.40–42 As a large
number of placebogenic variables had been
Mechanisms advanced in the literature,27,34,35,58–69 they were
organised into four groups: patient charac-
We turned to the psychological literature to teristics; practitioner characteristics; patient–
establish which mechanisms had been associ- practitioner interaction; and the treatment and
ated with the placebo effect. Four were identi- treatment setting. Within each group, factors
76 The Advanced Handbook of Methods in Evidence Based Healthcare
categorise health outcomes, and only studies without reference to expectancy. Therefore,
that reported one or more of these outcomes mere demonstrations of placebo effects were
were retained for analysis. Studies that only excluded, such as all randomised trials of
reported intermediate behavioural outcomes new treatments and drugs that are intended to
(such as adherence) without any reference to demonstrate the superiority of the new therapy
health effects, were excluded. above and beyond placebo effects. In addition,
the neurophysiological mechanisms by which
Conceptual Framework for the expectancy affects biological processes were
considered beyond the scope of the review
Placebo Effect because they are controversial (for example, the
role of endogenous opiates in placebo analgesia
The model of the determinants, expectancy has been debated50,89–92).
mechanism and outcomes which provided the The second literature removed from the frame
framework used to guide the review is shown in was the psychological research using laboratory
Figure 1. based experiments to deliberately manipulate
By choosing expectancy as the placebo mech- subject expectancies93 or to manipulate the effect
anism the review focused on the process by of instruction or expectancy on patients.94,95 It
which placebos achieve their therapeutic effec- was felt that findings from the laboratory would
tiveness. It could thus test the hypothesis that not generalise to clinical settings, particularly
changes in health status attributed to placebo since crucial psychological variables present in
effects are achieved by the manipulation of patients undergoing treatment, such as distress,
patient expectations. The theoretical exposition, were not likely to be found in non-patient
although time consuming, unravelled the func- volunteers.66
tional relationships of the placebo effect render- The third literature that was excluded was
ing it accessible to the review process. As the studies concerned with measuring the effective-
purpose of the review was to harness the ness of psychotherapy over placebo treatment.
placebo effect for use in the NHS, the final This literature presents serious challenges to
decision made at this stage was to confine the researchers, particularly because of the uncer-
literature to studies concerned with healthcare tainties that it raises. It has been argued that the
delivery in the clinical sector. placebo effect works like psychotherapy through
transference,96 and that since psychotherapy
Excluded Literature affects patients’ expectancies by providing sup-
port, compassion, reassurance, advice, and shar-
In the process of developing the model of the ing knowledge, it is in fact analogous to a
placebo effect, four literatures were removed placebo. Whilst some commentators argue that
from the search frame. the specific and non-specific (or expectancy or
The first literature that was removed was a set placebo) effects of psychotherapy cannot be
of studies that had investigated placebo effects separated because of the problem of finding a
78 The Advanced Handbook of Methods in Evidence Based Healthcare
credible placebo,8,97–102 others argue that cred- little empirical evidence to assist the debate, and
ible placebos are possible32 and that a treatment what exists is subject to methodological difficul-
effect can be discerned.103,104 There is, however, ties, not least the problem of finding appropriate
significant debate about this conclusion, and placebo controls.113–118 In view of the diversity
many investigations have found no evidence of of the complementary sector and the methodo-
a treatment effect from psychotherapy above the logical difficulties associated with identifying
placebo effect.105–109 the placebo effect in it, this fourth literature was
There is a similar controversial debate about also excluded from the review.
the alternative, or complementary medical sec-
tor, which is frequently discussed as an entity,
although in reality it includes a range of varied SEARCH, RETRIEVAL AND SCREENING OF
therapeutic modalities. Some commentators LITERATURE
suggest that the therapeutic value of these
alternative approaches may be largely, if not Search and Retrieval
entirely, accounted for by the placebo effect,
rather than by specific physiological effects of The search and retrieval procedure that was used
the treatments themselves.59,75,84,110–112 There is is illustrated in Figure 2. The initial scoping
exercise used terms which included placebo, METHODS USED TO ANALYSE THE PRIMARY
synonyms of placebo, expectancy, and syno- RESEARCH PAPERS
nyms of expectancy and identified 47 600 ab-
stracts. The terms were subsequently restricted
to placebo combined with synonyms of expect- The primary research papers (93) identified in
ancy, and to determinants of the placebo mech- the search strategy described studies in which a
anism and their synonyms. After testing these number of research approaches had been used to
restricted terms for recall (sensitivity) and pre- test a range of different clinical hypotheses. In
cision (specificity) the retrieval process was order that the review could test the hypothesis
repeated. This reduced the references that were that changes in health status were achieved by
located to 689. the manipulation of patient expectations, a
method of extracting data appertaining to the
expectancies created by the investigations was
Sources used in the search needed. To this end, a working definition of
The search included journal articles, books, expectancy was developed, together with criteria
chapters of books, letters and editorials. The that verified expectancy as a key target of the
major source was the electronic databases chosen investigation.
to span a broad range of disciplines and including
both European and American material. Although Working Definition of Expectancy
electronic sources were not searched before
1980, pre-1980 studies were included if they had Definitions of expectancy as the mechanism
been identified by other means. Moreover, a of the placebo effect, were developed from
comprehensive bibliography of placebo studies Bandura’s theorising about expectancy. Ban-
prior to 1980 was already available.119 The same dura120,121 distinguished two types of expectan-
search terms were used for each database, cies: outcome expectations and self-efficacy
although there were slight differences in the expectations. For the purposes of this review,
strategy that was used because of differences in his concept of outcome expectancy120–122 was
the databases. interpreted as a treatment-related expectancy
Other literature sources included reference and his concept of self-efficacy expectation120,122
lists of identified papers (known as exploding was interpreted as a patient-related self-efficacy.
references), personal contacts with known These two interpretations of expectancy were
experts in the field and conferences. The latter further developed as follows.
sources provided information about ongoing
work. Treatment-related expectancy
Treatment-related expectancy was developed
Screening and Data Extraction Process into three separate categories.
1 Process expectancy was used to refer to
The second flow diagram (Figure 3) illustrates expectations about medical interventions
the screening and data extraction process. Using created for patients either with no knowl-
a set of prompts, each abstract was screened by edge of an unfamiliar medical procedure, or
at least two members of the team. This process with inaccurate expectations about the actual
reduced the numbers of papers to be retrieved processes involved.
from 689 to 472, 93 of which were primary 2 Positive outcome expectancy was used to
research papers and 379 background papers. refer to expectancies created by practitioners
Using a set of guidelines developed from pilot when they convey their own faith and enthu-
work, data from the primary research papers siasm in the treatment. This particular con-
were extracted and entered onto a proforma. cept of outcome expectancy went further
Information concerning the quality of the than simply providing accurate information
research was extracted with the help of a check- about what experiences the patient can
list. The information entered onto the proforma expect.
was then used to develop a summary table for 3 Negative outcome expectancy was used to
each primary research paper. The table included refer to expectancies created either when the
information of first author, title and source, practitioner conveys uncertainty or even lack
study population, sample size, study design and of faith in the procedure or when the practi-
quality rating, key features of the intervention, tioner informs the patient of the negative
the health outcomes that had been measured, consequences of treatment such as possible
and the results. side effects.
The Placebo Effect: Methodological Process and Implications 81
Individual education about the heart and the impending Process expectancy created by procedural information
cardiac catheterisation and individual training in how
Management self-efficacy created by training patients
to cope using a cognitive-behavioural intervention
in cognitive-restructuring
Diabetic algorithm used in conjunction with medical records Interaction self-efficacy created by encouraging
to teach patients how to focus on treatment issues so that patients to be involved in decision making
they could improve their information seeking skills and
negotiate treatment with the doctor
Patients were either instructed that the pill was Positive outcome expectancy created by dentist’s belief in
very effective at reducing tension, anxiety and the success of treatment (oversell condition)
pain (oversell condition) or they were instructed
Negative outcome expectancy created by the dentist’s belief
that it may be effective (undersell condition)
that treatment was not very effective (undersell condition)
either told that the drug was very effective or ing the classification of types of expectancy
that it may be effective. Instructions which said by clinical area, clinical setting, and class of
the drug was very effective were said to be determinants. Some 60% of papers were located
positive outcome expectancies, and those which in three cells when types of expectancy were
only told patients that the drugs may be effective classified by class of determinants. Three
were said to be negative outcome expectancies. main relationships were identified: management
self-efficacy and the treatment and treatment
setting (30); process expectancy and the treat-
Grouping the Studies for Analysis ment and treatment setting (24); and positive
outcome expectancy and practitioner factors
Because of the heterogeneity of the studies used (15) (Table 4).
in the review, the papers needed to be grouped Three clinical categories classified all the
so that the benefits of the placebo effect could studies: preparation for medical procedures;
be evaluated across subgroups of similar studies. management of illness; and medical treatment.
A number of strategies were employed includ- Studies in which patient expectancies were
Table 4 Frequency with which type of expectancya was associated with class of
determinant
Patient Practitioner Practitioner–patient Treatment and
expectation factors interaction treatment setting Total
Process expectancy – 2 1 24 27
Positive outcome expectancy 5 15 5 9 34
Negative outcome expectancy 2 5 3 4 14
Interaction self-efficacy – 6 1 7
Management self-efficacy 3 1 – 30 34
Total 10 23 15 68 116
a
Some investigations included more than one type of expectancy.
The Placebo Effect: Methodological Process and Implications 83
created by the treatment and treatment setting suggestion in a few of the studies that the effect
either concerned the preparation for medical of expectancy may be counteracted by person-
procedures or the management of illness. Studies ality. For the NHS, there are potential cost
in which patient expectancies were created by savings in reduced use of analgesics and reduc-
practitioner factors mostly concerned the effec- tions in the length of stay, although only one of
tiveness of medical treatments (usually drug the studies calculated the financial ramifications
therapy). of the education and training programme.
Forty studies were concerned with the man-
agement of illness, 23 of which examined
patient-centred management of chronic illness.
EVIDENCE Eighteen were clinical trials and five were obser-
vational studies. The effects of manipulating
Eighty-five studies were used to assess the role management self-efficacy either alone or in com-
of expectancies in the placebo effect. Of these bination with positive outcome expectancy or
studies, 25 concerned the preparation for medi- process expectancy are summarised in Table 5b.
cal procedures, 20 of which were clinical trials Fifteen studies examined aspects of the
and five observational studies. Using the criteria patient–provider consultation. Ten of the studies
for verifying expectancy (see earlier), there was were observational, and five were clinical trials.
evidence that process expectancy and manage- The effects of expectancies for these studies are
ment self-efficacy had either been created alone summarised in Table 5c.
or in combination. The results are summarised The remaining two of the 40 studies con-
in Table 5a. cerned with chronic illness examined patients’
Taken together, the results suggest that pro- levels of positive outcome expectations for treat-
cess expectancy, when combined with manage- ment, but were poor in quality and so no conclu-
ment self-efficacy, produces the greatest benefits sions could be drawn.
by enhancing the quality of the patients’ hospital Taken together, the results reported in the
experience. Management self-efficacy when studies concerned with the management of ill-
created alone, on the other hand, was most likely ness also show positive health outcomes. From
to reduce patient anxiety. There was also a the 23 studies which examined patient-centred
management there is the suggestion that it is the studies reviewed provide support for the hypoth-
creation of management self-efficacy that leads esis that expectancies are the mechanism by
to beneficial health outcomes. Positive outcome which placebos have their effects. However,
expectancy alone was not as successful. For the because of the heterogeneity of outcomes
NHS, the potential cost benefits are a reduction assessed and the uneven distribution of the
in the use of health services. From the 15 studies expectancies across the three clinical areas, it
which investigated the effects of creating inter- was not possible to use meta-analysis to com-
action self-efficacy there is also the suggestion bine effect sizes across studies.
of health benefits for the patients. However, the
mechanism through which these positive health
outcomes effects are achieved needs further
Quality Assessment
investigation. These conclusions must be tempered by a lack
Twenty studies concerned the outcomes of of methodological excellence characterising
medical treatment in which the effects of vari- both the clinical trials (62) and observational
ous factors on the patients’ responses, including studies (23) that were reported. Weaknesses of
their own beliefs, were reported. Eighteen were the clinical trials was their lack of methodo-
clinical trials and one was an observational logical detail, making it difficult to judge the
study, with one study providing insufficient validity of the findings. For instance, in some
information. The results are summarised in cases there was no information on whether
Table 5d. groups were comparable at baseline, so con-
Taken together, the results from the studies founding the results. Other weaknesses included
concerned with the outcomes of medical treat- a failure to ensure that the control group had not
ment show the power of positive outcome been given information similar to the treatment
expectancy to enhance the effects of medical group and a failure to use all cross-over order-
treatment. The effects of negative outcome ings in a cross-over trial. As well as their lack of
expectancy are ambiguous. methodological detail, the shortcomings of the
Given the evidence of the subjective and observational studies was their weak research
objective benefits of creating expectancy, the designs, high dropout rates, the lack of adequate
The Placebo Effect: Methodological Process and Implications 85
controls, multiple testing of hypotheses leading groups may not, therefore, show the beneficial
to statistically significant correlations that were effects which psychological health technologies
small in magnitude, a heavy reliance on retro- can have on health outcomes. Reviews of other
spective reporting, and the use of patient groups psychological health technologies which are
that were not comparable in terms of the also drawn from different clinical groups could
medical procedures they were to undergo. The contribute further understanding that could also
quality of the studies tended to show an be usefully harnessed by the NHS. We identified
improvement over time, with ones published gaps in the clinical literature; for example, we
more recently demonstrating greater rigour. It is located no studies concerned with the manage-
not surprising that studies drawn from such ment of chronic illness resulting from some
diverse literatures would be of mixed quality. of the major degenerative diseases. Further
Our conclusions must, therefore, be limited research is also needed in these and possibly
by the methodological shortcomings, and the other clinical areas as well.
scarcity of studies conducted in the UK since a The second area concerns the development of
large proportion of the studies (66%) were con- statistical methods for handling studies that are
ducted in the USA. heterogeneous with respect to outcome vari-
ables. For example, the studies we identified on
Implications for Health Service Staff preparation for medical procedures typically
and Practitioners used a measure of post-procedure recovery such
as number of pain-killing medications requested
The principal implication of the results of the over a certain period of time period, days in
review for both health service staff and practi- hospital, or self-ratings of physical discomfort.
tioners is the importance of the psychological Rarely did two studies use exactly the same
components of healthcare delivery. The most outcome measure. Therefore, our analysis was
important message coming out of the review is limited to a narrative review and we were not
the influence on health outcomes of: (a) the able to perform any quantitative meta-analysis.
information/training provided (or not provided) Methods for combining diverse outcomes that
for patients; (b) the patient’s ability to manage are essentially measuring the same effects (in
their own care; and (c) the attitudes and inter- this case patient recovery) are currently being
personal skills of practitioners during the deliv- considered by our group so that statistical meth-
ery of care. All three aspects need to be seen as ods of meta-analysis can be used to quantify
health technologies in their own right, not as findings from reviews that encompass such
incidental, non-specific aspects of the medical/ diversity as we found for the placebo effect. One
nursing care. As such, when evidence is being such statistical method has developed a tech-
sought, whether it be to inform health policy or nique for combining multiple outcomes involv-
medical decision making, the effectiveness of ing, in particular, a mixture of discrete and
these psychological interventions should be continuous responses.126 In the simplest case
included in cost–benefit analyses. these outcomes can be bivariate, comprising a
binary and continuous response as, for example,
Suggestions for Further Research the patient’s health status reported as improved/
stable or deteriorated (binary response) and their
Two particular areas for further research follow level of anxiety (continuous response). Another
from the review. difficulty which our review pinpointed was the
Firstly, there is a need for reviews of health complex two-level structure (determinants and
technologies which focus on the psychological mechanisms) of factors affecting outcomes of
as well as the physical therapies. The health the placebo effect. Statistical methods that
technologies identified from the review are the handle one level of dependence and require that
various self-management interventions used to the determinants be specified in advance as co-
help patients cope with stress in acute clinical variates are readily available.127,128 However,
studies, particularly as preparation for medical more research is needed to develop statistical
procedures, and those which help patients man- methods to handle routinely the type of data
age their illness. Insights into the benefits of encountered here.
these psychological health technologies were
drawn from studies reporting health effects from
a range of clinical areas which included illness ACKNOWLEDGEMENTS
associated with cardiac rehabilitation, diabetes,
asthma, chronic headache, chronic pain, rheuma-
toid arthritis and Parkinson’s disease. Reviews of The review of the determinants of the placebo
the literature which focus only on single clinical effect and its use in the delivery of health care
86 The Advanced Handbook of Methods in Evidence Based Healthcare
was supported by a grant from the Health Tech- 16. Elander G. Ethical conflicts in placebo treat-
nology Assessment Programme, part of the ment. J. Advanced Nursing 1991; 16: 947–51.
NHS R and D Executive. The Surrey review 17. Handfield-Jones, M. A bottle of medicine from
team would also like to acknowledge Dr I.G. the doctor. Lancet 1953; 2: 823–5.
Vlachonikolis, Reader in Medical Statistics, 18. Kluge EH. Placebos: Some ethical considera-
EIHMS contribution concerning the new devel- tions. Can. Med. Assoc. J. 1990; 142: 293–5.
opments in meta-analytical techniques. Our 19. Krouse JH, Krouse HJ. Placebo debate. Am. J.
thanks are also due to Mrs J. King for her help Nursing 1981; 81: 2146–8.
in the preparation of this manuscript. 20. Lynoe N, Mattsson B, Sandlund M. The atti-
tudes of patients and physicians toward placebo
treatment – A comparative study. Soc. Sci. Med.
1993; 36: 767–74.
REFERENCES 21. Oh VMS. The placebo effect: Can we use it
better? Br. Med. J. 1994; 309: 69–70.
1. Beecher HK. The powerful placebo. JAMA 22. Rawlinson MC. Chapter 22, Truth-telling and
1955; 159: 1602–6. paternalism in the clinic: Philosophical reflec-
2. Borkovec TD. Chapter 4, Placebo: redefining tions on the use of placebos in medical practice.
the unknown. In: White L, Tursky B, Schwartz In: White L, Tursky B, Schwartz GE (eds).
GE (eds). Placebo: Theory, Research and Mech- Placebo: Theory, Research and Mechanisms.
anisms. New York: The Guilford Press; 1985; New York: The Guilford Press; 1985; pp. 403–
pp. 59–66. 18.
3. Gotzsche PC. Is there logic in the placebo? 23. Simmons B. Problems in deceptive medical pro-
Lancet 1994; 344: 925–6. cedures: An ethical and legal analysis of the
4. Grünbaum A. The placebo concept. Behav. Res. administration of placebos. J. Med. Ethics 1978;
Ther. 1981; 19: 157–67. 4: 172–81.
5. Grünbaum A. Chapter 2, Explication and impli- 24. Bakan D. Chapter 11, The apprehension of the
placebo phenomenon. In: White L, Tursky B,
cations of the placebo concept. In: White L,
Schwartz GE (eds). Placebo: Theory, Research
Tursky B, Schwartz GE (eds). Placebo: Theory,
and Mechanisms. New York: The Guilford
Research and Mechanisms. New York: The
Press; 1985; pp. 211–14.
Guilford Press; 1985; pp. 9–36.
25. Shapiro AK. Attitudes toward the use of
6. Kienle GS, Kiene H. Placebo effect and placebo
placebos in treatment. J. Nerv. Mental Disord.
concept: a critical methodological and concep-
1960; 130: 200–11.
tual analysis of reports on the magnitude of the
26. Shapiro AK. Factors contributing to the placebo
placebo effect. Alternative Therapies in Health
effect. Am. J. Psychother. 1961; 18: 73–88.
and Medicine 1996; 2: 39–54. 27. Shapiro AK. Chapter 22, The placebo response.
7. Byerly H. Explaining and exploiting placebo In: Howells JG (ed.). Modern Perspectives in
effects. Perspect. Biol. Med. 1976; 19: 423–36. World Psychiatry. London: Oliver and Boyd;
8. Wilkins W. Chapter 6, Placebo controls and 1968, pp. 596–619.
concepts in chemotherapy and psychotherapy 28. Shapiro AK. Iatroplacebogenics. Int. Pharmaco-
research. In: White L, Tursky B, Schwartz GE psychiatry 1969; 2: 215–48.
(eds). Placebo: Theory, Research and Mechan- 29. Shapiro A, Morris L. Chapter 10, The placebo
isms. New York: The Guilford Press; 1985; effect in medical and psychological therapies.
pp. 83–109. In: Garfield S, Bergin A (eds). Handbook of
9. Wolf S. The pharmacology of placebos. Psychotherapy and Behavioural Change. New
Pharmacol. Rev. 1959; 22: 689–704. York: Wiley; 1978, pp. 369–410.
10. Pogge RC. The toxic placebo. Medical Times 30. Levine J, Gordon N. Chapter 21, Growing pains
1963; 91: 773–8. in psychobiological research. In: White L,
11. Rosenzweig P, Brohier S, Zipfel A. Pharmaco- Tursky B, Schwartz GE (eds). Placebo: Theory,
epidemiology and drug utilization. Clin. Pharma- Research and Mechanisms. New York: The
col. Ther. 1993; 54: 578–83. Guilford Press; 1985; pp. 395–402.
12. Wolf S, Pinsky RH. Effects of placebo admin- 31. Shepherd M. The placebo: from specificity to
istration and occurrence of toxic reactions. the non-specific and back. Psychol. Med. 1993;
JAMA 1954; 155: 339–41. 23: 569–78.
13. Barham Carter A. The placebo: its use and 32. Critelli JW, Neumann KF. The placebo: Con-
abuse. Lancet 1953; 1: 823. ceptual analysis of a construct in transition. Am.
14. Bok S. The ethics of giving placebos. Scientific Psychol. 1984; 39: 32–9.
American 1974; 231: 17–23. 33. White KL. Factor X. In: Epidemiology, Medicine
15. Brody H. The lie that heals: The ethics of giving and the Public’s Health. New York: 1991;
placebos. Ann. Intern. Med. 1982; 97: 112–18. pp. 150–66.
The Placebo Effect: Methodological Process and Implications 87
34. White L, Tursky B, Schwartz GE. Chapter 1, 49. Ernst E. Make believe medicine: The amazing
Placebo in perspective. In: White L, Tursky B, powers of placebos. Eur. J. Phys. Med. Rehab.
Schwartz GE (eds). Placebo: Theory, Research 1996; 6: 124–5.
and Mechanisms. New York: The Guilford 50. Peck C, Coleman G. Implications of placebo
Press; 1985; pp. 3–7. theory for clinical research and practice in pain
35. Shapiro AK, Shapiro E. Chapter 25, Patient- management. Theor. Med. 1991; 12: 247–70.
provider relationships and the placebo effect. In: 51. Richardson PH. Placebo effects in pain manage-
Matarazzo JD, Weiss SM, Herd JA, Miller NE ment. Pain Rev. 1994; 1: 15–32.
(eds). Behavioural Health – A Handbook of 52. Wall PD. The placebo effect: an unpopular
Health Enhancement and Disease Prevention. topic. Pain 1992; 51: 1–3.
1984; pp. 371–83. 53. Wall PD. Pain and the placebo response. In:
36. Wickramasekera I. Chapter 15, A conditioned Bock GR, Marsh J (eds). Experimental and
response model of the placebo effect: Predic- Theoretical Studies of Consciousness. 174th edn.
tions from the model. In: White L, Tursky B, Chichester: John Wiley; 1993; pp. 187–216.
Schwartz GE (eds). Placebo: Theory, Research 54. Voudouris NJ, Peck CL, Coleman G. Condi-
and Mechanisms. New York: The Guilford tioned response models of placebo phenomena:
Press; 1985; pp. 255–87. further support. Pain 1989; 38: 109–16.
37. Brody H, Waters DB. Diagnosis is treatment. J. 55. Voudouris NJ, Peck CL, Coleman G. The role
Family Practice 1980; 10: 445–9. of conditioning and verbal expectancy in the
38. Brody H. Chapter 3, Placebo effect: An exam- placebo response. Pain 1990; 43: 121–8.
ination of Grunbaum’s definition. In: White L, 56. Brewer WF. There is no convincing evidence
Tursky B, Schwartz GE (eds). Placebo: Theory, for operant or classical conditioning in adult
Research and Mechanisms. New York: The humans. In: Weimer WB, Palermo DS (eds).
Guilford Press; 1985; pp. 37–58. Cognition and the symbolic processes. New
39. Moerman DE. Perspectives on the placebo Jersey: Lawrence Erlbaum; 1974; pp. 1–42.
phenomenon. Medical Anthropology Q. 1983; 57. Turkkan JS, Brady JV. Chapter 18, Mediational
14: 3–19. theory of the placebo effect. In: White L, Tursky
40. Gelbman F. The physician, the placebo and the B, Schwartz GE (eds). Placebo: Theory,
placebo effect. Ohio State Med. J. 1967; 63: Research and Mechanisms. New York: The
1459–61. Guilford Press; 1985; pp. 324–31.
41. Chaput de Saintonge M, Herxheimer A. Har- 58. Benson H, Epstein MD. The placebo effect – a
nessing placebo effects in healthcare. Lancet neglected asset in the care of patients. JAMA
1994; 344: 995–8. 1975; 232: 1225–7.
42. Ernst E, Herxheimer A. The power of placebo. 59. Benson H, Friedman R. Harnessing the power of
Br. Med. J. 1996; 313: 1569–70. the placebo effect and renaming it ‘remembered
43. Vogel AV, Goodwin JS, Goodwin JM. The wellness’. Annu. Rev. Med. 1996; 47: 193–9.
therapeutics of placebo. Am. Family Physician 60. Berg AO. Placebos: a brief review for family
1980; 22: 105–9. physicians. J. Family Practice 1977; 5: 97–100.
44. Ross M, Olson JM. An expectancy-attribution 61. Cousins N. Belief becomes biology. California
model of the effect of placebos. Psychol. Rev. School of Medicine 1989; 6: 20–9.
1981; 88: 408–37. 62. Ernst E. Placebos in medicine. Lancet 1994;
45. Evans F. Chapter 12, Expectancy, Therapeutic 345: 65–6.
Instructions and the Placebo Response. In: White 63. Ernst E, Resch KL. The science and art of the
L, Tursky B, Schwartz GE (eds). Placebo: placebo effect. Curr. Therap. 1994; 10: 19–22.
Theory, Research and Mechanisms. New York: 64. Honigfeld G. Nonspecific factors in treatment: I.
The Guilford Press; 1985; pp. 215–28. Review of placebo reactions and placebo reac-
46. Kleijnen J, de Craen AJM, Van Everdingen J, tors. Dis. Nervous System 1964a; 25: 145–56.
Krol L. Placebo effect in double-blind clinical 65. Honigfeld G. Non-specific factors in treatment:
trials: A review of interactions with medications. II. Review of social-psychological factors. Dis.
Lancet 1994; 344: 1348–9. Nervous System 1964b; 25: 225–39.
47. Kleijnen J, de Craen AJM. Chapter 3, The 66. Jospe M. The placebo effect in healing. Toronto:
importance of the placebo effect: A proposal for Lexington Books; 1978.
further research. In Ernst E (ed.). Complement- 67. Liberman R. An analysis of the placebo
ary Medicine: An Objective Appraisal. Oxford: phenomenon. J. Chron. Dis. 1962; 15: 761–83.
Butterworth Heinemann; 1996; pp. 30–41. 68. Miller NE. Placebo factors in treatment: Views
48. Roberts AH, Kewman DG, Mercier L, Hovell of a psychologist. In: Shepherd M, Sartorius N
M. The power of nonspecific effects in heal- (eds). Non-specific aspects of treatment. Hans
ing: Implications for psychosocial and bio- Huber Publishers; 1989; pp. 39–56.
logical treatments. Clin. Psychol. Rev. 1993; 13: 69. Turner JA, Deyo RA, Loeser JD, Von Korff M,
375–91. Fordyce WE. The importance of placebo effects
88 The Advanced Handbook of Methods in Evidence Based Healthcare
in pain treatment and research. JAMA 1994; 271: 89. Graceley RH, Dubner R, Deeter WR, Wolskee
1609–14. PJ. Clinicians expectations influence placebo
70. Buckalew LW, Ross S, Starr BJ. Nonspecific analgesia. Nature 1985; 305: 43.
factors in drug effects: Placebo personality. Psy- 90. Grevart P, Goldstein A. Chapter 19, Placebo
chological Reports 1981; 48: 3–8. analgesia, naloxone, and the role of endogenous
71. Gryll SL, Katahn M. Situational factors contrib- opioids. In: White L, Tursky B, Schwartz GE
uting to the placebo effect. Psychopharmacology (eds). Placebo: Theory, Research and Mech-
1978; 57: 253–61. anisms. New York: The Guilford Press; 1985;
72. Letvak R. Putting the placebo effect into prac- pp. 332–50.
tice. Patient Care 1995; 29: 93–102. 91. Levine JD, Gordon NC, Fields HL. The mech-
73. Finkler K, Correa M. Factors influencing patient anism of placebo analgesia. Lancet 1978; 2:
perceived recovery in Mexico. Soc. Sci. Med. 654–7.
1996; 42: 199–207. 92. Straus JL, Von Ammon Cavanaugh S. Placebo
74. Thomas KB. The placebo in general practice. effects: Issues for clinical practice in psychiatry
Lancet 1994; 344: 1066–7. and medicine. Psychosomatics 1996; 37: 315–26.
75. Benson H. Commentary: placebo effect and 93. Ross S, Buckalew LW. The placebo as an
remembered wellness. Mind/Body Medicine agent in behavioral manipulation: A review of
1995; 1: 44–5. problems, issues, and affected measures. Clin.
76. Di Matteo MR. The physician-patient relation- Psychol. Rev. 1983; 3: 457–71.
ship: Effects on the quality of health care. Clin. 94. Butler C, Steptoe A. Placebo responses: An
Obstet. Gynecol. 1994; 37: 149–61. experimental study of psychophysiological pro-
77. Friedman HS, Dimatteo MR. Chapter 4, Patient- cesses in asthmatic volunteers. Br. J. Clin.
physician interactions. In: Shumaker SA, Schron Psychol. 1986; 25: 173–83.
95. Luparello T, Leist N, Lourie CH, Sweet P. The
EB, Ockene JK (eds). The Handbook of Health
interaction of psychologic stimuli and pharma-
Behavior Change. New York: Springer Publish-
cologic agents on airway reactivity in asthmatic
ing Company; 1990; pp. 84–101.
subjects. Psychosom. Med. 1970; 32: 509–13.
78. Rogers S. Facilitative affiliation: Nurse-client
96. Gordon EE. The placebo: an insight into mind-
interactions that enhance healing. Issues in
body interaction. Headache Q. 1996; 7: 117–25.
Mental Health Nursing 1996; 17: 171–84.
97. Kazdin AE. Therapy outcome questions re-
79. Ong LML, de Haes JCJM, Hoos AM, Lammes
quiring control of credibility and treatment-
FB. Doctor-patient communication: A review of
generated expectancies. Behavior Therapy 1979;
the literature. Soc. Sci. Med. 1995; 40: 903–18.
10: 81–93.
80. Beecher HK. Surgery as placebo: A quantitative
98. Kirsch, I. Unsuccessful redefinitions of the term
study of bias. JAMA 1961; 176: 1102–7. placebo. Am. Psychol. 1986; 4: 844–5.
81. Cobb LA, Thomas GI, Dillard DH, Merendino 99. Kirsch, I. The placebo effect as a conditioned
KA, Bruce RA. An evaluation of internal- response: Failures of the ‘Litmus Test’. Behav.
mammary-artery ligation by a double-blind tech- Brain Sci. 1991; 14: 200–4.
nic. N. Engl. J. Med. 1959; 260: 1115–18. 100. Laporte JR, Figueras A. Placebo effects in psy-
82. Johnson AG. Surgery as a placebo. Lancet 1994; chiatry. Lancet 1994; 344: 1206–9.
344: 1140–2. 101. O’Leary KD, Borkovec TD. Conceptual, meth-
83. Fessel WJ. Strategic aspects of prescription writ- odological, and ethical problems of placebo
ing. Postgrad. Med. 1981; 70: 30–7. groups in psychotherapy research. Am. Psychol.
84. Brown WA. The placebo effect. Scientific 1978; 33: 821–30.
American 1998; 278: 68–73. 102. Wilkins W. Placebo problems in psychotherapy
85. Buckalew LW, Ross S. Relationship of per- research: Social psychological alternatives to
ceptual characteristics to efficacy of placebos. chemotherapy concepts. Am. Psychol. 1986; 41:
Psychological Rep. 1981; 49: 955–61. 551–6.
86. Buckalew LW, Coffield KE. An investigation of 103. Bowers TG, Clum GA. Relative contribution of
drug expectancy as a function of capsule color specific and non specific treatment effects:
and size and preparation form. J. Clin. Psycho- Meta-analysis of placebo-controlled behavior
pharmacol. 1982; 2: 245–8. therapy research. Psychol. Bull. 1988; 103:
87. de Craen AJM, Roos PJ, de Vries AL, Kleijnen 315–23.
J. Effect of colour of drugs: Systematic review 104. Smith ML, Glass GV, Miller TL. The Benefits of
of perceived effect of drugs and of their effec- Psychotherapy. Johns Hopkins University Press,
tiveness. Br. Med. J. 1996; 313: 1624–6. 1980.
88. Jacobs KW, Nordan FM. Classification of 105. Bootzin RR. Chapter 10, The Role of Expect-
placebo drugs: Effect of color. Perceptual and ancy in Behavior Change. In: White L, Tursky
Motor Skills 1979; 49: 367–72. B, Schwartz GE (eds). Placebo: Theory,
The Placebo Effect: Methodological Process and Implications 89
Research and Mechanisms. New York: The meta-analysis. J. Clin. Epidemiol. 1990; 43:
Guilford Press; 1985; pp. 196–210. 1191–9.
106. Brody N. Is psychotherapy better than a placebo? 119. Turner JL, Gallimore R, Fox-Henning C. An
Behav. Brain Sci. 1984; 7: 758–63. annotated bibliography of placebo research. J.
107. Brown WA. Placebo as a treatment for de- Supplement Abstract Service Am. Psychol.
pression. Neuropsychopharmacology 1994; 10: Assoc. 1980; 10: 22.
265–9. 120. Bandura A. Self-efficacy: Toward a unifying
108. Dago PL, Quitkin FM. Role of the placebo theory of behavioral change. Psychol. Rev. 1977;
response in the treatment of depressive dis- 84: 191–215.
orders. CNS Drugs 1995; 4: 335–40. 121. Bandura, A. Social Foundations of Thought and
109. Prioleau L, Murdock M, Brody N. An analysis Action: A Social Cognitive Theory. Englewood
of psychotherapy versus placebo studies. Behav. Cliffs, NJ: Prentice Hall, 1986.
Brain Sci. 1983; 7: 756–7. 122. Bandura, A. Self-efficacy: The exercise of con-
110. Frank JD. Biofeedback and the placebo effect. trol. New York: Freeman, 1997.
Biofeedback and Self-Regulation 1982; 7: 123. Lorig KR, Mazonson PD, Holman HR. Evi-
449–60. dence suggesting that health education for self-
111. Kaptchuk TJ, Edwards RA, Eisenberg DM. management in patients with chronic arthritis
Chapter 4, Complementary medicine: Efficacy
has sustained health benefits while reducing
beyond the placebo effect. In Ernst E (ed.).
health care costs. Arthritis Rheum. 1993; 36:
Complementary Medicine. An Objective Ap-
439–46.
praisal. Oxford: Butterworth Heinemann; 1996;
124. Kendall PC, Williams L, Pechacek TF, Graham
pp. 42–70.
LE, Shisslak C, Herzoff N. Cognitive-behavioral
112. Lynoe N. Is the effect of alternative medical
and patient education interventions in cardiac
treatment only a placebo effect? Scand. J. Soc.
Med. 1990; 18: 149–53. catheterization procedures: The Palo Alto medi-
113. Buckman R, Lewith G. What does homeopathy cal psychology. J. Consult. Clin. Psychol. 1979;
do and how? Br. Med. J. 1994; 309: 103–6. 47: 49–58.
114. Joyce CRB. Placebo and complementary medi- 125. Greenfield S, Kaplan SH, Ware JE, Yano EM,
cine. Lancet 1994; 344: 1279–81. Frank HJL. Patient participation in medical care:
115. Kleijnen J, Knipschild P, Ter Riet G. Clinical Effects on blood sugar control and quality of
trials of homeopathy. Br. Med. J. 1991; 302: life in diabetes. J. Gen. Intern. Med. 1988; 3:
316–23. 448–57.
116. Reilly TD, Taylor MA, McSharry C, Aitchison 126. Berkey CS, Anderson JJ, Hoaglin DC. Multiple-
T. Is homoeopathy a placebo response? Con- outcome meta-analysis of clinical trials. Statist.
trolled trial of homoeopathic potency, with Med. 1996; 15: 537–57.
pollen in hayfever as model. Lancet 1986; 2: 127. Berkey, CS, Hoaglin, DC, Mostellar, F., Corditz,
881–6. CA. A random-effects regression model for
117. Smith, I. Commissioning complementary medi- meta-analysis. Statist. Med. 1995; 14: 395–411.
cine. Br. Med. J. 1995; 310: 1151–2. 128. Thompson, SG (1994) Why sources of hetero-
118. ter Riet G, Kleijnen J, Knipschild P. Acu- geneity should be investigated. Br. Med. J. 1994;
puncture and chronic pain: A criteria-based 309: 1351–5.
Part II
OBSERVATIONAL
AND QUALITATIVE
METHODS
I N T R O D U C T I O N by R A Y F I T Z P A T R I C K
As the previous section of this volume has volves data-gathering in the context of a research
demonstrated, the randomised controlled trial study established to address specific questions.
(RCT) can in principle be considered the opti- Such data form a continuum with a second type
mal study design in order to evaluate the bene- of quantitative data, that in summary terms may
fits of an intervention. A well-designed and be termed routine. In the healthcare systems of
conducted RCT comparing outcomes in two or most advanced societies, there is a wide array of
more groups of patients receiving different inter- quantitative data gathered routinely and regu-
ventions is most likely to be able to minimise larly for administrative, financial, surveillance
the biases that may arise because groups differ or other purposes rather than to address specific
in other respects such as disease severity, co- research questions. Because they have largely
morbidity or other factors that may influence not been set up specifically to address evaluative
outcomes in addition to study interventions. questions, the role of routine data in health
However, there are circumstances where an technology assessment needs to be examined
RCT may be unethical because potential partici- very carefully. The third kind of data considered
pants do not have equipoise to permit random
in this section is quite distinctive in that it is
allocation or may not be feasible because
adverse outcomes are so rare or so distant in qualitative; in other words, data which may
time from the intervention that the study costs relate to or be capable of being rendered into
become prohibitive.1 Given that the scope for quantitative form but are primarily of signifi-
RCTs to be poorly designed is potentially enor- cance as non-numerical narrative or verbal evi-
mous, it is essential to re-examine the advant- dence. Generally qualitative data most resemble
ages, disadvantages and future potential of our first type of quantitative data since they are
observational (i.e. non-randomised) evidence in invariably collected in the context of dedicated
health technology assessment. research rather than as a by-product of health
This section considers the distinctive role and service routines.
merits of three broadly distinguished types of The first two chapters of this section provide
observational evidence. Firstly, the section con- important and complementary analyses of the
siders quantitative data derived from dedicated relative advantages and disadvantages of ran-
observational studies set up especially to address domised compared with observational evidence
particular research questions. Usually this in- to evaluate healthcare interventions. Sanderson
92 The Advanced Handbook of Methods in Evidence Based Healthcare
and colleagues report a number of reviews potential relevance to health technology assess-
addressing aspects of this comparison. Firstly, ment. In the best instances it is possible to make
they examine studies on four specific and con- linkages between interventions and health out-
trasting healthcare interventions: surgery in the comes; commonly however, data on health out-
form of coronary artery bypass grafting and comes do not exist or cannot be linked informa-
coronary angioplasty, calcium antagonists, stroke tively. Raftery and colleagues point out that the
units, and malaria vaccines. They conclude that evaluation of healthcare interventions is not
differences in results between randomised and solely concerned with effectiveness. We are
observational evidence in these four fields are often concerned with other issues such as how
smaller than might be expected and show no successfully effective treatments diffuse and
consistent tendency for RCTs to report smaller how fairly and equitably between social groups
effects, as claimed by some earlier reviews.2 they are provided. Routine data sets may be a
From other evidence, Sanderson and colleagues unique resource to address such broader evalu-
conclude that many differences in results ative questions.
between the two types of studies may arise from Lewsey and colleagues also address the cen-
differences in samples recruited into studies tral question of the value of routinely collected
rather than study design. Where appropriate data in evaluative research. They examine the
adjustments are made in non-randomised studies, role of linked Scottish healthcare and health
results may more closely resemble those of status data in two case studies, one to address
randomised designs. The overall message of this questions regarding surgery for subarachnoid
chapter is to caution the reader against the haemorrhage, the other comparing coronary
instinctive assumption that observational evi- artery bypass grafting with percutaneous coron-
dence of effectiveness of interventions is inher- ary angioplasty. They conclude that limitations
ently more biased and flawed. in the quality of routinely collected data (for
Reeves and colleagues address similar ques- example in terms of diagnostic imprecision)
tions about inferences that may be drawn from make it difficult to perform appropriate adjust-
randomised compared with observational de- ments to produce unbiased estimates of treat-
signs, using data from research on effectiveness ment benefits. However, they also make a case
in other fields, for example, mammographic for the benefits that would flow from strengthen-
screening and folic acid supplementation. When ing the quality of routine data.
observational studies rated by the team as of Finally, the role of qualitative evidence in
high methodological quality were compared health technology assessment is addressed by
with results obtained from RCTs, evidence of Murphy and Dingwall. Qualitative research is a
effectiveness was found to be rather similar in relatively recent arrival to the field of health
the two types of study. Only when observational technology assessment and, as they acknowl-
studies of poorer quality were included did evi- edge, its role is more contested. In an author-
dence of effectiveness of interventions begin to itative overview, they consider why this should
diverge more significantly. Whilst they are be the case. They consider and challenge views
cautious in generalising beyond the areas of that qualitative research is inherently less gen-
healthcare captured in their literature review, eralisable and its inferences less well validated.
their conclusions are strikingly similar to those They argue that qualitative research has a vital
of Sanderson and colleagues. With appropriate role to play in the evaluation of healthcare, in
use of conventional methods of adjustment for for example, identifying mechanisms underlying
possible confounding factors, well-designed the effects of services and, more generally, pro-
observational quantitative studies can prove as viding policy-makers and other audiences with
valid estimates of effectiveness of interventions deeper understanding of the nature of healthcare
as do RCTs. organisations that deliver interventions. As the
Routine data of the kind collected by health- authors put it, qualitative methods address the
care systems for administrative or quality assur- ‘how’ rather than ‘how many’ questions that
ance purposes normally attract the most negative modern healthcare constantly raises. Currently
comments in terms of potential value in health the quality of qualitative contributions to health-
technology assessment.3 Raftery and colleagues care research is more variable, but the potential
use the diverse array of routine data available in role is considerable.4
the UK national health service as a test case to There have been long-standing debates about
reassess the value of such data. Their review the most informative and most creative study
demonstrates what is likely to be the case for designs to examine causal relationships both in
many other healthcare systems, which is that the wider field of human behaviour, and more
routine data vary very considerably in their narrowly in the evaluation of the impact of
Part II: Observational and Qualitative Methods 93
health services.5,6 The lesson of this section is affects outcomes in comparisons of therapy: I
that evidence must replace dogma in choice of Medical. Statist. Med. 1989; 8: 441–54.
study design to optimise the value and impact of 3. Sheldon TA. Please bypass the PORT. Br. Med. J.
data to inform decisions about health services. 1994; 309: 142–3.
4. Boulton M, Fitzpatrick R, Swinburn C. Qualitative
research in health care. II A structured review and
evaluation of studies. J. Eval. Clin. Pract. 1996; 2:
REFERENCES 171–9.
5. Cook T, Campbell D. Quasi-experimentation
1. Black N. Why we need observational studies to Design and Analysis Issues for Field Settings.
evaluate the effectiveness of health care. Br. Med. Boston: Houghton Mifflin, 1979.
J. 1996; 312: 1215–18. 6. Cochrane A. Effectiveness and Efficiency. London:
2. Colditz G, Miller J, Mosteller F. How study design Nuffield Provincial Hospitals Trust, 1972.
6
Randomised and Non-Randomised
Studies: Threats to Internal and
External Validity
C O L I N S A N D E R S O N , M A R T I N MC K E E ,
A N N I E B R I T T O N , N I C K B L A C K , K L I M MC P H E R S O N
and C H R I S B A I N
‘life’); and that ethical recruitment is impossible conscious or otherwise – may creep into the
when individual clinicians and/or patients do not allocation process, with those receiving the new
accept that there is sufficient uncertainty about treatment tending to have greater underlying
the effectiveness of alternative treatments. Even potential to benefit. The first question is, what
where randomisation is practical, there can be evidence is there for this?
problems with internal validity if the process of
randomisation affects outcome in different ways
for different treatment groups, or when subjects Selection of Subjects and
have preferences and cannot be effectively External Validity
blinded to the intervention;2 and there can be
problems with external validity if the process Evaluation studies vary widely in the extent to
through which patients become involved in ran- which the range of potential future recipients of
domised studies is highly selective.3 Also in treatment is represented. Some groups of sub-
practice, randomised trials may have been less jects may be excluded as part of the design for
free from bias than many have supposed, with medical reasons (a high risk of adverse effects in
systematic differences arising from faulty ran- relation to expected benefits; a belief that benefit
domisation and incomplete blinding.4,5 will be relatively small or absent – or has
Thus, it is contended, there can be an import- already been established for the group in ques-
ant complementary role for non-randomised tion). Exclusions of this kind can be problematic
studies, provided that the same attention is paid if there is a tendency towards, say, a more
to data collection processes as in good random- cautious attitude to risk of adverse effects in
ised studies, and that adequate steps can be trials than in regular clinical practice.
taken to avoid or allow for material baseline Subjects may also be excluded for scientific
differences. Such studies have the potential, at reasons (increased precision in estimates of
least, for less selective recruitment and more effect by studying relatively homogeneous
rapid accumulation of subjects at lower admin- groups; reduced vulnerability to bias by exclud-
istrative cost – or greater sample size. The ing subjects who may be at high risk of failure
randomisers’ riposte is that external validity is to comply with treatment, or loss to follow-up9).
pointless without internal validity, that one can The pursuit of homogeneity can lead to exclu-
never be sure that adequate steps have been sion of large groups of people who are quite
taken to deal with baseline differences, that plausible candidates for treatment.
selective recruitment does not seriously bias A further threat to the external validity of
measures of relative effect, and that pragmatic both randomised and non-randomised studies
randomised designs give as much external valid- comes from selective participation, the cumu-
ity as is necessary. lative effect of decisions by centres, practi-
This debate has been characterised by tioners and eligible patients about whether to
strongly held views based on limited empirical take part.
evidence. In this review we seek to define some This suggests a second set of questions: what
of the unresolved questions; to examine the evidence is there on the contribution of eligibil-
evidence in support of the differing positions; ity criteria and selective participation to differ-
and to suggest priorities for further research. ences between randomised and non-randomised
The material presented here is a summary of a studies in measured relative effect? And if the
report that addresses these issues in more evidence is sparse – which it turns out to be –
detail.6 what is known about differences between these
study designs in terms of: (i) the extent and
nature of the selection involved; and (ii) subjects
at baseline?
DEFINING THE QUESTIONS
One test of this is whether, given similar eligi- were followed-up. As published evidence was
bility criteria, adjustment for confounding brings lacking, data from two recently completed sys-
the results of non-randomised studies closer to tematic reviews that included both RCTs and
those from RCTs; and if the answer is ‘some- non-randomised studies17,18 were reanalysed to
times’, then under what circumstances? assess the effect of differential participation by
teaching and non-teaching facilities.
Patient Preferences In addition, four case studies were under-
taken, examining the results of RCTs and non-
The idea that randomisation itself can give rise randomised studies of interventions represent-
to biased results about outcome may seem sur- ing four different modes of therapeutic action:
prising. However, when treatments are allocated coronary artery bypass grafting (CABG) and
randomly, practitioners and patients are percutaneous transluminal coronary angioplasty
deprived of expressing a preference and, if (PTCA) (surgical); calcium antagonists (phar-
choice and control are of therapeutic benefit, macological); stroke units (organisational); and
then a blinded randomised comparison might malaria vaccines (preventive). Of necessity, the
provide a misleading estimate of the relative overview provided here is a brief summary of
effect of treatment in ‘regular’ practice.10 The the evidence examined, and a full review pro-
difficulty is that if such preference effects exist, vides more details.6
attempts to detect them are compromised by Our thinking about external validity has been
potential confounding by selection:11 people informed by a simple conceptual model (Figure
who tend to prefer something may well tend to 1). The horizontal axis represents an indicator or
be different in other ways, plausibly related to predictor of potential benefit from treatment A
prognosis.12,13 There are several possible mech- relative to treatment B, such as blood pressure
anisms for patient preferences having an effect. or symptom score. The vertical axis represents
These have been examined in research on the numbers of people. The population is rep-
psychology of the placebo14 as well as in studies resented by an outer triangular ‘envelope’, of
showing that the extent of a person’s social or which only the right-hand part is shown; most of
professional control over their lives has a role in the population are to the left of the broken line.
the aetiology of coronary heart disease.15 Where This represents schematically the common situ-
people have strong preferences, randomising is ation in which there are relatively few people
difficult.16 Thus, an important and neglected part with the highest levels of potential benefit and
of the research agenda is to disentangle the main increasing numbers with lower levels. In reality
physiological effect of a treatment from any the envelope will take a variety of shapes
possible preference-dependent effect. depending on the intervention being studied.
In summary, the questions that will be For a substantial proportion of the population
addressed in this review are first, ‘do non- there will be little doubt that the expected poten-
randomised studies differ from RCTs in meas- tial advantage of treatment A over treatment B is
ured magnitude of treatment effect?’, and if they outweighed by risks or other ‘costs’. These,
do, ‘to what extent can these differences be indicated as me (medical exclusions), should be
explained by: excluded by design – from randomised studies
as ineligible, and from non-randomised studies
(a) selective recruitment; as inappropriate. (The diagram also shows an
(b) baseline differences between groups in ‘upper’ threshold, above which there is little
non-randomised studies (and can these be doubt that benefits outweigh risks.) That leaves
allowed for); and the grey band in the diagram, indicating the ref-
(c) the differential impact, in a randomised erence population for which results are needed.
study, of any therapeutic effect arising However, as we have seen, some subjects (se:
from patient participation in treatment scientific exclusions) may be ‘designed out’ to
decisions?’ improve precision or reduce bias. Then, eligible
patients may not participate for a variety of
reasons: non-participation in study by treatment
METHODS centre or doctor (d); patient not invited to par-
ticipate (i); patient choice not to participate (p).
The search strategy used to address the four The study population becomes a progressively
questions identified above has been described in smaller subset of the reference population, with
detail elsewhere6 but, in brief, it involved sear- opportunities for recruitment bias arising at
ches of Medline (1966–1996) and EMBASE each stage. There may also be some deliberate
(1980–1996) using, among others, variants on sampling process, but as random sampling, at
the words participation, inclusion, exclusion, least, should be ‘neutral’, this is not shown. In
recruitment and preference. Cited references the diagram the distributions of (and hence
Randomised and Non-Randomised Studies: Internal and External Validity 99
mean values of) potential benefit in the study because different papers involved different types
and reference populations are different, sig- of comparison; but in a very crude analysis of
nalling the question of whether it is valid to differences in effect, seven papers reported
generalise from one to the other. larger effect sizes in non-randomised studies,
but in only three of these was the difference
significant (or test-based conclusions from the
THE EVIDENCE two types of study different) at the 5% level; in
three, differences were not significant, and in
Do Non-Randomised Studies and RCTs one, significance was not reported. Eleven
papers reported larger effect sizes for RCTs,
Produce Systematically Different but only four of these were significant; six
Results? were not significant, and one did not report
significance.
The proposition that non-randomised studies
‘produce’ greater treatment effects than random- ● For CABG versus PTCA, the RCTs tended
ised studies appears to be based on two kinds of to favour CABG although confidence inter-
evidence. The first consists of a few frequently vals were wide; unadjusted results of the
quoted comparisons of the two methods. The non-randomised studies favoured PTCA.
second comes from studies of RCTs with vary- ● For calcium antagonists, effects for ran-
ing quality of randomisation, which have shown domised studies (overall risk ratio for nine
that inadequate or unclear methods are asso- RCTs: 1.10 [0.95 to 1.27]) and non-
ciated with exaggerated intervention effects. randomised (unadjusted RR for one study:
In this review, 18 papers were identified that 1.18 [1.04 to 1.34]) were similar.
directly compared the results of RCTs and ● For stroke units versus conventional care,
non-randomised studies of the same interven- the Cochrane meta-analysis for RCTs gave
tion.12,19–35 The results are summarised in Table OR 5 0.81 (0.68 to 0.96) for survival to final
1. The table is difficult to summarise further follow-up (median 12 months) as against 0.6
100
Table 1 Comparison of effects from RCTs and non-randomised studies
Randomised studies Non-randomised studies
Same Different
eligibility Adjustment Treatment Outcome relative relative effect size
McKay et al. 12 Alcohol yes yes day care % days . 3 drinks 24 20.0% 0.36 ? 65 35.1% 0.90 ? ,0.05
(1995) rehab inpatients at 1 yr 24 55.0% 31 39.1%
day care % treated again 24 15.0% 0.50 ? 65 12.3% 0.40 ? (ns)
inpatients in rehab at 1 yr 24 10.0% 31 30.4%
CASS 19 CHD yes no Surgery 5-yr survival 390 95.0% 1.03 ns 570 94.0% 1.02 ? ns
(1984) Medical 390 92.0% 745 92.0%
Hlatky et al. 20 CHD yes yes Surgical 5-yr survival 686 83.0% 1.064 ? 719 85.5% 1.057 ? ns
(1988) Medical 78.0% 80.9%
CHD yes yes Surgical 5-yr survival 767 92.0% 1.10 ? 512 91.9% 1.06 ?
Medical 84.0% 86.3%
Horwitz et al. 21 Myocardial yes no Beta-blockers Mortality 1916 7.3% 0.79 ? 417 7.2% 0.67 ? ns
(1990) infarction no BB at 24 m 1912 9.2% 205 10.7%
yes yes Beta-blockers Mortality 1916 7.3% 0.79 ? 417 7.6% 0.78 ?
no BB at 24 m 1912 9.2% 205 9.7%
Paradise et al. 22 Throat yes no Tonsillectomy Infections/pers 38 1.24 na 0.001 44 1.77 na 0.04 ?
(1984) infection Control in yr 1 35 3.09 34 3.09
Paradise et al. 23 Otitis yes no Adenoidectomy % of days with 52 15.0% 0.53 0.04 47 17.8% 0.76 0.55 sdtr
(1990) media Control otitis media in y1 47 28.5% 67 23.3%
Schmoor et al. 24 Breast yes yes chemotherapy treatment 289 ? 0.90 0.45 176 ? 0.90 ? 0.99
(1996) cancer control effect ? ?
hormone therapy treatment 184 ? 0.75 0.085 71 ? 0.53 ? 0.20
control effect ? ?
yes yes radiotherapy treatment 199 ? 0.79 0.32 129 ? 0.76 ? 0.94
control effect ? ?
Yamamoto et al. 25 Peptic yes no Eder-Puestow recurrence of 16 69.0% 0.86 ,0.05 58 88.0% 0.94 .0.05 nsdtr
(1992) strictures Balloon dysphagia 15 80.0% 34 94.0%
yes no Eder-Puestow % needing 16 38.0% 1.41 ,0.05 58 43.0% 0.86 .0.05
Balloon redilation 15 27.0% 34 50.0%
Nicolaides et al. 26 Pregnancy yes no cvs spontaneous 250 1.2% 0.20 ,0.05 320 3.1% 0.61 ns sdtr
(1994) amino loss 238 5.9% 493 5.1%
Emanuel 27 Terminal no no hospice cost/patient 247 $16,000 na ? 5853 $7,719 na ? ?
(1996) cancer conventional $15,493 $11,729
Garenne et al. 28 Measles no no vaccine measles 740 ? 0.97 ? 1224 ? 0.93 ? ns
(1993) none absent 348 ? 4403 ?
Antman et al. 29 Sarcoma yes no chemotherapy disease-free 20 75.0% 1.03 0.81 21 61.9% 1.39 0.001 sdtr
(1985) control at follow-up 22 72.7% 27 44.4%
yes yes chemotherapy disease-free 20 ? ? 0.26 21 ? ? 0.03
control at follow-up 22 ? 27 ?
Shapiro et al. 30 Breast no no chemotherapy % with acute ? ? 24.0 ,0.01 ? ? 8.10 ,0.01 nsdtr
(1994) cancer surgery non-lymphocytic ? ? 2.6 ns ? ? 1.40 ns
radiation leukaemia ? ? 10.3 ,0.01 ? ? 3.70 ,0.01
Jha et al. 31 Cardio- no yes vitamin E % CVD mortality 14564 5.0% 0.98 ns 11342 5.2% 0.61 ,0.05 sdtr
(1995) vascular no vitamin suppl per 10000 yrs fu 14569 6.0% 75903 8.5%
Pyorala et al. 32 Crypto- no no LHRG descent of 872 21.0% 1.11 ns 2410 47.0% 1.42 ,0.05 sdtr
(1995) orchism hCG testes 19.0% 33.0%
RMTIG 33 Spontaneous no no immuno live birth rates 240 61.7% 1.19 0.024 877 59.0% 1.07 0.216 sdtr
(1994) abortion control 209 51.7% 256 55.1%
Watson et al. 34 Infertility no no hysterosalpin pregnancy 287 30.3% 1.62 ? 1072 41.8% 1.61 ? ns
(1994) control rate 513 18.7% 734 25.9%
Reimold 35 Atrial yes no quinidine sinus rhythm 372 69.4% 1.54 ,0.00 471 44.3% 1.26 ,0.05 ns
(1992) fibrillation control at 3 m 354 45.1% 290 35.1%
quinidine sinus rhythm 372 50.2% 2.03 ,0.00 471 13.7% 1.26 0.29
control at 12 m 354 24.7% 290 10.9%
(0.4 to 0.9) and 0.59 (0.42 to 0.84) for the whole recruitment process) on measured effects.
two unadjusted non-randomised studies. The direct evidence is very limited. We found
● For malaria episodes within 12 months of only one paper, by Horwitz et al.21 in which the
vaccination, a meta-analysis of five RCTs relative effectiveness of two treatments found in
gave an odds ratio of 0.62 (0.53 to 0.71) as a randomised trial was compared with results
against 0.78 (0.52 to 1.17) for a non- from non-randomised cohorts subject to first
randomised study. some, and then all, of the same eligibility cri-
In summary, contrary to the proposition, we teria. Their findings are summarised in Table 2.
found no evidence for a consistent pattern or It shows that the effect estimate for the trial was
bias when the results of RCTs and non- closer to that for a ‘restricted’ observational
randomised studies were compared. Some will cohort (who satisfied all the trial’s eligibility
find this result surprising, but there are a number criteria) than that for an ‘expanded’ observa-
of possible explanations: tional cohort (with the trial’s age and diagnostic
criteria, but without its contra-indications and
1 The biases that the RCT is designed to guard factors designed to enhance follow-up).
against are in practice too small to have been Results from the study by Ward et al.36 may
detected in this analysis, and have been throw some light on this because although they
swamped by ‘random’ variation or hetero- did not address relative effectiveness, their trial
geneity. of adjuvant chemotherapy for stomach cancer
2 Most of the RCTs in this study were of such found no significant effect. They compared 960
poor quality that in practice the differences cancer registry patients with 493 trial-eligible
in study design became blurred. patients and 249 trial patients, and found median
3 There are various biases involved in non- survivals of 9, 11 and 13, months respectively.
randomised studies, but they tend to cancel Some indirect light is thrown on this issue by
out. the growing interest in heterogeneity of trials
All of these explanations may have a part to that has accompanied meta-analyses. Thompson
play. If the last is a significant one, the implica- for example has shown a relationship between
tion is that although non-randomised designs measured risk reduction and average age at
may be unbiased, their results may be subject to coronary event for subjects in different trials.37
larger mean absolute error. However, heterogeneity is a product of the
whole process of trial design and recruitment,
To What Extent can Differences not just of variations in eligibility criteria.
In terms of extent of selection, very little is
between the Results of Randomised and known about non-randomised studies, even
Non-Randomised Studies be Explained though they are (at least potentially) less vulner-
by Differences in Eligibility Criteria? able to selective recruitment than RCTs. This
may be at least partly because they have been
The key question here is the possible impact of widely seen as so flawed as to be of little interest
eligibility criteria specifically (as opposed to the to methodologists. For RCTs, the impact of
BB no BB BB no BB BB no BB
eligibility criteria on numbers recruited appears RCTs have ‘blanket’ exclusions, such as
to vary widely. For example, one of the early women,42 the elderly43 and ethnic minorities, for
studies of losses prior to randomisation38 was reasons that are often unspecified. For example,
based on 16 trials in the 1979 inventory of the Gurwitz et al. reviewed 214 RCTs of specific
National Institute of Health. Overall, 73% of pharmacotherapies for treatment of acute myo-
those screened were deemed ineligible, the rate cardial infarction between 1960 and 1991, and
ranging from 10% to 99% for different studies. found that over 60% formally excluded people
In a more recent overview of eight trials of aged 75 or more. However, explicit exclusion
intravenous thrombolysis by Muller and Topol,39 from trials on the basis of age seems to be less
the percentage of patients screened that was common in cancer trials.44
deemed eligible ranged from 9% to 51%. In summary, high levels of exclusions can
These results are difficult to interpret because bring some benefits for the researcher in terms
much of this variation may be attributable to of precision and risk of bias, as well as the
differences in use of the word ‘screening’. It medical benefits of avoiding potential harm, but
may, for example, mean surveillance of all hos- there are also important disadvantages:
pital admissions to provide possible candidates (a) uncertainty about extrapolation of results
for proper diagnostic screening rather than diag- to other populations;
nostic screening itself. This is illustrated by (b) denial of potentially effective treatment to
Table 3, which draws largely on material from those who might benefit; and
our case studies. The generally low proportions (c) delay in obtaining definitive results and/or
excluded from vaccine trials in which there was the likelihood of added research costs
no sickness- or risk-related selection prior to because of reduced recruitment rate.
screening, and very high proportions for primary
prevention of hypertension in which those
screened were again whole populations, are To What Extent can Differences
unsurprising. More striking are the fairly high between the Results of Randomised and
proportions excluded from a set of secondary Non-Randomised Studies be Explained
prevention trials in which screening was typic- by Selective Participation?
ally based on chest pain on hospital admission,
and for cardiac surgery, in which those screened Again, there are very few data on the impact of
were patients with multiple-vessel disease. selective participation on measured effect in
The relative importance of exclusions for randomised and non-randomised trials. There
scientific reasons (to improve precision and/or have, however, been some studies of the extent
reduce vulnerability to bias) rather than medical and nature of the selection involved.
ones (to focus on groups without contraindica- In the 1979 National Institutes of Health
tions and for whom outcome is uncertain) is study, in addition to the 73% deemed ineligible,
unclear, partly because of lack of data, but partly 15% were not randomised (4% withdrawn by
because for many common exclusion criteria the doctor, 4% patient refusals, and 7% with-
both medical and scientific factors are involved. drawn by the investigators),38 leaving 12% of
For example, people with serious co-morbidities those screened participating in studies.
may be excluded not only because the expected A few reports of individual studies have given
balance of benefit and risk is adverse, but also the proportions of patients under treatment not
because they may not survive to the trial end- recruited for different reasons: Lee and Breuax:
point, and they may add to the heterogeneity of 34% excluded, 29% clinician refusal, 5% patient
likely treatment effect. refusal;45 Begg et al.: 70% excluded, 8%
Those with severe disease40 or strong prefer- clinician refusal, 3% patient refusal;46 Martin
ences are also commonly excluded,41 and many et al.: 69% excluded, 15% clinician or patient
104 The Advanced Handbook of Methods in Evidence Based Healthcare
refusal;47 EORTC: 60% ineligible, 4% patient than those in the Duke non-randomised study
refusal.48 If at all representative, these figures and the nine RCTs .
suggest that doctor or patient preferences are Overall, the amount of direct evidence on
less of a barrier to recruitment than eligibility the effects of selective participation was very
criteria. limited. However, it does seem that where there
Of course studies of participation which take are differences between the results of random-
patients under treatment in participating treat- ised and non-randomised studies, explanations
ment centres or hospitals as their denominator should be sought in differences in the popula-
reveal nothing about selective participation by tions participating, as well as in the method of
centres. In a reanalysis of two recently com- allocating subjects.
pleted systematic reviews17,18 it was found that
non-randomised studies were more likely to be
multicentre than RCTs, with some weaker sup- To What Extent can Differences in
port for the proposition that non-randomised Results be Attributed to Baseline
studies were more likely to extend beyond
Differences Between Intervention
teaching or specialist centres.6
There is some evidence for socio-demographic Groups, and be Allowed for?
differences between those who participate in
trials and those who do not. Earlier reviews There are a number of theoretical objections to
found that subjects of RCTs that evaluated treat- relying on adjustment to deal with baseline
ments tended to be less affluent and less edu- differences. Particular problems arise where
cated than those who declined to participate, large numbers of factors contribute to the risk of
whilst those in RCTs evaluating preventive an adverse outcome.53 Many of these factors
interventions tended to be more affluent, and may be unknown, stratified analysis may fail
better educated.49 Rochon et al. have shown that because of small numbers in each stratum, and
although 64 out of 83 RCTs of non-steroidal adjustment models may be vulnerable to high
anti-inflammatory drugs had no specified upper levels of correlation between treatment and
age limit, only 10% of the 9664 subjects were prognostic factors.
reported as being aged 50 or more.50 The papers reviewed again gave mixed
In our own review of 16 RCTs of treatments, results. Horwitz et al., in the Beta-Blocker Heart
we also found studies in which patients were Attack Trial,21 found that adjustment of non-
younger (significantly so in four studies out of randomised findings for age and severity, when
the 11 in which age was examined), more likely combined with exclusion of subjects that did not
to be male (five studies out of 12), non-white meet the corresponding trial eligibility criteria,
(three out of seven) and of lower socio- gave similar results to their randomised trial (see
economic status (three out of four). The three Table 2). However, the authors of a paper evaluat-
studies that were biased in favour of younger ing treatment for sarcoma were unable to recon-
males all involved post-infarction anticoagulants cile the results of the different approaches.30
or anti-arrhythmics. By contrast, in one RCT of In our four case studies:
diabetic care51 the trial subjects tended to be
older than the trial-eligible patients in a primary 1 For CABG versus PTCA, there were marked
care database. baseline differences between treatment
In our review the four prevention trials all groups in the non-randomised studies. Ad-
addressed the issue of age. Subjects were rela- justment improved the consistency of the
tively young in one. Of the three that considered ‘dose–response’ relationship between base-
gender selection, the excess of males was sig- line severity and relative mortality in the
nificant in one and borderline in another. non-randomised studies (low-risk: PTCS
There have also been studies of whether base- better, high-risk: CABG better), but did
line severity or prognostic indicators were similar nothing to improve agreement with RCTs.
for subjects in RCTs and non-randomised studies. This is an area in which results may well be
In our review, 14 treatment trials examined this sensitive to selective recruitment.
in one way or another. In seven trials the partici- 2 For the calcium antagonist non-randomised
pants tended to have the poorer prognoses, but study, the patients were generally similar at
in one (non-Hodgkin’s lymphoma) prognoses baseline, apart from there being more hyper-
were better. The prevention trials were more tensives in the treatment group and more
mixed in this respect. patients on beta-blockers and digoxin among
In our case-study of CABG versus PTCA, the controls. The effect on the crude result
patients in the non-randomised Medicare study52 (1.18; 1.04 to 1.34) of adjustment for base-
had markedly higher baseline risks of mortality line differences was progressive as more
Randomised and Non-Randomised Studies: Internal and External Validity 105
The implications for external validity of selec- (b) the steps they have taken to ensure that the
tive participation by treatment centre are unclear. study population is representative of this
Strategies have been developed to evaluate the wider population in relevant respects; and
effect of individual institutions in multi-centre (c) information on how study and reference
trials,63,64 and these could be instructive if used populations differ.
more widely. Further, until these issues are better understood
Meanwhile, this review confirms the proposi- it would be desirable for authors to describe:
tion that participants in RCTs can be highly
selected groups that have had to pass through a (a) the characteristics of those centres partici-
series of hurdles of eligibility, invitation and pating, and any that declined; and
decision to participate. Thus clinical decision- (b) the numbers and characteristics of those
makers, and also policy-makers, need to: eligible to be included, those who were not
invited, and those invited who declined.
(a) pay particular attention to eligibility cri-
teria, and any analyses of the character- There will be advantage doing this for both
istics of those in the trial as compared to RCTs and non-randomised studies.
those who were eligible and to any more Effective implementation of the CONSORT
broadly defined reference population; and statement,67 which requires authors to report the
(b) be aware of the distinction between efficacy number in the eligible population and the num-
studies (of effectiveness under ideal cir- ber not randomised for each reason, will partly
cumstances) and pragmatic studies (of address this issue. The statement does not, how-
effectiveness in ‘normal’ practice); and ever, require information on the proportion of
base their decisions on pragmatic studies the treated population screened, and the charac-
where possible. teristics of those included and excluded.
One obstacle to explicit assessment of rep-
resentativeness will be the lack of information
IMPLICATIONS FOR RESEARCHERS on the spectrum of potential candidates for treat-
ment. Involving more ‘normal’ treatment set-
Exclusive versus Inclusive Trials tings in trials would help, although extraneous
factors, such as competition between profession-
The pros and cons of restrictive eligibility cri- als or hospitals can work against this; in the
teria for trials need to be more widely debated and USA, research is increasingly concentrated in
understood. Commentators such as Chalmers the less competitive parts of the healthcare mar-
have long been arguing for less restrictive ket as pressures from healthcare funders squeeze
trials.65 Unless subgroups are pre-specified, less out research.68
restrictive trials provide less precise information
on a wider range of patient groups rather than Improving External Validity
more precise information on one group and none
at all on others. The irony seems to be that if The evidence concerning behaviour change,
heterogeneity is a real problem, extrapolation either lay or professional, suggests that it may
from one subgroup to a wider population may be be difficult to overcome the failure (whether
particularly hazardous, and restrictive trials are deliberate or unintended) to invite certain centres
of limited relevance. Meanwhile, Yusuf et al. or subjects to participate, or to obtain consent.
among others have argued that worthwhile gains Any strategy must address not only the lack of
in homogeneity require good predictors of treat- recognition among both practitioners and the
ment effect, and that since these are seldom public that patients often benefit simply from
available, in practice restricting eligibility does being in RCTs and that a new treatment may be
little to improve precision.66 harmful instead of beneficial,69 but also attitudes
and practices. It should take into account the
emerging body of research on patients’ expecta-
Assessment of External Validity tions of researchers and how recruitment might
be improved.70–72 In particular, where there is
One immediate priority is to improve the quality
opposition to randomisation from practitioners,
of reporting evaluative research. Information on
methods such as that advocated by Zelen (Figure
eligibility and selective participation is often
2), in which subjects are randomised before
lacking in spite of the evidence that recruitment
being asked to give consent, should be con-
rates vary widely. As a minimum, authors
sidered.73 This has been shown to increase par-
should be required to state:
ticipation by practitioners in certain circum-
(a) the reference population to whom their stances, although it is difficult to predict how
results may be applied; many subjects will be recruited to each category
108 The Advanced Handbook of Methods in Evidence Based Healthcare
and thus to estimate how long it will take to by increasing the number of variables adjusted
accumulate a given sample size. The approach for, but often sample sizes were too small for
may also be cause for ethical concern.74 confidence in the findings.
Evidently, adjustment methods should be
developed in a rigorous fashion, based primarily
Dealing with Baseline Differences in on an understanding of the biological, clinical
Non-Randomised Studies and social, as well as the statistical issues
involved. A range of precautionary measures is
The proposition that in the past, imperfections in suggested by Concato76 (see Table 5).
the randomisation process75 have resulted in Certain key messages emerge:
baseline differences in RCTs is increasingly 1 heterogeneity among studies, both in terms
accepted by researchers and, in major trials at of the populations and the interventions
least, such imperfections are being rectified. studied, should be addressed explicitly.
Thus, the focus here is on non-randomised 2 Where there is a difference in effect size
studies. between randomised and non-randomised
The theoretical superiority of RCTs in dealing studies, differences in the study population
with unknown confounders is clear, but how or lack of power should be considered as
effective can adjustment be in dealing with well as differences in treatment allocation
unbalanced treatment groups? The evidence procedures.
reviewed is not sufficient to provide an answer.
One test would be progressive convergence with Any differences in the measurement of effect
the results of RCTs, but this requires the sub- size by the two methods are likely to reflect the
jects recruited to non-randomised studies to be interaction of several different factors. The more
similar at baseline to those in RCTs, and in most obvious threats to validity, and some of the
of the studies reviewed they were not. Some accepted counter-measures have been set out in
papers have shown how results can be affected Table 6.
and/or asymmetric, but these effects are poorly 2 If non-randomised studies are less vulner-
understood. able in this respect, this would suggest a
The physiological effect can be estimated potential complimentary role for such studies,
from the randomised arm, but with an unknown even where RCTs are ethical, feasible and
preference component. A preference effect affordable.
might thus be estimated from comparison of the
results from the two arms, but the error struc- To clarify this issue, the first step for researchers
tures are formidable.78 would be to define reference populations more
It may be easier to put into practice the trial explicitly. Studies could then be carried out for
design described by Brewin and Bradley79 different conditions (heart disease, cancer, etc.),
(Figure 4), but this will produce results for different types of intervention (primary and
which a preference effect cannot be disentangled secondary prevention, pharmaceutical, surgical,
from the possible confounding arising from organisational, etc.) and different types of study
differences between patients with strong treat- (RCT, non-randomised) to establish the impact
ment preferences. Alternative methods13 involve on measured effect of selection at each ‘stage’
recording preferences before randomisation as a in Figure 1. Ideally, all the categories of
co-variate and estimating a preference effect by patient would be followed and their outcomes
including a term in a regression equation esti- compared.
mating the main effects. Unless the trial is Those commissioning research could have
extremely large, such estimates will probably be an important role in encouraging organisers of
too imprecise to distinguish them reliably from evaluative studies to gather and report such
the main physiological effects. information.
Understanding Heterogeneity in
AREAS FOR FUTURE RESEARCH Study Findings
analysis would have to be based, for each cri- What, If Any, is the Effect of Patient
terion, on the most restrictive range used in any Preference?
of the studies involved, and large numbers of
cases would be excluded. However, it would The case for patient preferences affecting thera-
provide some impression of the contribution of peutic efficacy is theoretically plausible. How-
differences in eligibility criteria to differences in ever, even if such effects exist, finding the
measured effect. empirical evidence will require a major effort,
and interested researchers are in the common
bind of there being insufficient evidence for
The Role of Non-Randomised Studies their importance to convince funders of the need
for the complex trials needed to establish their
There will continue to be situations in which importance.
RCTs cannot be carried out, in particular where There is, however, some scope for greater use
the ethical concerns are beyond remedy. It is in of preference trials. With this design, patients
these situations that a good understanding of are carefully and fully informed about the rele-
when, why and how non-randomised studies can vant scientific uncertainties and encouraged to
be misleading would be invaluable. actively choose their treatment. Those with no
It has been suggested earlier that non- clear preference for treatment are encouraged to
randomised studies could also have an important accept randomisation. The systematic follow-up
role to play if they turn out to be less vulnerable of such cohorts would offer the opportunity to
than RCTs to problematic selective recruitment. establish through randomisation the physio-
Another requirement for this is that problems of logical effects of treatment among those with
allocation bias can be resolved or managed. no preference, and to learn whether patients
The literature reviewed here does not support with apparently similar prognostic character-
the contention that RCTs tend to produce istics who actively choose their treatments
smaller effects than non-randomised studies, but have different outcomes than those predicted by
margins of statistical error were large and the randomisation.
evidence not very discriminating. Even if they
are not biased overall, non-randomised studies
may be subject to larger absolute errors than
RCTs, but these should be reduced if the ACKNOWLEDGEMENTS
same standards of data collection, blinding and
follow-up are applied. This, of course, may We are grateful to Simon Thompson, Tom
create barriers to recruitment, and such studies Jefferson and Diana Elbourne for helpful advice
would begin to lose their advantage unless it is during the project.
randomisation per se that is the prime cause of
problematic recruitment in RCTs.
The effectiveness of adjustment in dealing
with allocation bias was difficult to judge from REFERENCES
studies comparing RCTs with adjusted results
from non-randomised studies because of differ- 1. Black N. Why we need observational studies to
ences in subject selection. There is a need for evaluate the effectiveness of health care. Br. Med.
more studies like that of Horwitz et al.,21 in J. 1996; 312: 1215–18.
which adjustment was made after applying con- 2. Black NA. The relationship between evaluative
sistent selection criteria. It should be possible to research and audit. J. Publ. Health Med. 1992;
do this retrospectively on existing datasets. 14: 361–6.
Where there are likely to be difficulties with 3. Tools for effectiveness research. In: Identifying
an RCT due to selective or slow recruitment, health technologies that work. Office of Tech-
one approach that has been adopted in the past is nology Assessment, Washington DC, 1994.
to nest the RCT within a prospective cohort 4. Schulz KF, Chalmers I, Hayes RJ, Altman DG.
study, ideally using the same participating Empirical evidence of bias: dimensions of meth-
centres and methods of data collection, ensuring odological quality associated with estimates of
compliance, blinding, follow-up, etc. If the treatment effects in controlled trials. JAMA 1995;
results for the subset of the cohort that match the 273: 408–12.
RCT subjects are found to be similar to those of 5. Majeed AW, Troy G, Nicholl JP, Smythe A, Reed
the RCT, possibly after adjustment for known MW, Stoddard CJ, Peacock J, Johnson AG. Ran-
confounders, confidence in the results of the domised, prospective, single-blind comparison of
wider cohort (and to a lesser extent for other laparoscopic versus small-incision cholecystect-
subsets within it) may be increased. omy. Lancet 1996; 347: 989–94.
112 The Advanced Handbook of Methods in Evidence Based Healthcare
6. Britton A, McKee M, Black N, McPherson K, trials of coronary artery bypass surgery. J. Am.
Sanderson C, Bain C. Choosing between random- Coll. Cardiol. 1988; 11: 237–45.
ised and non-randomised studies: a systematic 21. Horwitz RI, Viscoli CM, Clemens JD, Sadock
review. Health Technol. Assessment 1998; 2(13). RT. Developing improved observational methods
7. Colditz GA, Miller JN, Mosteller F. How study for evaluating therapeutic effectiveness. Am. J.
design affects outcomes in comparisons of ther- Med. 1990; 89: 630–8.
apy. I: Medical. Stat. Med. 1989; 8: 441–54. 22. Paradise JL, Bluestone CD, Bachman RZ, et al.
8. Miller JN, Colditz GA, Mosteller F. How study Efficacy of tonsillectomy for recurrent throat
design affects outcomes in comparisons of ther- infection in severely affected children. Results of
apy. II: Surgical. Stat. Med. 1989; 8: 455–66. parallel randomised and nonrandomised clinical
9. Lumley J, Bastian H. Competing or complimen- trials. N. Engl. J. Med. 1984; 310: 674–83.
tary? Ethical considerations and the quality of 23. Paradise JL, Bluestone CD, Rogers KD, et al.
randomised trials. Int. J. Tech. Assess. Health Efficacy of adenoidectomy for recurrent otitis
Care 1996; 12: 247–63. media in children previously treated with
10. McPherson K. The best and the enemy of the tympanostomy-tube replacement. Results of par-
good: randomised controlled trials, uncertainty, allel randomised and nonrandomised trials. JAMA
and assessing the role of patient preference in 1990; 263: 2066–73.
medical decision making. The Cochrane Lecture. 24. Schmoor C, Olschewski M, Schumacher M. Ran-
J. Epidemiol. Community Health 1994; 48: 6–15. domised and non-randomised patients in clinical
11. Rucker G. A two-stage trial design for testing trials: experiences with comprehensive cohort
treatment, self-selection and treatment preference studies. Stat. Med. 1996; 15: 236–71.
effects. Stat. Med. 1989; 8: 477–85. 25. Yamamoto H, Hughes RW, Schroeder KW, et al.
12. McKay JR, Alterman AI, McLellan T, Snider EC, Treatment of esophageal stricture by Eder-
O’Brien CP. Effect of random versus nonrandom Puestow or balloon dilators. A comparison
assignment in a comparison of inpatient and day between randomized and prospective nonrandom-
hospital rehabilitation for male alcoholics. J. ized trials. Mayo Clin. Proc. 1992; 67: 228–36.
Consult. Clin. Psychol. 1995; 63: 70–8. 26. Nicolaides K, de Lourdes Brizot M, Patel F,
13. Torgerson DJ, Klaber-Moffett J, Russell IT. Snijders R. Comparison of chorionic villus
Patient preferences in randomised trials: threat or sampling and amniocentesis for fetal karyotyping
opportunity. J. Health Serv. Res. Policy 1996: 1: at 10–13 weeks’ gestation. Lancet 1994; 344:
194–7. 435–9.
14. Redd WH, Anderson BL, Bovbjerg DH, Carpenter 27. Emanuel, EJ. Cost savings at the end of life. What
PJ, Dolgin M, Mitnick L, Schover LR, Stevens J. do the data show? JAMA 1996; 275: 1907–14.
Physiologic and psychobehavioural research in 28. Garenne M, Leroy O, Beau JP, Sene I. Efficacy of
oncology. Cancer 1991; 67: 813–22. measles vaccines after controlling for exposure.
15. Marmot MG, Bosma H, Hemingway H, Brunner Am. J. Epidemiol. 1993; 138: 182–95.
E, Stansfield S. Contribution of job control and 29. Antman K, Amoto D, Wood W, et al. Selection
other risk factors to social variations in coronary bias in clinical trials. J. Clin. Oncol. 1985; 3:
heart disease incidence. Lancet 1997; 350: 235–9. 1142–7.
16. Lilford RJ, Jackson G. Equipoise and the ethics 30. Shapiro CL, Recht A. Late effects of adjuvant
of randomisation. J. Roy. Soc. Med. 1995; 88: therapy for breast cancer. J. Natl Cancer Inst.
552–9. Monographs 1994; 16: 101–12.
17. Downs SH, Black NA, Devlin HB, Royston 31. Jha P, Flather M, Lonn E, et al. The antioxidant
CMS, Russell RCG. Systematic review of vitamins and cardiovascular disease – a critical
the effectiveness and safety of laparoscopic review of epidemiological and clinical trial data.
cholecystectomy. Ann. R. Coll. Surg. (Engl.) Ann. Intern. Med. 1995; 123: 860–72.
1996; 78: 241–323. 32. Pyorala S, Huttunen NP, Uhari M. A review and
18. Black NA, Downs SH. The effectiveness of sur- meta-analysis of hormonal treatment of crypt-
gery for stress incontinence in women. A system- orchidism. J. Clin. Endocrinol. Metab. 1995; 80:
atic review. Br. J. Urol. 1996; 78: 497–510. 2795–9.
19. CASS Principal Investigators and their Asso- 33. The Recurrent Miscarriage Immunotherapy Trial-
ciates. Coronary artery surgery study (CASS): A ists Group. Worldwide collaborative observa-
randomised trial of coronary artery bypass sur- tional study and meta-analysis on allogenic leuko-
gery. Comparability of entry characteristics and cyte immunotherapy for recurrent spontaneous
survival in randomised patients and nonrandom- abortion. Am. J. Repro. Immunol. 1994; 32:
ised patients meeting randomisation criteria. J. 55–72.
Am. Coll. Cardiol. 1984; 3: 114–28. 34. Watson A, Vail A, Vandekerckhove P, et al. A
20. Hlatky MA, Califf RM, Harrell FE, et al. Com- meta-analysis of the therapeutic role of oil soluble
parison of predictions based on observational data contrast media at hysterosalpingography: a sur-
with the results of randomised controlled clinical prising result? Fertil. Steril. 1994; 61: 470–7.
Randomised and Non-Randomised Studies: Internal and External Validity 113
35. Reimold SC, Chalmers TC, Berlin JA, Antman trials of arthritis. Arch. Intern. Med. 1993; 153:
EM. Assessment of the efficacy and safety of 243–8.
antiarrhythmic therapy for chronic atrial fibrilla- 51. The Diabetes Control and Complications Trial
tion: Observations on the role of trial design and Research Group; Klein R, Moss S. A comparison
implications of drug-related mortality. Am. Heart of the study populations in the diabetes control
J. 1992; 124: 924–32. and complications trial and the Wisconsin epi-
36. Ward LC, Fielding JWL, Dunn JA, Kelly KA for demiologic study of diabetic retinopathy. Arch.
the British Stomach Cancer Group. The selection Intern. Med. 1995; 155: 745–54.
of cases for randomised trials: a registry survey 52. Hartz AJ, Kuhn EM, Pryor DB et al. Mortality
of concurrent trial and non-trial patients. Br. J. after coronary angioplasty and coronary artery
Cancer 1992; 66: 943–50. bypass surgery (the National Medicare Experi-
37. Thompson SG. What sources of heterogeneity in ence). Am. J. Cardiol. 1992; 70: 179–85.
meta-analyses should be investigated? Br. Med. J. 53. Iezzoni LI (ed.). Dimensions of risk. In: Risk
1994; 309: 1351–5. adjustment for measuring health outcomes. AHSR,
38. Charlson ME, Horwitz RI. Applying results of Ann Arbor, 1994.
randomised trials to clinical practice: impact of 54. Fallowfield LJ, Hall A, Maguire GP, Baum M.
losses before randomisation. Br. Med. J. 1984; Psychological outcomes of different treatment
289: 1281–4. policies in women with early breast cancer out-
39. Muller DWM, Topol EJ. Selection of patients side a clinical trial. Br. Med. J. 1990; 301:
with acute myocardial infarction for thrombolytic 575–80.
therapy. Ann. Intern. Med. 1990; 113: 949–60. 55. Davey Smith G, Song F, Sheldon T. Cholesterol
40. Col NF, Gurwitz JH, Alpert JS, Goldberg RJ. lowering and mortality: the importance of con-
Frequency of inclusion of patients with cardio- sidering initial risk. Br. Med. J. 1993; 206:
genic shock in trials of clinical therapy. Am. J. 1367–73.
Cardiol. 1994; 73: 149–57. 56. Brand R, Kragt H. Importance of trends in the
41. Bailey KR. Generalizing the results of random- interpretation of an overall odds ratio in the meta-
ised clinical trials. Controlled Clinical Trials analysis of clinical trials. Stat. Med. 1992; 11:
1994; 15: 15–23. 2077–82.
42. Stephenson P, McKee M. Look twice. Eur. J. 57. Antman EM, Lau J, Kupelnick B, Mosteller F,
Publ. Health 1993; 3: 151–2. Chalmers TC. A comparison of results of meta-
43. Institute of Medicine. Effectiveness and outcomes analyses of randomized controlled trials and rec-
in health care: Proceedings of an invitational ommendations of clinical experts: Treatments for
conference by the Institute of Medicine Division myocardial infarction. JAMA 1992; 268: 240–8.
of Health Care Services. Heithoff KA, Lohr KN 58. Glasziou PP, Irwig LM. An evidence based
(eds). National Academy Press, Washington DC, approach to individualising treatment. Br. Med. J.
1990. 1995; 311: 1356–9.
44. Begg CB, Engstrom PF. Eligibility and extrapola- 59. Miettinen OS. The need for randomization in the
tion in cancer clinical trials. J. Clin. Oncol. 1987; study of intended effects. Stat. Med. 1983; 2:
5: 962–8 267–71.
45. Lee JY, Breaux SR. Accrual of radiotherapy pa- 60. Cowan CD, Wittes J. Intercept studies, clinical
tients to clinical trials. Cancer 1983; 52: 1014–16. trials, and cluster experiments: to whom can we
46. Begg CB, Zelen M, Carbonne PP, et al. Co- extrapolate? Controlled Clinical Trials 1994; 15:
operative groups and community hospitals: meas- 24–9.
urement of impact on community hospitals. 61. Davis CE. Generalizing from clinical trials.
Cancer 1983; 52: 1760–7. Controlled Clinical Trials 1994; 15: 11–14.
47. Martin JF, Henderson WG, Zacharaski LR et al. 62. Sharp SJ, Thompson SG, Altman DG. The
Accrual of patients into a multi hospital cancer relationship between treatment benefit and under-
clinical trial and its implications. Am. J. Clin. lying risk in meta-analysis. Br. Med. J. 1996; 313:
Oncol. 1984; 7: 173–82. 735–8.
48. Fentiman IS, Julien JP, Van DJ, et al. Reasons 63. Gray RJ. A Bayesian analysis of institutional
for non-entry of patients with DCIS of the breast effects in a multi-centre clinical trial. Biometrics
into a randomised trial (EORTC 10853). Eur. J. 1994; 50: 244–53.
Cancer 1991; 27: 450–2. 64. Pacala JT, Judge JO, Boult C. Factors affecting
49. Hunninghake DB, Darby CA, Probstfield JL. sample selection in a randomised trial of balance
Recruitment experience in clinical trials: literature enhancement: the FISCIT study. J. Am. Geriatr.
summary and annotated bibliography. Controlled Soc. 1996; 44: 377–82.
Clinical Trials 1987; 8: 6S–30S. 65. Chalmers TC. Ethical implications of rejecting
50. Rochon PA, Fortin PR, Dear KBG, Minaker patients for clinical trials (editorial). JAMA 1990;
KL, Chalmers TC. Reporting age data in clinical 263: 865.
114 The Advanced Handbook of Methods in Evidence Based Healthcare
66. Yusuf S, Held P, Teo KK. Selection of patients 81. Wilcox RG, Hampton JR, Banks DC, et al. Trial
for randomized controlled trials: implications of of early nifedipine in acute myocardial infarction.
wide or narrow eligibility criteria. Stat. Med. The TRENT study. Br. Med. J. 1986; 293:
1990; 9: 73–86. 1204–8.
67. Begg C, Cho M, Eastwood S, Horton R, Moher 82. CABRI Trial Participants. First year results of
D, Olkin I, et al. Improving the quality of CABRI (Coronary Angioplasty vs Bypass Re-
reporting of randomized controlled trials: the vascularisation Investigation). Lancet 1995; 346:
CONSORT Statement. JAMA 1996; 276: 637–9. 1179–84.
68. McKee M, Mossialos E. The impact of managed 83. SHEP Cooperative Research Group. Prevention
care on clinical research. Pharmacoeconomics of stroke by antihypertensive drug treatment in
1998; 14: 19–25. older persons with isolated systolic hypertension.
69. Chalmers I. Assembling comparison groups to JAMA 1991; 265: 3255–64.
assess the effects of health care. J. Roy. Soc. Med. 84. D’Alessandro U, Leach A, Drakeley CJ, et al.
1997; 90: 379–86. Efficacy trial of malaria vaccine SPf66 in Gambian
70. Corbett F, Oldham J, Lilford R. Offering patients infants. Lancet 1995; 346: 462–7.
entry in clinical trials: preliminary study of the 85. Goldbourt U, Behar S, Reicher-Reiss H, et al.
views of prospective participants. J. Med. Ethics Early administration of nifedipine in suspected
1996; 22: 227–31. acute myocardial infarction: the SPRINT 2 study.
71. Snowden C, Garcia J, Elbourne D. Reactions of Arch. Intern. Med. 1993; 153: 345–53.
participants to the results of a randomised con- 86. King SB, Lembo NJ, Kosinski AS et al. A ran-
trolled trial: exploratory study. Br. Med. J. 1998; domised study of coronary angioplasty compared
337: 21–6. with bypass surgery in patients with symptomatic
72. Snowden C, Garcia J, Elbourne D. Making sense multi-vessel coronary disease. N. Eng. J. Med.
of randomisation: responses of parents of critic- 1994; 331: 1044–50.
ally ill babies to random allocation of treatment 87. Sempertegui F, Estrella B, Moscoso J, et al.
Safety, immunogenicity and protective effect of
in a clinical trial. Soc. Sci. Med. 1998; 45:
the SPf66 malaria synthetic vaccine against
1337–55.
Plasmodium falciparum infection in a random-
73. Zelen M. Randomized consent designs for
ized double-blind placebo-controlled field trial in
clinical trials: an update. Stat. Med. 1990; 9:
an endemic area of Ecuador. Vaccine 1994; 12:
645–56.
337–42.
74. Snowden C, Elbourne D, Garcia J. Zelen random-
88. The Israeli SPRINT Study Group. Secondary Pre-
isation: attitudes of parents participating in a
vention Reinfarction Israeli Nifedipine Trial
neonatal clinical trial. Controlled Clinical Trials
(SPRINT). A randomised intervention trial of
1999; 20: 149–71. nifedipine in patients with acute myocardial
75. Schulz KF, Chalmers I, Grimes DA, Altman DG. infarction. Eur. Heart J. 1988; 9: 354–64.
Assessing the quality of randomization from 89. Hamm CW, Reimers J, Ischinger T, et al. A
reports of controlled trials published in obstetrics randomised study of coronary angioplasty com-
and gynecology journals, JAMA 1994; 272: pared with bypass surgery in patients with symp-
125–8. tomatic multi-vessel disease. N. Eng. J. Med.
76. Concato J, Feinstein AR, Holford TR. The risk of 1994; 331: 1037–43.
determining risk with multivariate models. Ann. 90. Valero MV, Amador LR, Galindo C, et al.
Intern. Med. 1993; 118: 201–10. Vaccination with SPf66, a chemically synthesised
77. Wennberg JE. What is outcomes research? In: vaccine, against Plasmodium falciparum in
Gelijns AC (ed.) Medical innovations at the Columbia. Lancet 1993; 341: 705–10.
crossroads. Vol. 1. Modern methods of clinical 91. Branigan JP, Walsh K, Collins WC, et al. Effect
investigation. Washington DC: National Acad- of early treatment with nifedipine in suspected
emy Press, 1990, pp. 33–46. acute myocardial infarction. Eur. Heart J. 1986;
78. McPherson K, Britton AR, Wennberg J. Are 7: 859–65.
randomised controlled trials controlled? Patient 92. Goy JJ, Eeckhout E, Burnand B, et al. Coronary
preferences and unblind trials. J. Roy. Soc. Med. angioplasty versus left main internal mammary
1997; 90: 652–6. grafting for isolated proximal left anterior
79. Brewin CR, Bradley C. Patient preferences and descending artery stenosis. Lancet 1994; 343:
randomised clinical trials. Br. Med. J. 1989; 299: 1449–53.
313–15. 93. Valero MV, Amador R, Aponte JJ, et al. Evalu-
80. Alonso PL, Smith T, Schellenberg JRM et al. ation of SPf66 malaria vaccine during a 22-month
Randomised trial of efficacy of SPf66 vaccine follow-up field trial in the Pacific coast of Colum-
against Plasmodium falciparum malaria in chil- bia. Vaccine 1996; 14: 1466–70.
dren in southern Tanzania. Lancet 1994; 344: 94. Sirnes PA, Overskeid K, Pedersen TR, et al.
1175–81. Evaluation of infarct size during the early use of
Randomised and Non-Randomised Studies: Internal and External Validity 115
Nifedipine in patients with acute myocardial bypass surgery with angioplasty in patients with
infarction: The Norwegian Nifedipine Multicentre multi-vessel disease. N. Eng. J. Med. 1996; 335:
Trial. Circulation 1984; 70: 638–44. 217–25.
95. Rodriguez A, Boullon F, Perez-Balino N, et al. 97. MRC Working Party. Medical Research Council
Argentine randomised trial of percutaneous trans- trial of treatment of hypertension in older
luminal coronary angioplasty versus coronary patients: principal results. Br. Med. J. 1992; 304:
artery bypass surgery in multi-vessel disease 405–12.
(ERACHI): in-hospital results and 1-year follow 98. MRC Working Party. Medical Research Council
up. J. Am. Coll. Cardiol. 1993; 22: 1060–7. trial of treatment of mild hypertension: principal
96. BARI Investigators. Comparison of coronary results. Br. Med. J. 1985; 291: 97–104.
7
A Review of Observational,
Quasi-Experimental and Randomised
Study Designs for the Evaluation of the
Effectiveness of Healthcare Interventions
B A R N A B Y C . R E E V E S , R A C H E L R . MAC L E H O S E ,
IAN M. HARVEY, TREVOR A. SHELDON,
I A N T . R U S S E L L and A N D R E W M . S . B L A C K
Methods Conclusions
Relevant studies were identified by searches of These findings suggest that QEO estimates of
electronic databases, reference lists of papers effectiveness from high-quality studies can be
Observational, Quasi-Experimental Study Designs to Evaluate Healthcare 117
valid. Although the majority of the QEO evi- non-randomised designs are seen as alternatives
dence reviewed was of poor quality, we recom- to experimentation, rather than as a set of
mend that QEO designs should not be rejected complementary approaches. QEO studies can
for evaluations of effectiveness just because provide estimates of the effectiveness of inter-
most past studies have had serious flaws in their ventions when RCTs are not possible, and can
designs or analyses. help to interpret the findings of RCTs and
their applicability in circumstances that do not
exactly match those of the RCTs (external valid-
INTRODUCTION ity). A recent editorial attempted to reconcile
the two points of view, arguing that the com-
plementary nature of different study designs
The debate about the respective advantages and reflects the different types of research questions
disadvantages of different research designs for that they are best suited to address.18
evaluating the effectiveness of healthcare inter- The internal validity of these non-randomised
ventions is one of the most controversial in the approaches must always be suspect, since it is
field of health services research, with strong impossible to be certain that all important
advocates for randomised controlled trials confounding factors have been identified and
(RCTs) on the one hand1–7 and observational or adequately controlled for.4,5 QEO studies also
quasi-experimental (QEO) research designs on offer less opportunity to control for biases since,
the other.8–10 although outcome assessment can often be
High-quality RCTs are widely perceived as blinded, healthcare providers and study partici-
the gold standard research design for evaluating pants are usually aware of treatment allocations.
effectiveness5,11–14 because they increase the The extent of the distrust of evidence derived
comparability of the groups being compared and from these non-experimental approaches is illus-
so minimise confounding of the intervention of trated by the above quotations, although Black10
interest by differences in known and unknown comments that ‘it is unclear how serious and
prognostic factors between groups. This prop- how insurmountable a methodological problem
erty is particularly important for evaluations of the threat to internal validity is in practice’.
interventions for two reasons. Firstly, small Previous comparisons of estimates of effec-
effects of comparable size to those arising from tiveness derived from RCTs and QEO studies
bias and confounding are often clinically have suggested that the latter tend to report
important.2,13 Secondly, quantifying the effect of larger estimates of the benefit of the treatment
an intervention, rather than merely whether or being evaluated.19–21 The perception that QEO
not it is statistically significantly different from estimates consistently favour new treatments has
an alternative, is very important; interventions led to the assumption that discrepancies arise
have financial costs, and often side-effects or from bias. However, comparisons from the lit-
complications, as well as benefits and the deci- erature are often cited selectively. They may not
sion to adopt an intervention therefore depends have been carried out systematically and may
on weighing up the relative magnitudes of the have failed to consider alternative explanations
benefits and the costs. These considerations for the discrepancies observed, for example dif-
have led some proponents of RCTs to adopt an ferences in the populations, interventions, or out-
extreme position about the value of non- comes studied, a greater susceptibility of QEO
randomised study designs, stating for example studies than RCTs to publication bias, or a tend-
that ‘. . . observational methods provide no ency for RCTs to underestimate effectiveness.10
useful means of assessing the value of a ther- A ‘sister’ review22 (see Chapter 6) focused on
apy’,14 or that ‘. . . they [observational studies] ways in which possible differences in external
cannot discriminate reliably between moderate validity between RCT and QEO studies can give
differences and negligible differences in out- rise to discrepant estimates of effect size. Our
come, and the mistaken clinical conclusions that review23 is complementary, and focuses on the
they engender could well result in the under- internal validity of QEO estimates of effect size.
treatment, over-treatment or other mistreatment The objective was to compare systematically
of millions of future patients’.13 estimates of effectiveness for specific interven-
QEO research designs15 have nevertheless tions derived from RCTs and QEO studies, and
been advocated for evaluating effectiveness, we adopted two strategies to minimise the influ-
because there are perceived to be many circum- ence of differences in external validity:
stances in which RCTs may be unnecessary,
inappropriate, impracticable or inadequate.9 1. Compare RCT and QEO estimates of effec-
Black and others9,10,16,17 contend that the polarity tiveness for an intervention where both esti-
of the debate about the value of QEO methods mates were reported in a single paper, on
for evaluating effectiveness arises because the assumption that such papers were more
118 The Advanced Handbook of Methods in Evidence Based Healthcare
likely to compare ‘like with like’ than com- Medline searches to estimate the amount of
parisons across papers; evaluations of any literature available for each of the interventions.
intervention were eligible. Our final choice of interventions required three
2. Compare RCT and QEO estimates of effec- aspects of evaluations to be uniform across
tiveness for an intervention where the respec- studies: (i) the intervention under consideration;
tive estimates were reported in different (ii) the outcome used to measure effectiveness;
papers and where the intervention, popula- and (iii) the population in which the intervention
tion and outcome investigated were antici- was investigated. The two interventions which
pated to be homogeneous across studies. best satisfied these criteria were selected for
Our primary interest was in the size and direc- review:
tion of discrepancies between RCT and QEO 1. Mammographic screening (MS; interven-
estimates, and the extent to which the discrep- tion) for women aged 50–64 years (popula-
ancies were associated with variations in meth- tion) to reduce mortality from breast cancer
odological quality, study design and other attri- (outcome);
butes of studies. We reasoned that being able to 2. Periconceptional folic acid supplementation
quantify such associations would be valuable to (FAS; intervention) for women trying to
healthcare decision makers in guiding the weight conceive (population) to prevent neural tube
that they should attach to observational evidence defects (outcome).
about the effectiveness of interventions. Papers were eligible if they evaluated either of
the interventions being reviewed, and matched
the definitions for the intervention, population
METHODS and outcome described above. We considered
this latter point to be very important, since only
Eligibility by ensuring that all studies were evaluating the
same intervention on the same population could
Two types of evidence were eligible for the first we be confident of minimising any confounding
strategy: of the association between study quality and
1. Primary studies, which reported both RCT effect size by differences in these attributes.
and QEO estimates of effect size for the
same intervention. Searching for Evidence
2. Secondary studies, which compared pooled
RCT and QEO estimates of effect size for A range of methods were used to identify rele-
the same intervention by synthesising evi- vant literature, including (i) searches of elec-
dence from primary studies which reported tronic databases (Medline 1966–June 1996;
either an RCT or QEO estimate of effect EMBASE 1980–June 1996; Cochrane Library);
size for the intervention. (The second strat- (ii) reference lists of relevant papers already
egy carried out this type of study for two identified; (iii) experts; and (iv) ‘hand searches’
interventions.) of databases of abstracts (Cochrane Library,
Secondary studies, which compared pooled RCT Database of Abstracts of Reviews of Effective-
and QEO estimates of effect size for many inter- ness, and a database of 1535 abstracts created by
ventions by aggregating evidence from primary another review team24) which we anticipated
studies which reported either an RCT or QEO might contain a higher proportion of relevant
estimate of effect size,20,21 were not eligible. material than specific journals. Searches were
These studies were excluded because of the not limited to the English language.
possibility of ‘confounding’ of any association Our efforts to design electronic search strat-
between study design and discrepancy in effect egies for the first strategy were frustrated by the
size by intervention. We also excluded studies inadequacy of the indexing of methodological
which reviewed RCT and QEO studies for the aspects of studies. Terms typically used to iden-
same intervention in a comparative way, but tify RCTs, such as ‘comparative studies’ and
which did not attempt to calculate an estimate of ‘prospective studies’,25 are ambiguous and can
effect size for different study designs. also describe observational studies. The term
For the second strategy, our choice of inter- ‘randomised controlled trial’ is available as a
ventions to review was made through a multi- publication type, but this classification was only
stage selection process. An initial list of 31 introduced in 1992. Because of these limitations,
interventions was drawn up by the project steer- search strategies designed to identify papers that
ing group on the basis of their personal expert- included both RCT and QEO study design ele-
ise, focusing on the availability of a substantial ments in fact detected studies that used a single
number of both RCTs and QEO studies. This study design. Traditional hand-searching meth-
number was reduced to seven by preliminary ods were rejected for this strategy because it was
Observational, Quasi-Experimental Study Designs to Evaluate Healthcare 119
not obvious which journals should be searched, (yes/no) score. For example, in some papers
and because eligible papers were exceedingly authors may have adjusted for some, but not all,
rare. Most references were therefore identified prognostic factors. No attempt was made to
from expert knowledge and from the reference mask the authorship or other details of these
lists of papers already obtained for the review. papers because many were already familiar to
For the second strategy, papers were identi- the reviewers, having been identified through
fied primarily by electronic searches of Medline their expertise. We did not use a more complex
and EMBASE, and the databases of abstracts. method of measuring quality because none
EMBASE searches were always carried out last of the available checklists were suitable for
and were designed to be less restrictive. They assessing the RCT versus QEO comparison (as
were intended to represent a final ‘trawl’ for opposed to the individual design elements).
papers which may have been missed by other For the second strategy, that is comparison of
searches. RCT and QEO estimates of effectiveness for the
same intervention but published in different
Quality Assessment papers, three reviewers used a detailed instru-
ment to evaluate all eligible papers. Details of
Papers reviewed for the first strategy often journal, author, year of publication, and all ref-
reported more than one comparison between erences were removed before the papers were
RCT and QEO effect sizes, for example for distributed for evaluation. If the methods of
different outcomes, populations or interventions. the study were reported in detail in a separate
The ‘quality’ of each comparison was assessed paper, relevant sections of the additional paper
by three reviewers using a checklist of six items were also distributed with the paper under
to identify the comparability of the populations evaluation.
studied (1–2 below) and the susceptibility of the The instrument used for measuring methodo-
comparison to bias (3–6 below): logical quality was based on an existing instru-
1. Were the same eligibility criteria used ment,26 which had been developed to provide a
for RCT and non-randomised study measure of external as well as internal validity
populations? for both randomised and observational study
2. Were RCT and non-randomised populations designs. The instrument was modified to try to
studied contemporaneously? satisfy of our aim of quantifying the relationship
3. Was the assessment of outcome ‘blinded’ between study quality and effect size; additional
(both RCT and non-randomised studies)? questions were included to measure the extent
4. Did the non-randomised estimates take to which factors likely to influence effect size
account of the effects of confounding by (for example adjustment for confounders) were
disease severity? addressed, and the original questions about
5. Did the non-randomised estimates take external validity were also modified. Questions
account of the effects of confounding by were classified as pertaining to one of four
comorbidity? dimensions of quality on the basis of face valid-
6. Did the non-randomised estimates take ity: (i) quality of the reporting of the study;
account of the effects of confounding by (ii) external validity; (iii) features of the study
other prognostic factors? designed to reduce bias; and (iv) features of the
For each comparison, a reviewer indicated study designed to reduce confounding. A score
whether the items were satisfied or not, and an for each question for each paper was calculated
overall score was calculated by averaging the by averaging the scores allocated by the three
scores allocated by the three reviewers. Scores reviewers.
greater than 3/6 were classified as high quality, For both strategies, additional information
and scores less than or equal to 3/6 as low about the interventions, populations studied,
quality. sample sizes and estimates of effectiveness were
The first three items were considered to be extracted from the papers by one reviewer and
matters of fact, and the few discrepancies checked by another.
between assessors were resolved by discussion
and reference to the papers. The extent to which
QEO estimates of effect size were adjusted for Data Analyses and Synthesis
confounding was considered to be a judgement,
and disagreements between assessors were For the first strategy, measures of the size of
respected. Disagreements mainly concerned discrepancies between RCT and QEO estimates
adjustment for co-morbidity and other prog- of effect size and outcome frequency for inter-
nostic factors, and probably arose because vention and control groups were calculated,
assessors were asked to assign a dichotomous where possible, for each comparison. There are
120 The Advanced Handbook of Methods in Evidence Based Healthcare
Table 1 Indices used to compare the findings of RCT and QEO study design
elements using strategy one
1 RR(RCT) RR effect size for the RCT element divided by the RR effect size for the
RR(QEO) QEO element
2 RD(RCT) 2 RD(QEO) Difference between RD effect sizes (risk or rate difference), i.e. RD for the
QEO element subtracted from RD for the RCT element
3 IOF(RCT) 2 IOF(QEO) Difference between OFs for intervention groups, i.e. the OF for the QEO
intervention group subtracted from the OF for the RCT intervention group
4 COF(RCT) 2 COF(QEO) Difference between OFs for control groups, i.e. the OF for the QEO control
group subtracted from the OF for the RCT control group
RR – relative risk (risk, odds or rate ratio); RD – risk difference; IOF – outcome frequency (risk, rate) for
intervention group; COF – outcome frequency (risk, rate) for control group.
no established indices for comparing the find- (three points were awarded where the three
ings from two studies. We therefore constructed reviewers agreed, and one point where two of
four indices, described in Table 1, to reflect the three reviewers agreed; points were summed
different ways of quantifying discrepancies in across papers and expressed as a percentage).
measures of effect size and outcome frequency. The internal consistency of the different quality
The rationale for considering discrepancies in dimensions was investigated using Cronbach’s
outcome frequencies, as well as discrepancies in alpha.
measures of effect size, arose from our primary Simple multiple regression was used to
objective of evaluating the internal validity of investigate predictors of different dimensions of
the QEO estimates. The quality criteria required quality and the sum of the scores for different
the RCT and QEO elements to compare the dimensions of quality, in order to try to validate
same populations, as well as to control for the instrument. Independent variables that were
confounding. Under these circumstances, the investigated included the intervention (MS or
outcome frequencies as well as the effect esti- FAS), study design (RCT, cohort or case control
mates should be the same for RCT and QEO study; other designs were omitted from the
elements. If we had considered only effect size analysis) and the interaction of intervention and
estimates, these might have disguised substantial study design.
differences in outcome frequencies.
Meta-regression was used to investigate fac-
It was not possible to calculate all indices for
tors associated with the magnitude of the effect
all comparisons considered, because some of the
data required were not reported either in the size estimates, taken as the natural logarithm of
papers which were reviewed, or in previous the relative risk estimate (ln(RR)), weighting
papers by the same research teams describing each estimate by the inverse of its variance.
the same investigations. At least one index was Dummy variables were created for cohort
calculated for each comparison. studies and case control studies, using RCTs as
Distributions showing the size and direction the baseline comparison. Associations between
of discrepancies were compared for high- and other independent variables, characterising the
low-quality comparisons. No analytical statistics quality of studies and their heterogeneity with
were calculated because several papers included respect to population and intervention, and
multiple comparisons, and it was considered effect size were also investigated.
unlikely that comparisons within papers were
statistically independent.
For the second strategy, we described the
performance of the instrument as well as the
RESULTS
relationship between effect size estimates and
study design, quality and other attributes. Inter- Strategy One
rater reliabilities for each question (pooled
across all papers for the two interventions) were Fourteen papers were identified from a variety
determined by two methods: (i) unweighted of sources and a wide range of journals. Three
Kappa (calculated using the statistical package papers were reviews,19,27,28 four were compre-
STATA Version 4.0; the use of three assessors hensive cohort studies,29–32 and seven compared
precluded the calculation of weighted Kappas primary data for RCT and QEO populations
for ordinal responses); (ii) percent agreement indirectly.33–39
Observational, Quasi-Experimental Study Designs to Evaluate Healthcare 121
The comprehensive cohort studies consisted The papers reported 38 comparisons between
of a cohort study of all subjects who were RCT and QEO findings. The numbers of com-
eligible for a RCT, with the subjects who par- parisons based on different study designs and
ticipated in the RCT nested in the cohort.40 their average quality are shown in Table 2.
Schmoor and colleagues32 used this study design Some comparisons were indirect, when the QEO
to compare mastectomy and breast preservation element of a study consisted of a single group;
treatments, and other adjuvant treatments for this usually occurred when the RCT found no
breast cancer. difference between intervention and control
The study by Hlatky and colleagues,34 which groups. For these comparisons, we compared
compared coronary artery bypass surgery versus one or both groups in an RCT with the single
medical treatment for coronary artery disease, non-trial group; the QEO population could
provides an example of other studies which either be the same (randomisable patients who
received a single treatment) or different from the
compared RCT and QEO primary data indirectly.
RCT population. Three of the 14 papers focused
Their QEO data were obtained from a pros-
on a comparison between trial and non-trial
pective database, which included information patients,32,37,39 but did not include all of the
about important clinical and prognostic factors; details which were considered relevant to the
data from the database were compared with data review. In these cases, relevant data were extrac-
from three independent RCTs. Outcomes for ted from earlier papers41–43 reporting the findings
medical and surgical patients in the prospective of the study by the same research groups.
database who would have been eligible for one The inclusion of studies which compared one
of the RCTs were analysed to provide a treat- or both groups in an RCT with a single non-trial
ment effect which was then compared with the group might be considered controversial, since
treatment effect found by the RCT. The three they cannot provide a QEO estimate of effect
RCTs used different eligibility criteria, allowing size directly. QEO estimates for these studies
the researchers to make three independent were calculated by comparing the single QEO
comparisons. outcome frequency with the outcome frequency
Table 2 Quality scores for different types of study reviewed using strategy one.
The percentage for each cell is calculated from the number of quality items scored
as present by three assessors for all comparisons of each study type
Reviews CCH IC:IC IC:I IC:C
Quality criterion (n 5 13) (n 5 8) (n 5 8) (n 5 6) (n 5 3)
CCH – comprehensive cohort; IC:IC – indirect comparison where the QEO study design element comprised both
intervention and control groups; IC:I – indirect comparison where the QEO study design element comprised an
intervention group only; IC:C – indirect comparison where the QEO study design element comprised a control
group only.
†
Only two papers mentioned blinding of outcome assessment;32,37 this criterion was scored as ‘yes’ for other
papers only where the outcome was unequivocal, i.e. survival/all-cause mortality.
‡
Three IC:IC comparisons33 were given credit for contemporaneity because the analyses took account of the date
of entry of each subject.
§
One review paper18 considered a number of outcomes; credit for blinding of outcome assessment was only given
for all-cause mortality.
$
Credit for adjustment of co-morbidity (and other confounders) was given to two IC:I37 and two IC:C com-
parisons.36 Because the QEO and RCT samples were drawn from the same population and the QEO population
received only one treatment, the QEO sample should have balanced on all factors with both intervention and
control RCT samples.
#
Credit was given for blinding of outcome for only one of the outcomes considered in each of the two IC:I
studies, i.e. for endometrial cancer but not any disease event37 and for all-cause mortality but not intra-breast
recurrence, distant metastasis or contralateral breast cancer.38
122 The Advanced Handbook of Methods in Evidence Based Healthcare
for the ‘missing’ group from the RCT element direction of discrepancies in RR and RD cannot
of the study. We believe that this approach is be inferred from the figure, because some out-
justified because the criteria for eligibility for comes were adverse and some favourable. Dis-
this strategy meant that RCT and QEO elements crepancies are therefore shown in Table 3,
had to investigate the same intervention and according to whether the RCT or the QEO effect
outcome and similar populations (except, per- size estimate was more extreme. For high-
haps, in the case of reviews). quality comparisons, there was no evidence at
The distributions of discrepancies in relative all that QEO effect size estimates were more
risk (RR), risk difference (RD) and outcome extreme, although there was some evidence of
frequencies for high- and low-quality compar- this tendency for low-quality comparisons.
isons are shown in Figure 1. Discrepancies for The same caveat applies to the distributions
all indices were much smaller for high- than of discrepancies in intervention and control out-
low-quality comparisons. The distribution of the come frequencies, which were clearly skewed,
(a)
(b)
indicating that QEO outcome frequencies tended or lower favourable, QEO outcome frequency.
to be larger than RCT outcome frequencies. The latter finding, is consistent with the above
Because of the prevailing view that eligibility view.
criteria for RCTs tend to exclude patients who It is interesting to consider the extent to which
have a higher risk of an adverse outcome, dis- differences in internal validity and differences in
crepancies in outcome frequency are shown in the populations contributed to discrepancies. In
Table 4 according to whether or not the RCT or two papers,35,36 the authors reported how the
QEO outcome frequency was greater for an effect size estimate changed for the QEO ele-
adverse outcome or smaller for a favourable ment as differences between the RCT and QEO
outcome. Discrepancies in outcome frequency populations and confounding factors were taken
for high-quality comparisons were equally dis- into account. In both these cases, restricting the
tributed, but those for low-quality comparisons QEO population to subjects who would have
were more likely to indicate a higher adverse, been eligible for the RCT brought about the
(c)
(d)
(c) differences between RCT and QEO intervention group outcome frequencies
(OF); and (d) differences between RCT and QEO control group outcome
frequencies (OF).
124 The Advanced Handbook of Methods in Evidence Based Healthcare
Table 3 Direction of discrepancies between discrepancies between RCT and QEO elements
RCT and QEO estimates of effect size. The may itself be ‘confounded’ by other factors; that
number of comparisons in which the RCT and is, other attributes of comparisons may be asso-
QEO elements gave the more extreme estimate ciated both with size and direction of discrep-
of effect size are tabulated separately (a) for ancy and with study quality. It was not possible
measures of relative risk and (b) risk or rate to investigate confounding formally, because of
difference the likely non-independence of the discrepancies
between RCT and QEO design elements across
(a) comparisons within papers. Instead, we described
either the relationship between quality and size
RCT more QEO more
and direction of discrepancies within strata of
extreme 5 extreme Total†
a possible confounding factor (see I and III
Low quality 9 0 13 22 below), or the relationship between possible
High quality 6 1 6 13 confounding factors and the size and direction
†
of discrepancies (see II below). The former
Relative risk estimates were not available for three approach allows inspection of the relationship
comparisons. in the absence of the confounding factor; using
the latter approach, if no relationship between
(b) the confounding factor and size or direction
RCT 5 QEO Total† of discrepancy is apparent, confounding is
extremely unlikely. Three confounding factors
Low quality 9 0 12 21 were considered.
High quality 7 0 3 10
†
Risk or rate difference estimates were not available I. Confounding due to the inclusion
for seven comparisons. of reviews
By their nature, reviews were scored as being of
Table 4 Direction of discrepancies between low quality. If comparisons in review papers
RCT and QEO estimates of (a) intervention
tend to give rise to larger discrepancies than
and (b) control outcome frequencies. The
comparisons made on the basis of primary data,
number of comparisons for which the RCT and
QEO outcome frequencies of an adverse event the inclusion of reviews may be a confounding
was higher, or a favourable event lower, are factor. It should also be pointed out that,
shown because reviews included data from several
primary studies, discrepancies for review com-
(a) parisons were more likely to arise from differ-
ences in the characteristics of the populations
RCT 5 QEO Total† studied. Because all of the review comparisons
were of low quality, the findings for the high-
Low quality 5 0 16 21 quality comparisons were unaltered. The dis-
High quality 3 0 5 8 tributions of the discrepancies for the low-
†
RCT or QEO intervention outcome frequencies were quality comparisons based on primary data still
not available for nine comparisons. showed a greater spread and a slight tendency to
produce more extreme effect size estimates.
(b)
RCT 5 QEO Total† II. Confounding by whether or not the
intervention is effective
Low quality 4 1 13 18
High quality 4 1 3 8 When an intervention is ineffective, the RCT
versus QEO comparison becomes harder to
†
Risk or rate difference estimates were not available interpret. If prognostic factors are reasonably
for twelve comparisons. balanced, the outcome frequencies should be
very similar for intervention and control groups
largest step in the convergence between RCT in both RCT and QEO populations and the
and QEO estimates, with adjustment for con- discrepancies between RCT and QEO effect size
founding producing relatively little subsequent estimates should be small. The observation that
change. high-quality comparisons show smaller discrep-
The comparison of studies of high and low ancies might therefore arise because they tend,
quality with respect to the size and direction of on average, to have evaluated less effective
Observational, Quasi-Experimental Study Designs to Evaluate Healthcare 125
interventions. If confounding by the effective- Stratifying by sample size reduced the num-
ness of an intervention were having an import- ber of comparisons in each stratum and makes it
ant effect, we would expect discrepancies to be more difficult to interpret the distributions of the
larger, on average, for comparisons involving discrepancies. Overall, however, the distribu-
effective interventions. On the basis of the pre- tions of the size of discrepancies appeared to be
vailing view that QEO studies are more likely to similar across strata, and showed the same pat-
yield exaggerated effect sizes, we might also tern as without stratification, that is larger dis-
expect the discrepancies involving effective crepancies for low-quality comparisons. With
interventions to be more extreme for QEO study respect to the direction of discrepancies, the
design elements. tendency for QEO estimates to be more extreme
Confounding by the effectiveness of the inter- was most apparent for low-quality comparisons
vention was explored by tabulating size and based on large sample sizes, as predicted and
direction of discrepancies for comparisons by a shown in Table 5. There was no evidence from
proxy marker of the effectiveness of an inter- high-quality comparisons, whether based on
vention, namely whether or not the RCT study small or large sample sizes, to support the view
design element yielded a statistically significant
result. The distributions of the discrepancies for Table 5 Direction of discrepancies between
‘effective’ and ‘ineffective’ comparisons showed RCT and QEO results classified by quality and
no marked difference in either their dispersion stratified by the sample size of the RCT study
or symmetry, suggesting that confounding by design element, (a) for measures of relative
the effectiveness of an intervention was unlikely risk and (b) risk or rate difference
to explain the difference in the size of discrep-
ancies for high- and low-quality comparisons. (a)
RCT sample
III. Confounding by sample size
size #750 RCT more QEO more
One would expect a direct relationship between (n 5 21) extreme 5 extreme Total†
the size of discrepancies and sample size,
because sample size (in conjunction with out- Low quality 8 0 4 12
come frequency) determines the precision of an High quality 2 1 3 6
effect size estimate; that is, discrepancies for †
Relative risk estimates were not available for three
comparisons based on small sample sizes would comparisons.
be expected on average to be larger than those
based on large samples. If high-quality com-
RCT sample
parisons tend on average to be based on larger size .750 RCT more QEO more
sample sizes, the relationship between quality (n 5 16) extreme 5 extreme Total
and size of discrepancy might be substantially
reduced by taking account of sample size. (It Low quality 1 0 8 9
could be argued that sample size is itself a factor High quality 4 0 3 7
that should be considered in assessing quality,
but it was not included in the quality assessment
which was carried out for this strategy.) With (b)
respect to the direction of discrepancies, one RCT sample
would expect any general tendency for QEO size #750 RCT more QEO more
estimates to produce more extreme estimates of (n 5 21) extreme 5 extreme Total†
effect size to be more evident when the effect
size estimates are more precise, that is for larger Low quality 7 0 4 11
sample sizes. High quality 3 0 0 3
Confounding by sample size was explored by †
Risk or rate difference estimates were not available
examining size and direction of discrepancies
for seven comparisons.
for comparisons by low and high quality separ-
ately for ‘small’ and ‘large’ sample size strata.
Classification was based on the total RCT
sample size only, because QEO sample size was RCT sample
affected for some comparisons by the absence of size .750 RCT more QEO more
either an intervention or control group; a cut-off (n 5 16) extreme 5 extreme Total
of 750 subjects was chosen since this created Low quality 2 0 7 9
two groups with approximately equal numbers High quality 4 0 3 7
of comparisons in each.
126 The Advanced Handbook of Methods in Evidence Based Healthcare
that QEO effect size estimates tend to be more 6. Both cohort and case control studies had
extreme than RCT estimates. significantly poorer total quality than RCTs
(mean quality differences 24.9, 95% CI 26.5 to
23.3, P , 0.0001, and 21.8, 95% CI 23.5 to
Strategy Two 20.1, P 5 0.03, respectively). The difference in
total quality between cohort and case control
Thirty-four eligible papers were identified, 17 studies was also statistically significant (mean
evaluating MS44–60 and 17 evaluating FAS.61–77
difference 3.1, CI 1.2 to 5.0, P 5 0.003). There
For MS and FAS respectively, eight44,45,48,51,54,56,60
was no independent effect of intervention,
and four papers61,70,72,73 were RCTs,
49,52,55,58,59 nor interaction of intervention and study
five and six62,63,64,67,68,69 were non-
design, i.e. the findings were the same for both
randomised trials or cohort studies, and
three46,47,50,52 and six65,66,71,74,75,77 were matched interventions.
or unmatched case control studies; one study of Considering the individual dimensions of
MS57 and one of FAS76 used some other design. quality, cohort studies but not case control studies
Inter-assessor agreement was similar for MS had worse reporting scores than RCTs (mean
and FAS, and is reported here for all papers difference 22.3, CI 23.2 to 21.4, P , 0.0001).
combined. The majority of Kappa statistics Cohort studies also had poorer reporting scores
indicated only slight (0–0.2) or fair (.0.2–0.4) than case control studies (mean difference 22.0,
agreement, yet percent agreement exceeded 60% CI 23.4 to 20.9, P , 0.001). For confounding,
for all but two questions. One explanation for both cohort and case control studies had poorer
this apparent contradiction is that the majority of scores than RCTs (mean quality differences
papers were given the same rating on many 22.8, 95% CI 23.4 to 22.3, P , 0.001, and
items (see Note a). The moderate overall level of 22.4, 95% CI 23.0 to 21.8, P 5 0.001,
agreement can be attributed to problems experi- respectively); cohort and case control studies did
enced by reviewers in using the instrument; not differ significantly. There were no differ-
despite piloting, some questions remained ences in external validity or bias between study
ambiguous and difficult to apply to all study designs or interventions.
designs.23 The fact that RCTs were given significantly
Cronbach alpha values for different quality higher scores than both cohort and case control
dimensions were low, suggesting that the items studies provides some evidence of the content
were not assessing a homogeneous aspect of validity of the instrument. However, the finding
quality. Surprisingly, the highest alpha value that case control studies scored higher than
was obtained for the sum of the four dimen- cohort studies ran counter to the widely
sions, although the increase in the value of alpha accepted view that cohort studies provide higher
for the total quality score is due in part to the quality evidence than case control studies. The
larger number of items. There were modest explanation for this finding may lie in the fact
correlations between the quality of reporting and that cohort studies were a mix of quasi-
measures taken to avoid bias (0.28) and con- experiments and observational studies, whereas
founding (0.56), but no correlation between bias all case control studies were observational.
and confounding (20.03). We had expected a Some reviewers commented that observational
strong inverse correlation between external studies seemed to be of higher quality than
validity and confounding (since RCTs, which quasi-experiments, and this perception is sup-
are often considered to have poor external valid- ported by the difference in reporting scores for
ity, were automatically given credit for control- different study designs, which followed the
ling for confounding) but found none (20.07). same pattern. The observation may, therefore,
Average scores for different aspects of quality having nothing to do with study design per se,
for each type of study design are shown in Table that is cohort versus case control study, but
Table 6 Scores for different quality dimensions and total quality scores (and
standard deviations) for different types of study reviewed for strategy two
Cohort Case control Other study
RCTs studies studies designs
result from the fact that the quasi-experiments across studies, although we tried to minimise
tended to be of low quality. this source of heterogeneity by choosing the
The number of papers was limited by the duration of follow-up which was closest to
strict criteria which we laid down in order to 10 years when findings for a particular study
achieve homogeneity of the intervention, popu- were reported for different follow-up durations.
lation and outcome investigated. There were Alexander and colleagues56 also highlighted the
some difficulties in applying these criteria, and possibility that the definition of a breast cancer
they were relaxed in some instances; we also death might have varied across studies.
found that studies differed in ways which we For FAS, eight studies investigated women
had not anticipated at the outset. without a previous history of a neural tube
For MS, we originally defined the population defect, that is the effect of FAS on the occur-
of interest to be women aged 50–64 years. rence of neural tube defects, and nine studied
However, several of the studies which were women with a previous history of a neural tube
identified did not limit their study populations to defect, that is the effect of FAS on the recur-
this age group, and did not report findings for rence of neural tube defects. Since our eligibility
this age range separately. If the original popula- criteria did not specify whether FAS should be
tion criterion had been applied strictly, several intended to prevent occurrence or recurrence,
studies would have had to have been excluded. and because excluding either would have halved
The population criterion was therefore relaxed the number of studies available, both types of
to include women aged between 35 and 74 study were included. Researchers defined a
years, the range of ages investigated in the neural tube defect in different ways so that there
studies which we identified. may have been further heterogeneity within
Variation in the intervention/exposure was populations of women with a previous history of
also a major source of heterogeneity across a neural tube defect. The definition of a control
studies. Some studies used 1-view mammo- subject in case control studies also varied; some
graphy, some used 2-view mammography, and studies recruited a control group of healthy
others used both (on different screening visits, babies, some recruited malformed control babies
or when the screening protocol changed during without a neural tube defect, and others recruited
the course of the study); in some studies, mam- both types of control groups.
mography was combined with breast examina- As was found in studies of MS, several sources
tion by a physician or with the teaching of breast of heterogeneity in the FAS interventions were
self-examination. The interval between screens observed. Some studies used folic acid alone
varied across studies from once each year to and others used multivitamins which contained
once every four years, and some studies used a folic acid; one study used a factorial design to
variable screening interval. The duration of the investigate the effect of both folic acid and
intervention varied across studies from four to multivitamins simultaneously. The dose of folic
17 years. Provision for the control group varied acid that was used varied from 0.3 mg/day to
across studies; in some cases control participants 5 mg/day. The time period during which women
received nothing, but in others they were taught had to have taken FAS in order to be considered
breast self-examination. Case control studies ‘fully supplemented’ varied, particularly with
were carried out in the context of an entire respect to supplementation prior to conception.
population being offered screening, and the (If FAS was used prior to conception, most
only exposure that could be compared across researchers generally assumed that supplementa-
studies was ‘ever screened’ (compared to tion continued during the critical first three
never screened). Screening provision for partici- weeks after conception.) In some studies the
pants at the end of the studies varied; in some control group received multivitamins or a trace
studies, both intervention and control groups supplement, and in others the control group
were offered screening, while in others both received nothing; this variation was con-
intervention and control groups received no founded by study design, since RCTs or quasi-
screening. experiments were more likely to give some
A key factor that limited the number of MS intervention to the control group than observa-
studies reviewed for this strategy was the choice tional studies.
of breast cancer mortality as the outcome For MS, the pooled RR estimate for case
measure. Breast cancer mortality was chosen to control studies (0.51) was about 0.6 times less
minimise the problems of lag and lead time (indicating greater benefit) than for RCTs
biases that affect survival studies of the effec- (RR 5 0.85; mean difference ln(RR) 5 20.50,
tiveness of screening interventions. Despite 95% CI 5 21.04 to 0.03, P 5 0.06). In contrast,
choosing this outcome, there was some variation the pooled RR estimate for cohort studies was
across studies in the precise outcome measured. the same as for RCTs (RR 5 0.82; mean differ-
The duration of follow-up varied considerably ence ln(RR) 5 20.03, 95% CI 5 20.34 to 0.28,
128 The Advanced Handbook of Methods in Evidence Based Healthcare
P 5 0.82). Neither total quality, nor any compo- effect size between different study designs. The
nent of the quality score, were significantly lack of relationship between study quality and
associated with effect size. Including any of effect size for different study designs can be
these variables resulted in lower values of seen in Figure 2. None of the other independent
r2(adj), without affecting the mean differences in variables, singly or in combination, were sig-
Figure 2 Blobbogram showing the effect size point estimates and confidence
intervals for MS studies reviewed for strategy two. Pooled estimates are shown
separately for RCTs, cohort studies and case control studies. Within each type of
design, studies are ranked in order of quality.
Observational, Quasi-Experimental Study Designs to Evaluate Healthcare 129
nificantly associated with effect size. Nor did to 0.13 respectively) compared to their standard
they affect the mean differences in effect size errors.
between different study designs. Coefficients for For FAS, the pooled RR estimate for case
case control and cohort studies in different control studies (0.72) was about 2.6 times higher
models varied little (20.47 to 20.61 and 20.12 (indicating less benefit) than for RCTs (RR 5
Figure 3 Blobbogram showing the effect size point estimates and confidence
intervals for FAS studies reviewed for strategy two. Pooled estimates are shown
separately for RCTs, cohort studies and case control studies. Within each type of
design, studies are ranked in order of quality.
130 The Advanced Handbook of Methods in Evidence Based Healthcare
0.28; mean difference ln(RR) 5 0.96, 95% interventions for cardiovascular disease pre-
CI 5 0.02 to 1.89, P 5 0.05). As in the analysis dominated amongst the comparisons which were
of MS studies, the pooled RR estimate for reviewed.
cohort studies was the same as for RCTs Secondly, most high-quality comparisons
(RR 5 0.27; mean difference ln(RR) 5 20.03, studied RCT and QEO populations with the
95% CI 5 21.13 to 1.08, P 5 0.96), and neither same eligibility criteria; this may have had the
total quality, nor any component of the quality effect of creating relatively uniform risk strata,
score, were significantly associated with effect reducing the possibility of confounding. In con-
size. The lack of relationship between study trast, QEO studies typically use more relaxed
quality and effect size for different study selection criteria than RCTs, recruiting popula-
designs can be seen in Figure 3. tions which are more heterogeneous with respect
Time of starting and stopping supplementa- to prognostic factors.
tion, but none of the other independent vari- Thirdly, the papers that were reviewed may
ables, were marginally associated with effect have been subject to some form of publication
size; however, including these variables had no bias. Authors did not appear to be disinterested
impact on the coefficients for cohort and case about the comparison between RCT and QEO
control studies. The RR was increased (that is estimates, and findings appeared to support
less benefit) by a factor of 1.04 for each unsup- authors’ viewpoints. Therefore the papers may
plemented week prior to conception (coefficient not be a representative sample of all instances in
5 0.04, 95% CI 20.001 to 0.08, P 5 0.05), and which researchers have set out to compare RCT
the RR was reduced (that is greater benefit) by a and QEO effect size estimates, and researchers
factor of 0.95 for each additional week supple- may have chosen to report examples that sup-
mented after conception (coefficient 5 20.05, ported their points of view.
95% CI 20.10 to 0.01, P 5 0.10). These asso- The second strategy found no difference in
ciations are not entirely consistent with the effect size between RCTs and cohort studies,
critical period for closure of the neural tube, suggesting that evidence from cohort studies can
but may simply reflect that women who be valid when this evidence is considered col-
were supplemented for longer were more fully lectively. However, for both interventions that
supplemented. were reviewed, significantly different RR effect
size estimates were obtained for case control
studies compared with either RCTs or cohort
studies. Interestingly, the direction of the dis-
CONCLUSIONS crepancy was not consistent across inter-
ventions, with RR estimates from case control
studies indicating on average more benefit than
The results of the first strategy suggest that QEO those from RCTs and cohort studies for MS, but
estimates of effectiveness may be valid (as less benefit for FAS. No association was found
measured by agreement with the results of between study quality and effect size for either
RCTs), providing that important confounding MS or FAS interventions, after taking account
factors are measured and controlled for using of different study designs.
standard epidemiological methods of analysis, The lack of association between quality and
that is stratification or regression modelling. The effect size is difficult to interpret because this
small size of the discrepancies for high-quality result could have arisen for a variety of reasons.
comparisons between RCT and QEO estimates The first, and most relevant to the review, is that
also suggests that psychological factors, for study quality is truly not associated with RR in a
example treatment preferences or willingness predictable way. Second, the two reviews may
to be randomised, had a negligible effect on have had inadequate power to detect an import-
outcome. Previous comparisons reported in ant association. Third, the instrument may have
the literature may have overemphasised the been an insensitive measure of quality, and may
differences between RCT and QEO estimates of have failed to characterise relevant study attri-
effectiveness because of the poor quality of butes adequately. Fourth, using a quality score
most non-randomised, and especially quasi- that was the sum of many items may have
experimental, evidence. disguised the opposing influences of particular
We do not recommend generalising the subgroups of items on effect size.
results found using the first strategy to other We believe that the second and fourth explana-
contexts, for three main reasons. Firstly, few tions are unlikely, since there was no evidence
papers were reviewed, and our findings may of any trend in effect with study quality, nor of
depend on the specific interventions evaluated associations between components of study qual-
in these papers. For example, evaluations of ity and effect size. The fact that the quality
Observational, Quasi-Experimental Study Designs to Evaluate Healthcare 131
scores for papers reviewed varied considerably We therefore conclude, in common with
from 11.6 to 23.7, and that RCTs were given standard methodological texts,78 that case con-
significantly higher quality scores on average trol studies designed to estimate effectiveness
than both the cohort and case control studies, should be interpreted with caution, and that the
provides some evidence of the construct validity direction of any discrepancy between RR esti-
of the instrument and goes some way to refute mates for case control studies and other study
the third explanation. However, the difficulty we designs is likely to depend on the intervention
experienced using the instrument for measuring being evaluated. There was no evidence that
methodological quality means that we are not discrepant estimates for case control studies
confident in concluding that the first explanation could be attributed to confounding by quality or
is correct. sources of heterogeneity.
There are similar difficulties in interpreting
the failure to show associations between aspects
of heterogeneity, considered likely to be asso-
ciated with outcome for a priori reasons, and RECOMMENDATIONS
effect size. The analyses may have had in-
adequate power, or the independent variables Our aim was to establish ‘safety limits’ for QEO
included in the analysis may have been un- studies of different levels of quality, that is the
reliable measures of the relevant attributes. size of effect which would be unlikely to be
Information about other variables of interest, explained by bias, in order to guide interpreta-
for example dose of folic acid, was not always tion of QEO evidence. This aim was thwarted
available or was inadequately specified. There-
for both strategies, for different reasons. For
fore, despite being unable to demonstrate
strategy one, we felt unable to draw strong
significant associations between sources of
conclusions because of the paucity of evidence,
heterogeneity and effect size, we are wary of
and the potentially unrepresentative nature of
interpreting the pooled results for different study
designs. Investigating reasons for discrepancies, the evidence we reviewed. For strategy two, the
rather than providing a pooled estimate, was the aim was thwarted by our inability to distinguish,
primary objective of the review. and measure, the influences of variations in
There are several possible reasons for the internal and external validity between studies.
different direction of the discrepancy for the two Our recommendations relate directly to these
interventions, although it should be recognised obstacles.
that any such explanations are post hoc. Case Our first recommendation is a general one.
control studies of MS provide estimates of We strongly recommend that the use of quasi-
screening with 100% coverage, albeit ever experimental designs to evaluate healthcare
versus never screened; both RCTs and cohort interventions should not be rejected on the basis
studies included substantial proportions of un- of past QEO evidence because, for both strat-
screened women in their intervention groups for egies, most quasi-experiments that were
intention-to-treat analyses, with coverage for reviewed were of poor quality.
MS varying considerably between studies that This leads to our second recommendation,
reported this information. Case control studies namely that there is a need for more direct
of FAS also provide estimates of supplementa- evidence about the comparability of findings
tion with 100% coverage. However, coverage from RCTs and QEO studies. We believe that
amongst women assigned to supplementation in the comprehensive cohort study is the best study
RCTs and cohort studies is likely to have been design to use to obtain such evidence with the
much higher than for MS. addition that ineligible patients who receive one
Almost all case control studies also required or other of the treatments being investigated, not
women to recall their ‘exposure’. Recalling sup- usually included in comprehensive cohorts,
plementation, which often required remember- should also be followed-up. Studies need to be
ing the particular supplement taken in order to carried out in areas where RCTs are the pre-
confirm the folic acid content, is likely to have ferred design, as well as areas where RCTs are
been much less reliable than asking women to problematic, in order to assess the generalis-
recall whether they had ever had a mammogram. ability of evidence about the validity of QEO
Unreliable recall could easily lead to bias, with evidence. Analyses of comprehensive cohort
women whose pregnancy was affected by a studies should focus on using the RCT estimate
neural tube defect being more likely to report of effectiveness, and estimates of the effect of
they had taken supplementation when they had other prognostic factors from the entire cohort,
not than women who did not have an affected to predict outcome frequencies for different
pregnancy. groups of patients who were not randomised;
132 The Advanced Handbook of Methods in Evidence Based Healthcare
19. Chalmers TC, Matta RJ, Smith H, Kunzler AM. carcinoma. Preliminary results. Acta Oncol. 1988;
Evidence favouring the use of anticoagulants in 27: 671–7.
the hospital phase of acute myocardial infarction. 32. Schmoor C, Olschewski M, Schumacher M. Ran-
N. Engl. J. Med. 1977; 297: 1091–6. domized and non-randomized patients in clinical
20. Colditz GA, Miller JN, Mosteller F. How study trials: experiences with comprehensive cohort
design affects outcomes in comparisons of ther- studies. Stat. Med. 1996; 15: 263–71.
apy. I: Medical. Stat. Med. 1989; 8: 441–54. 33. Gray-Donald K, Kramer MS. Causality inference
21. Miller JN, Colditz GA, Mosteller F. How study in observational vs. experimental studies. Am. J.
design affects outcomes in comparisons of ther- Epidemiol. 1988; 127: 885–92.
apy. II: Surgical. Stat. Med. 1989; 8: 455–66. 34. Hlatky MA, Facc MD, Califf RM, et al. Com-
22. Britton A, McKee M, Black N, McPherson K, parison of predictions based on observational data
Sanderson C, Bain C. Choosing between ran- with the results of RCT of coronary artery bypass
domised and non-randomised studies: a system- surgery. J. Am. Coll. Cardiol. 1988; 11: 237–45.
atic review. Health Technol. Assessment 1998; 2: 35. Horwitz RI, Viscoli CM, Clemens JD, Sadock
1–115. RT. Developing improved observational methods
23. MacLehose RR, Reeves BC, Harvey IM, Black for evaluating therapeutic effectiveness. Am. J.
AM, Sheldon TA, Russell IT. A review of ran- Med. 1990; 89: 630–8.
domized controlled trials, quasi-experimental and 36. Kirke PN, Daly LE, Elwood JH, for the Irish
observational study designs for evaluating the Vitamin Study Group. A randomized trial of low
effectiveness of health care. Health Technol. dose folic acid to prevent NTD. Arch. Dis. Child.
Assessment; in press. 1992; 67: 1442–6.
24. Prescott R, Counsell CE, Gillespie WJ, et al. 37. Ward L, Fielding JWL, Dunn JA, Kelly KA, for
Factors that limit the quality, number and pro- the British Stomach Cancer Group. The selection
gress of randomized controlled trials. Health of cases for randomized trials: a registry of con-
Technol Assessment 1999; 3(20). current trial and non-trial participants. Br. J.
25. Dickersin K, Scherer R, Lefebvre C. Identifying Cancer 1992; 66: 943–50.
relevant studies for systematic reviews. Br. Med. 38. Fisher B, Costantino JP, Redmond CK, et al.
J. 1994; 309: 1286–91. Endometrial cancer in tamoxifen-treated breast
26. Downs SH, Black N. The feasibility of creating a cancer patients: findings from the National
checklist for the assessment of the methodo- Surgical Adjuvant Breast and Bowel Project
logical quality both of randomized and non- (NSABP) B-14. J. Natl Cancer Inst. 1994; 86:
randomized studies of health care interventions. J. 527–37.
Epidemiol. Community Health 1998; 52: 344–52. 39. Marubini E, Mariani L, Salvadori B, et al. Results
27. Wortman PM, Yeaton WH. Synthesis of results in of a breast cancer-surgery trial compared with
controlled trials of coronary artery bypass graft observational data from routine practice. Lancet
surgery. In: Light R (ed.) Evaluation Studies 1996; 347: 1000–3.
Review Annual. Sage Publications, 1983; 8: 40. Olschewski M, Scheurlen H. Comprehensive
536–57. cohort study: an alternative to randomized con-
28. Reimold SC, Chalmers TC, Berlin JA, Antman sent design in a breast preservation trial. Methods
EM. Assessment of the efficacy and safety of Inf. Med. 1985; 24: 131–4.
antiarrhythmic therapy for chronic atrial fibrilla- 41. Veronesi U, Banfi A, Salvadori B, et al. Breast
tion: observations on the role of trial design and conservation is the treatment of choice in small
implications of drug-related mortality. Am. Heart breast cancer: long-term results of a randomized
J. 1992; 124: 924–32. trial. Eur. J. Cancer 1990; 26: 668–70.
29. CASS principal investigators and their associates. 42. Schumacher M, Bastert G, Bojar H, et al. A
Coronary artery surgery study (CASS): A ran- randomized 2 3 2 trial evaluating hormonal treat-
domized trial of coronary artery bypass surgery. ment and the duration of chemotherapy in node-
Comparability of entry characteristics and sur- positive breast cancer patients. J. Clin. Oncol.
vival in randomized and non-randomized patients 1994; 12: 2086–93.
meeting randomization criteria. J. Am. Coll. 43. Allum WH, Hallissey MT, Kelly KA, for the
Cardiol. 1984; 3: 114–28. British Stomach Cancer Group. Adjuvant chemo-
30. Paradise JL, Bluestone CD, Bachman RZ, et al. therapy in operable gastric cancer. 5 year follow-
Efficacy of tonsillectomy for recurrent throat up of first British stomach cancer group trial.
infection in severely affected children: results of Lancet 1989; i: 571–4.
parallel randomized and non randomized clinical 44. Dales LG, Friedman GD, Collen MF. Evaluating
trials. N. Engl. J. Med. 1984; 310: 674–83. periodic multiphasic health check ups: a con-
31. Blichert-Toft M, Brincker H, Andersen JA, et al. trolled trial. J. Chron. Dis. 1979; 32: 385–404.
A Danish randomized trial comparing breast pre- 45. Shapiro S, Vent W, Strax P, Venet L, Roeser R.
serving therapy with mastectomy in mammary Ten- to fourteen-year effect of screening on breast
134 The Advanced Handbook of Methods in Evidence Based Healthcare
cancer mortality. J. Natl Cancer Inst. 1982; 69: breast cancer in women under 50 years at entry:
349–55. the Kotka pilot project in Finland. Int. J. Cancer
46. Collette HJA, Rombach JJ, Day NE, de Waard F. 1995; 63: 55–7.
Evaluation of screening for breast cancer in a 59. Peer PGM, Werre JM, Mravunac M, et al. Effect
non-randomized study (The DOM Project) by on breast cancer mortality of biennial mammo-
means of a case-control study. Lancet 1984; i: graphic screening of women under age 50. Int. J.
1224–6. Cancer 1995; 60: 808–11.
47. Verbeek ALM, Holland R, Sturmans F, et al. 60. Tabar L, Fagerberg G, Chen HH, et al. Efficacy
Reduction of breast cancer mortality through of breast cancer screening by age. New results
mass screening with modern mammography. First from the Swedish two-county trial. Cancer 1995;
results of the Nijmegen Project, 1975–1981. 75: 2507–17.
Lancet 1984; i: 1222–4. 61. Laurence KM, James N, Miller MH, Tennant GB,
48. Andersson I, Aspergren K, Janzon L, et al. Mam- Campbell H. Double-blind randomized controlled
mographic screening and mortality from breast trial of folate treatment before conception to pre-
cancer: the Malmo mammographic screening vent recurrence of neural-tube defects. Br. Med. J.
trial. Br. Med. J. 1988; 297: 943–8. 1981; 282: 1509–11.
49. Morrison AS, Brisson J, Khalid N. Breast cancer 62. Smithells R, Sheppard S, Schorah CJ, et al.
incidence and mortality in the Breast Cancer Apparent prevention of neural tube defects by
Detection Demonstration Project. J. Natl Cancer periconceptional vitamin supplementation. Arch.
Inst. 1988; 80: 1540–7. Dis. Child. 1981; 56: 911–18.
50. Palli D, Rosselli Del Turco M, Buiatti E, et al. 63. Smithells R, Seller MJ, Harris R, et al. Further
Time interval since last test in a breast cancer experience of vitamin supplementation for pre-
screening programme: a case-control study in vention of neural tube defect recurrences. Lancet
Italy. J. Epidemiol. Community Health 1989; 43: 1983; i: 1027–31.
241–8. 64. Seller MJ, Nevin NC. Periconceptional vitamin
51. Frisell J, Eklund G, Hellstrom L, et al. Random- supplementation and the prevention of neural tube
ized study of mammography screening – pre- defects in south-east England and Northern
liminary report on mortality in the Stockholm Ireland. J. Med. Genet. 1984; 21: 325–30.
trial. Breast Cancer Res. Treat. 1991; 18: 49–56. 65. Mulinare J, Cordero JF, Erickson JD, Berry RJ.
52. Collette HJA, de Waard F, Rombach JJ, Collette Periconceptional use of multivitamins and the
C, Day NE. Further evidence of benefits of a occurrence of neural tube defects. JAMA 1988;
(non-randomized) breast cancer screening pro- 260: 3141–5.
gramme: the DOM project. J. Epidemiol. Com- 66. Mills JL, Rhoads GG, Simpson JL, et al. The
munity Health 1992; 46: 382–6. absence of a relation between the periconcep-
53. Miller AB, Baines CJ, To T, Wall C. Canadian tional use of vitamins and neural-tube defects. N.
National Breast Screening Study: 2. Breast cancer Engl. J. Med. 1989; 321: 430–5.
detection and death rates among women aged 50 67. Milunsky A, Jick H, Jick SS, et al. Multivitamin/
to 59 years. Can. Med. Assoc. J. 1992; 147: folic acid supplementation in early pregnancy
1477–88. reduces the prevalence of neural tube defects.
54. Miller AB, Baines CJ, To T, Wall C. Canadian JAMA 1989; 262: 2847–52.
National Breast Screening Study: 1. Breast cancer 68. Smithells R, Sheppard S, Wild J, Schorah CJ.
detection and death rates among women aged 40 Prevention of neural tube defect recurrences in
to 49 years. Can. Med. Assoc. J. 1992; 147: Yorkshire: final report. Lancet 1989; ii: 498–9.
1459–76. 69. Vergel RG, Sanchez LR, Heredero BL, Rodriguez
55. UK Trial of Early Detection of Breast Cancer PL, Martinez AJ. Primary prevention of neural
Group. Breast cancer mortality after 10 years in tube defects with folic acid supplementation:
the UK trial of early detection of breast cancer. Cuban experience. Prenat. Diag. 1990; 10:
The Breast 1993; 2: 13–20. 149–52.
56. Alexander FE, Anderson TJ, Brown HK, et al. 70. MRC Vitamin Study Research Group. Prevention
The Edinburgh randomized trial of breast cancer of neural tube defects: results of the Medical
screening: results after 10 years of follow-up. Br. Research Council vitamin study. Lancet 1991;
J. Cancer 1994; 70: 542–8. 338: 131–7.
57. Thompson RS, Barlow WE, Taplin SH, et al. A 71. Bower C, Stanley FJ. Periconceptional vitamin
population-based case-cohort evaluation of the supplementation and neural tube defects; evi-
efficacy of mammographic screening for breast dence from a case-control study in Western
cancer. Am. J. Epidemiol. 1994; 140: 889–901. Australia and a review of recent publications.
58. Hakama M, Pukkala E, Kallio M, Godenhjelm K, J. Epidemiol. Community Health 1992; 46:
Svinhufvud U. Effectiveness of screening for 157–61.
Observational, Quasi-Experimental Study Designs to Evaluate Healthcare 135
72. Cziezel AE, Dudas I. Prevention of the first occurrent neural tube defects. JAMA 1993; 269:
occurrence of neural-tube defects by periconcep- 1257–61.
tional vitamin supplementation. N. Engl. J. Med. 76. Chatkupt S, Skurnick JH, Jaggi M, et al. Study of
1992; 327: 1832–5. genetics, epidemiology, and vitamin usage in
73. Kirke PN, Daly LE, Elwood JH, for the Irish familial spina bifida in the United States in the
Vitamin Study Group. A randomized trial of low 1990s. Neurology 1994; 44: 65–70.
dose folic acid to prevent neural tube defects. 77. Shaw G, Schaffer D, Velie EM, Morland K,
Arch. Dis. Child. 1992; 67: 1442–6. Harris JA. Periconceptional vitamin use, dietary
74. Martinez-Frias ML, Rodriguez-Pinilla E. Folic folate, and the occurrence of neural tube defects.
acid supplementation and neural tube defects. Epidemiology 1995; 6: 219–26.
Lancet 1992; 340: 620. 78. Elwood M. Critical Appraisal of Epidemiological
75. Werler MM, Shapiro S, Mitchell AA. Peri- Studies and Clinical Trials. 2nd edn. Oxford,
conceptional folic acid exposure and risk of Oxford University Press, 1998.
8
The Potential Use of Routine Datasets
in Health Technology Assessment
disease-specific databases and health surveys. available for analysis; and only through analysis
While adverse event datasets can be useful can data become information.
for assessing the performance of organisa- Conventional wisdom suggests ever more and
tions, they have little else to offer HTA larger randomised controlled trials (RCTs)
except where linkage can be established. measuring differences in efficacy using ‘custom-
4 Census-based datasets are important because ised’ data. The relative merits of trials have
they provide the population denominator for been pitted against other approaches (model-
other analyses. ling, observational studies) which rely on non-
In assessing the cost of health technologies, four customised or routinely collected data. The
different types of cost-related datasets can be limitations of RCTs in providing generalisable
distinguished: health technology plus unit cost information, particularly on effectiveness (as
(prices) directly linked to specific health tech- opposed to efficacy), plus the cost of trials, and
nologies; grouped health technology plus unit the large number of technologies to be assessed,
costs; linkable unit costs, i.e. unit costs based on mean that alternative approaches will continue
combining two datasets, at least one of which is to be explored.
routine; and survey unit costs, i.e. unit costs The information deficit is greater if HTA is
based on special surveys. recognised to include not only the measurement
In interpreting the scope of HTA widely, of efficacy and effectiveness, but – as argued
much evidence already exists of the use of here – also assessment of the diffusion, equity
routine data at least in some aspects of HTA. and cost of health technologies. Diffusion of
However, for mainstream evaluation of effec- health technologies can be seen as closely
tiveness and efficiency, the scope is restricted by related to assessment of equity in access or use,
data limitations and quality problems. There is, in that both issues require data on the use of
however, some potential widening of that scope technologies by time and place or group.
given the growth of high-quality, technology- Assessments of diffusion/equity and cost of
specific clinical datasets, developments in health technologies, by contrast with those
information technology and healthcare coding focused on efficacy, rely heavily on routine data.
schemes. But if routine datasets are to fulfil Routine data (see full definition below)
their potential, a strategic overview is required include not only administrative data, but also
plus support for potentially promising data- registry data, adverse event reporting and regu-
bases, greater linkage, assessment of gaps in lar health-related surveys. The quantity and
current datasets, and greater use of routine data quality of routine datasets vary widely. They
in the clinical (and therefore also potentially depend partly on how particular healthcare sys-
health technology assessment) setting need to tems are organised, with more administrative
explored. data in more market-oriented systems. Registers
of diseases, long established in clinical practice,
are partly an attempt to improve routine data.
Computerisation and automated data collec-
INTRODUCTION tion are dramatically increasing the scope for
registers.
Routine data have arguably been under-used in Improvement of routine datasets requires a
health technology assessment (HTA). A vicious knowledge of their flaws. These include prob-
circle applies; ‘the data are poor, so they cannot lems around data collection (coverage, accuracy,
be used, so they remain limited and of poor frequency, delays in publication), plus limited
quality’. One of the purposes of this report is to scope (imperfect capture of health technologies,
help instil a virtuous circle; ‘good quality data their outcomes, and likely confounding factors).
which, because used, must be improved’. Routine datasets are often poor at picking up
The international ‘effectiveness revolution’ in fine but clinically relevant distinctions between
healthcare cries out for much more research technologies, and at identifying new technolo-
and information. Better information is required gies (often constrained by the standard classifi-
on both the effectiveness and the costs of the cation systems). For example, coronary stenting
multiple and varied health technologies that would not be picked up as a distinct technology
make up healthcare. Healthcare exhibits the in routine hospital statistics despite its high cost
familiar data-information gap in an extreme and questions over its effectiveness. Concerns
way. While masses of detail on patients’ health over confidentiality of data, coupled historically
states and their use of health technologies are with inability to link different datasets, have
collected in most healthcare systems, they gen- prevented technologies and treatments being
erally fail to become ‘data’ in the sense of being connected to patients, and hence to health status.
138 The Advanced Handbook of Methods in Evidence Based Healthcare
Nonetheless, routine data are being used more KEY TERMS AND CONCEPTS
widely in HTA. An analysis of the effectiveness
of in-vitro fertilisation (IVF) based on data col-
lected routinely by the UK Human Fertility and Each of the key terms – routine data, health
Embryology Authority identified factors asso- technology and health technology assessment –
ciated with successful conception, including requires clarification.
age, number of previous children if any, and
duration of infertility1 and in turn formed the
basis of evidence based guidelines on managing Routinea Data
the infertile couple.2
In the UK, Hospital Episode Statistics (HES) In the absence of an unequivocal definition in
data have been used to explore variations in the literatureb, this chapter identifies the follow-
ing important characteristics in routine data-
clinical health status and diffusion patterns. The
setsc:
National Health Service (NHS) clinical indica-
tors published in 1999, covering specific types (a) regular and continuous collection;
of hospital mortality, readmission and discharge, (b) using standard definitions;
are entirely based on HES.3 The impact of GP (c) with comprehensive coverage of the target
Fundholding (arguably a health technology) on group, service or technology;
waiting times,4 prescribing5 and day case rates,6 (d) with some (usually considerable) degree of
has been assessed using NHS routine data. obligation to collect the data completely
In the US, the PORT studies7 used admin- and regularly; and
istrative claims data to assess the effectiveness (e) with collection at national or regional level
of a range of common treatments, but were (depending partly on the degree of compul-
criticised8 for potential bias. Many other studies sion involved).
have explored equity and diffusion issues in the
US context. This definition excludes single site (hospital or
This chapter proposes a more optimistic-than- health authority) datasets, of which there have
usual view of the potential use of routine data been many a often short-lived datasets of limited
in HTA, based partly on a wider definition of generalisability. Some deliberate ‘fuzziness’
health technology assessment, because of encour- remains around the level of compulsion and
aging evidence of several specific registries, dis- about the role of ‘regional’ datasets.
Routine datasets cover not only adminis-
cussed below.
trative data (such as hospital episode statistics),
but go beyond these to include disease or
technology-specific databases established for
AIMS AND METHODS epidemiological research, and evaluation pur-
poses, provided they meet the above criteria. It
thus includes disease registers such as those for
The aim of the project reported here was to
cancer, and technology-specific registers such as
explore how routine data can be relevant to those for renal replacement therapy, which have
health technology assessment, taking into ac- developed in many countries. This wider defini-
count its planned development, rules of access, tion breaks with the (largely US) literature
previous use in HTA, and potential for future which discusses mainly administrative data.
use. Methods used in this review to assess the Some indication of the scope of routine data
role of routine data included: in relation to health and healthcare can be
(a) a review of the literature on the potential gleaned by briefly reviewing the history of UK
use of routine data; routine data.
(b) a structured compilation of sources of The decennial Census of Population since
routine data in the UK, both NHS9 and 1801 has provided the most basic data on popu-
more generally;10 lation size and composition. Mortality data have
(c) information from key informants on char- long been collected, largely for probate reasons.
acteristics of each dataset planned develop- Mortality statistics collected at national level in
ment, and the rules for access; and England since the Births and Deaths Registra-
(d) annotation of each source according to the tion Act 1836, became compulsory in 1874.
aims outlined above. Natural and unnatural causes of death have been
distinguished and classified, partly for legal
While the main focus is on the UK, the proposed reasons, partly for epidemiological reasons.
classifications of HTA and the typology of Stillbirths were added to notifiable causes of
routine datasets will apply to most modern death in 1927, and amended in 1992 in line with
healthcare systems. the changed legal definition of stillbirth. Limited
Potential Use of Routine Datasets in HTA 139
morbidity data, particularly in relation to infec- survey data are available, even though as
tious diseases, have also been required by law to much as 20% of all elective surgery is carried
be collected. out in the private sector.18 Although nursing
Some disease-specific registries have long home and residential care are increasingly pro-
been established, notably for cancer. Registries vided privately (if often publicly funded), few
have been defined as ‘data concerning all cases routine data are required. Routine data on per-
of a particular disease or other health relevant sonal social services are also largely absent in
condition in a defined population such that the the UK.
cases can be related to a population base’.11 The potential usefulness of routine data has
Cancer registries were developed on a voluntary increased due to information technology, im-
basis from 196212 in the UK after follow-up proved record linkage, and coding.
studies of patients treated with radium in the
1930s. Registries for particular diseases (also
known as case registers) have come in and out Information Technology
of existence in the different countries of the
UK.13 Clinicians have often kept designated lists This has dramatically increased the scope of
of patients, identifying their illnesses, severity, routine data in healthcare, as in other sectors of
health service contact and health status. Psychi- the economy. Although much patient-specific
atric case registers were developed in the 1960s information is recorded in the course of health-
but fell out of use in the 1980s.14 More recently care delivery (such as patient characteristics,
there has been interest in registers of diabetes disease severity, co-morbidity, diagnostic and
and stroke.15 treatment data plus some information on out-
Some data on the use of healthcare have also comes), it is usually stored in inaccessible case-
long been routinely collected. Mental hospital notes. A useful distinction can be made between
admissions data date from the 1845 Lunacy Act. ‘data’ (given) and ‘capta’ (captured or taken).
Compulsory detention required such data for Health services capture enormous amounts of
legal reasons, and covered patients in both pub- detail, often in casenotes, but these are seldom
lic and privately funded hospitals.16 Data on translated into data. Casenotes in England still
admissions to other hospitals came later. NHS
tend to be handwritten,19 specific to the site of
acute hospital data have been collected and
treatment, and difficult to interpret. Problems
published from 1948; these include demo-
include lack of standard formats, variation
graphic, administrative and some clinical head-
ings (surgical interventions and ICD coding of between doctors in completeness and accuracy
conditions, but generally not medical interven- of data recording (such as the difficulty in defin-
tions or technologies). Although these are not ing cancer stage from medical notes), poor inte-
subject to legal statute, they have the force of gration with other professions’ entries (nurses,
administrative law. physiotherapists), or with other settings of care.
The establishment of the NHS in 1948, by A high proportion of casenotes have been
covering the entire population increased the untraceable in various studies. Computerisation
scope for use of comprehensive data. Of particu- of casenotes offers great scope for their
lar note is the NHS Central Register, which has ‘routinisation’. Electronic patient records have
been used since 1948 to pay GPs in relation to already become a reality in some hospitals and
the number of patients registered. This register, their widespread use is an objective of the NHS
based on a special 1939 Census of Population Information and Communications Technology
and subsequently used for rationing food in the strategy.20
Second World War, was computerised in 1991.
The Central Health Register Inquiry System
(CHRIS) contains ‘flags’ relating to both cancer Linking of Data
registration and membership of some 240 on-
going medical research studies, in which around Linking of data between routine datasets pro-
2 million people or some 4% of the UK popula- vides a powerful way of greatly extending their
tion, are included.17 use as well as validating data items. Besides the
Data on privately provided services have NHS Central Register, more detailed linkage
often been lacking in the UK except for services including hospital records has been pioneered
whose legal regulation required such data (com- in the Oxford Region and in Scotland.21,22 His-
pulsory psychiatric admissions, terminations, torically, the lack of a single computable NHS
IVF). In the 1990s, both US and France imposed patient number has limited health record link-
a legal obligation on private providers of acute age, but the change to a single NHS number
hospital services to provide minimum data- from 1997 greatly increased the scope for such
sets to the state. In England, only occasional linkage.
140 The Advanced Handbook of Methods in Evidence Based Healthcare
If healthcare ‘capta’ are to be turned into and to improve rehabilitation and long term
useful data, standardised coding systems of dis- care’.25
eases and health technologies are essential. The And in the US:
International Classification of Diseases (ICD)
was originally developed for classifying causes ‘the set of techniques, drugs, equipment and
of death but from 1948 was extended to cover procedures used by health care profession-
morbidity.23 ICD has been regularly updated als in delivering medical care to individuals
to keep up with new diseases (currently on and the systems within which such care is
ICD10), and has the major advantage of being delivered’.26
international. The ability of routine data to capture health
Standard definitions of treatments have been technologies thus depends on how well defined
slower to develop. Surgical interventions are the health technologies are, and the purposes of
captured in different ways in different countries. the dataset. Those technologies that are regu-
The UK uses the OPCS codes (Office of Popula- lated by law tend to have fuller, more precise
tion Census and Surveys – since changed to definitions. Pharmaceuticals make up a specific
Office of National Statistics), and the US the class of health technologies, due to having long
Classification of Procedures and Treatments been subject to licensing in most countries both
(CPT). Standard definitions apply to drugs via to ensure quality and to prevent harm. Most
the British National Formulary (which can be pharmaceuticals are manufactured and marketed
linked to both World Health Organization and by private enterprises. The regulation of phar-
European Union classifications, but which do maceuticals reflects a history of harm and sub-
not readily map to International Classification of sequent litigation. The most notorious episode
Disease headings). was probably the thalidomide scandal in the
Several systems,24 including Read, offer 1960s, which led to tighter controls in many
means of translating all terms used by doctors to industrialised countries. Within pharmaceuti-
describe diseases, conditions and treatments into cals, different regulations apply depending on
standard codes by way of an electronic thesaurus. whether drugs are available only on prescrip-
Despite the NHS adoption of Read codes, their tion, over-the-counter in pharmacies only, or in
development has been slow. all shops.
Even when codes exist, their usefulness is Medical and surgical devices tend to be less
limited by the lack of detail in most standard regulated, but since they are often privately
coding definitions. For example, ICD does not provided, they are also subject to litigation (the
cover disease severity and in practice, many extent of which varies internationally). Litiga-
conditions are coded as ‘unspecified’. Whilst tion has led to the establishment of registers
procedure coding systems (such as OPCS and of patients who have had certain devices
CPT) capture detail of the surgical interventions, implanted, for example cardiac pacemakers or
detail is lacking on the specific surgical tech- silicone breast implants. The main regulation of
nologies used. All standard coding systems diagnostic testing has to do with quality control,
inevitably lag new diseases and procedures. such as the national external quality scheme for
Health status is seldom recorded, partly due to pathology laboratories.
problems of the coding systems (no severity), Other health technologies are subject to less
but also due to limitations with ICD (no quality regulation and tend to be less well defined.
of life measurement, or disease-specific symp- Surgical procedures are recorded in routine data-
tom profiles). These difficulties are exacerbated sets for in-patient and day cases (but not for out-
by variations in medical diagnosis. To remedy patients or when carried out in GP surgeries),
these limitations, clinical trials usually go but are not regulated other than by clinical audit
beyond standard classifications by using more of outcomes (both in general and of adverse
detailed terms. As argued below, clinical trials events), and voluntary means such as the UK’s
provide criteria against which to gauge the Safety and Efficacy Register of New Interven-
potential usefulness of routine data in assessing tional Procedures (SERNIP).
efficacy and effectiveness. Changes in the setting or organisation of
healthcare are less regulated, but the recent
history of community care for the mentally ill
HEALTH TECHNOLOGIES may indicate a move towards auditing of
adverse events (spurred in part by media
reports). The move in many countries to
‘Health technology’ tends to be defined broadly. increased reliance on primary care may have the
In the UK health technologies are defined as: effect of reducing the extent to which routine
‘all methods used by health professionals to data systems capture treatment data (such as
promote health, prevent and treat disease, minor surgery). Against that, primary care
Potential Use of Routine Datasets in HTA 141
organisations may over time come to collect (a) clear identification of the health technology
more detailed relevant data than at present. along with relevant comparator health
technology;
(b) measurement of patients’ health state at
patient or population level, before and after
HEALTH TECHNOLOGY ASSESSMENT intervention; plus
(c) relevant patient characteristics such as
Health technology assessment is widely used to severity, co-morbidity and socio-demo-
include assessment of efficacy, effectiveness and graphics, and any other likely confounders.
cost effectiveness. This section suggests broad-
ening the definition explicitly to include equity, Assessment of Diffusion and Equity of
diffusion and cost. For the purposes of under- Health Technologies
standing the scope of routine data, assessments
can be classified as measuring: Equity can be defined in terms of which groups
(a) efficacy and/or effectiveness; use particular health technologies. Assessment
(b) equity and/or diffusion; and of equity requires that datasets show which
(c) the cost impact of health technologies, particular groups receive the relevant tech-
whether alone or in conjunction with their nology. Groups can be defined in many ways
efficacy/effectiveness. (age, sex, ethnicity, socio-economic group, geo-
graphy). Diffusion is concerned with the factors
influencing the uptake of health technologies by
Assessment of Efficacy and/or place and time, and requires broadly the same
Effectiveness information as equity. Both require data on the
use of technologies by time, place and relevant
Efficacy measures patient benefit and short-term group.
harm in scientific, usually experimental, studies. Health technology assessments concerned
Effectiveness, by contrast, is concerned with with equity or diffusion need less information
patient benefit and harm in everyday practice. than those concerned with efficacy or effective-
Health technology assessments concerned with ness. It remains necessary to have the health
efficacy place particular emphasis on RCTs and technology specifically identified at one or more
systematic reviews of trials. The conventional points in time and place. Some NHS routine
wisdom is that only trials can provide the datasets have been extended in recent years to
requisite data on efficacy/effectiveness, and per- include group identifiers (ethnic group). Post-
haps on cost effectiveness. However, effective- coding of patients’ records in the UK has
ness can only be measured in costly large prag- enabled service use to be linked to the Census of
matic trials with wide entry criteria. The degree Population enumeration districts allowing eco-
to which efficacy as established in trials trans- logical analysis of socio-economic factors.30
lates into everyday practice depends on many
factors, including the transferability of the Assessment of the Costs of Technologies
technology and its application to groups not
included in trials. To be useful for costing, datasets must identify
The advantages of trials in measuring efficacy the relevant health technology and its cost. Cost-
have been well rehearsed.27 The random alloca- related assessments include the range of econ-
tion of patients to intervention and control omic evaluations: cost efficacy, cost effective-
arms prevents differences in patients (including ness, ‘cost of illness’, cost consequences and
selection bias and confounders unknown to the cost impact studies. Costing is complicated by
experimenter) affecting the results. The blinding several factors, including the wide definition of
of outcome assessment minimises the possibility cost used by economists and the lack of prices
of information bias. However, the limitations of or unit costs (used equivalently here) for many
trials in relation to efficacy have also been health technologies. Economics defines costs in
outlined.28 Assessment of effectiveness requires terms of societal opportunity costs, which
both large simple trials and detailed registers of requires data on the range of knock-on effects of
particular diseases and their treatments.29 particular health technologies (both in terms of
If routine datasets are to help assess effective- the full range personal costs and benefits, and
ness, they should possess the key characteristics over the entire life of the patient). Difficulties
of data generated by trials, including informa- around the assessment of such knock-on cost
tion on patients’ health before and after treat- effects are most extreme in trials (which tend to
ment, together with data on relevant confounders. be short term and focused on outcomes), but is
Specifically this implies: shared by all approaches.
142 The Advanced Handbook of Methods in Evidence Based Healthcare
Only some datasets that identify a health Type I: Health Technology 1 Patient
technology have a price linked to the technol- and/or Health Status Datasets
ogy. Pharmaceuticals and some packages of
healthcare tend to be priced, depending on the These databases specify a particular health tech-
healthcare system, but few other health technol- nology plus patient characteristics (such as age,
ogies are priced. Costing then involves estimat- sex, condition, severity, co-morbidity, and socio-
ing unit costs for those technologies. Prices (and demographics). Although generally patients are
unit costs) vary, depending on patent, industrial identified directly, for some population-oriented
processes, and competition. programmes only aggregate population charac-
For costing purposes, the key criteria are that teristics are needed. Within this premier group
the health technology is identified in the dataset of databases, three subtypes can be distinguished:
and linked to either a price/unit cost or to a registries; enhanced administrative datasets; and
pattern of resource use which can be priced/ population-based datasets.
costed. Ideally, the knock-on effects of use of
one technology on use of others would be identi-
fied, but as the requisite long-term follow-up
Registry-type databases
data are rarely available, this usually involves New disease-specific or health technology-
modelling. specific databases have been developing, partly
due to the limitations of traditional routine
administrative datasets. Some are statutory and
ROUTINE DATASETS FOR EFFICACY/ many are led by clinicians. These are routine in
EFFECTIVENESS AND EQUITY/DIFFUSION the sense of being regularly collected, but are
registries in reporting additional clinically rele-
vant data and following-up patients.
The above discussion suggests four broad types Only registry-type datasets are likely to even
of datasets relevant to assessing issues around approach the criteria required for measurement
efficacy/effectiveness or equity/diffusion: of effectiveness. The criteria require measure-
ment of patients’ health state, before and after
I Health technology, plus patient and/or
intervention, plus relevant patient character-
health status;
istics such as severity, co-morbidity and socio-
II Health technology only;
demographics, and any other likely confounders.
III Health status only;
While good quality registry-type datasets in-
IV Census based datasets.
clude relevant patient health states, the possibil-
A similar but less refined classification has been ity of bias due to uncontrolled circumstances is
suggested by WHO researchers.31 This classi- difficult to overcome.
fication is applicable widely and not just to the The ability of registries to contain the data
UK. required for assessing effectiveness depends on
For assessing effectiveness, only datasets in the condition or disease and the relevant health
type I are potentially useful, due to their inclu- technologies. Some outcomes can be more read-
sion of both health technology and health status. ily linked to well-defined effective interventions
Health status can be measured directly (as in or patient characteristics. For example, in the
registers and to some extent in enhanced admin- Human Fertility and Embryology Authority
istrative datasets) or indirectly (through linkage database, successful conceptions are the out-
based on the individual patient records or at come of IVF, with relevant details of patient
aggregate population level). Type I datasets can groups and the interventions. In some data-
also be used to assess equity/diffusion, as they sets, extension of life is the main target out-
usually include data on groups and place, by come. Examples include the Renal Registries in
time. England & Wales and in Scotland, and the UK
The other types fail to include both health Transplant Support Services Authority (TSSA).
technology and health status. Type II contain The Renal Registry collects data on the character-
health technologies only and so can potentially istics of all new patients starting renal replace-
be used for equity/diffusion analysis. Type III ment therapy, types and details of treatment, and a
contain health status only and can only excep- range of health statuses including death, modal-
tionally be linked back to a particular health ity switches and intermediate biochemical and
technology. Type IV refer to Census of Popula- haematological markers. As planned, this data-
tion datasets which provide the denominator base will eventually give long-term health
data for rates and which can sometimes be statuses – including those rare outcomes which
linked to other datasets. Systems enabling link- are beyond trials. Some registries provide proxies
age to mortality or to Census of Population data for health status. The National Prospective
are discussed separately below. Monitoring Scheme for HIV provides not only
Potential Use of Routine Datasets in HTA 143
data on drugs prescribed but also on CD4 blood accompanied by quality checks, may over time
count (which defines the transition from HIV to lead to quality improvement.
AIDS). Although in-patients in NHS hospitals are
Cancer registries provide data on the incidence, recorded in HES, only limited data are collected
stage and progression of disease. Although they on out-patients (ambulatory care), GP and com-
contain some treatment data, their lack of munity health services. The Minimum Data Set
precise data on specific forms of chemotherapy for outpatients, introduced in 1996 (yet to
and radiotherapy, and on subsequent healthcare report) is limited due partly to the nature of
use, limits their usefulness in health technology ambulatory care – much of the service is either
assessment. The effectiveness of interventions diagnostic or monitoring with many repeat
varies both by cancer site and stage and specific visits. The large number of contacts with GPs
intervention. Nonetheless, differences in sur- means that no routine patient-based data are
vival rates internationally for particular cancers collected centrally, other than that for the pay-
have spurred action in the reorganisation of ment of prescriptions, so that analysis of trends
cancer services. Cancer registries arguably need relies on surveys such as the periodic GP Mor-
to develop their ability to capture the diffusion bidity survey or various ongoing data collections
and equity of use of major evidence based from samples of practices, such as the General
interventions such as adjuvant chemotherapy for Practice Research Database (GPRD) and the
particular stages of cancers. Doctors Independent Network (DIN). The scale
and complexity of diagnostic imaging means
that only rudimentary data are collected.
Enhanced administrative datasets
The main UK dataset under this heading is the Aggregated datasets useful for population
Hospital Episode Statistics (HES) and its Welsh effectiveness
and Scottish equivalents (PEDW and COPPISH).
These large databases (10 million records Although patients are not identified at an indi-
annually in England alone) contain demo- vidual level in these datasets, population level
graphic, administrative and clinical data of data can be used to assess the effectiveness of
cases treated as in-patients or daycases in NHS specific population-oriented interventions. The
hospitals. By including surgical interventions, as national immunisation and vaccination pro-
captured by OPCS4 surgical codes, a range of grammes are examples. Since the effectiveness
important health technologies is identified. Dis- of these interventions is well established, data
eases are noted using ICD codes, but medical on the coverage of these programmes enables
treatments and diagnostic tests, which are item- their overall effectiveness to be assessed. Screen-
ised in casenotes, are not included. ing programmes, such as national breast and
Within HES, important health technologies, cervical programmes, where again coverage,
such as stroke unit and ICU beds are not dis- incidence and mortality are the critical target
tinguished. Health status data in HES are limited variables, also fall under this heading.
to ICD coding of events plus deaths while in While such datasets can be used for selective
hospital, as in the 1999 NHS Clinical Indicators analyses of effectiveness/efficacy, they can also
which focus on in hospital mortality post sur- be used for equity/diffusion and costing. Those
gery or admission with a fractured hip or a discussed in the next section can only be used
myocardial infarction.3 Mortality rates post for diffusion/equity and costing.
discharge require linking to the NHS Central
Register. ‘Destination on discharge’ to normal Type II: Health-Technology-Only
home provides a crude proxy outcome for cer- Datasets
tain conditions – for stroke and fractured hip in
the NHS Clinical Indicators. More specific out- Datasets in this group contain data on use of a
come measures are being developed for the health technology by time and place, but without
NHS, including mortality and complications any linkage to patients or their conditions. Such
after discharge, which require additional linked datasets can be used to assess the diffusion by
data. The use of the single NHS number in this geographical area, and hence in assessment of
dataset will enable records to be linked from equity. A prime example is the Prescription
year to year and to other datasets. Lack of data Pricing Authority data which specifies the
on out-patients and on diagnostic data remain amount of each drug dispensed by GPs in each
major barriers. HES has been criticised, not least UK country each year. These data are fed back
for its definition of finished consultant epi- to GPs and play a role in monitoring and con-
sodes,32 but it has also performed well when trolling prescribing. Other databases in this
validated against one disease register.33 Its use heading include the National Breast Implant
in the Clinical Indicators, which has been Register (which records who has had silicone
144 The Advanced Handbook of Methods in Evidence Based Healthcare
implants) and the family planning returns (NHS technologies. Without linkage to health tech-
Return KT31). nologies these datasets have little to offer health
technology assessment. Some surveys such as the
Type III: Health-Status-Only Datasets Health Survey for England collect data on patient
health states and on prescribed drug use.
Databases in this group specify health status or
health states, but without necessarily reporting Type IV: Census-Related Datasets
either the health technology or the patient. Three
subtypes can be identified: adverse events/ The decennial Census of Population which con-
confidential enquiries, disease-specific databases, tains several questions on long-standing illness
and health surveys. While adverse event-type must be included in any inventory of datasets,
datasets can be useful for assessing the perform- not least because it provides the population
ance of organisations, the other two types have denominator for all epidemiological analyses.
little to offer HTA except where event linkage, Beyond this, a number of long-standing studies
such as to mortality, can be established. based on record linkage to the Census have
developed, most notably the Longitudinal Study,
Adverse events/confidential enquiries which links an anonymised 1% of the 1971
The Adverse Drug Reaction notification system Census to subsequent censuses, and to vital
identifies adverse events associated with particu- events (deaths, births) using the NHS Central
lar drugs, with doctors reporting on a voluntary Register. Only by linkage to datasets which
basis by filling in a widely available yellow identify a health technology can such datasets be
card. Individual pharmaceutical companies carry used to assess health technologies, and then
out similar monitoring via ‘post-marketing sur- mainly for assessment of equity/diffusion and
veillance’. The Medical Devices Agency has for effectiveness as captured by mortality.
recently developed a voluntary reporting scheme
for adverse events associated with use of par- Linkage-Enabling Systems
ticular devices.
The four Confidential Inquiries run by the The two main UK linkage-enabling systems are
English Department of Health which audit Central Health Register Inquiry Service (CHRIS)
selected causes of death (stillbirths, perioper- and postcodes. Linkage by CHRIS by means of
ative deaths, maternal deaths, suicides) also fit the NHS number enables records with that num-
under this heading. These provide an assessment ber to be linked. The most immediate scope
of the relevant system and can indicate the need applies to Hospital Episode Statistics, along the
for more detailed inquiry. While adverse event lines of the Oxford Record Linkage Study. As
reporting clearly has an important role to play, it the use of the NHS number gradually extends to
also has severe limitations. First, the process cover other health services, more integrated
means that only extreme events are likely to be records can be built up. The new NHS number,
detected. Second, reporting is usually voluntary. which became mandatory in 1997, will in time,
Third, lack of data on the totality of events assuming full coverage and accuracy, enable
means that rates cannot be estimated. Fourth, greatly improved linkage.
limited control population data prevents estima-
tion of the relative magnitude of any effects.
Finally, the necessary standardised definitions ASSESSING THE COST OF HEALTH
required for comparisons over time and place, TECHNOLOGIES
are not always available.
(b) Grouped health technology plus unit cost. Linked Health Technology Unit Costs
These are the unit costs of groups of
different, related health technologies or Many cost evaluations of health technologies
treatments. (other than drugs and hospital stays) have had
(c) Linkable unit costs, that is unit costs based to improvise, either by combining customised
on combining of two datasets, at least one data with routine data or by special surveys.
of which is routine. Examples of the former include those evalu-
(d) Survey unit costs, that is unit costs based ations which have combined data on number
on special surveys, whether one-off or and duration of contacts with healthcare person-
repeated or updated annually. nel and valued these using national pay scales.
Remarkably few cost-related datasets specify The most usual source for pay scales has been
health technologies. Of the 127 NHS Financial the Annual Pay Review Body Reportsd.36,37 Any
Returns for England only eight could be linked moves to local pay levels would pose problems
to health technologies. Two main exceptions for this approach. The reported 14 ways of
exist – the costs of drugs dispensed by GPs (as costing a GP consultation38 result from the
published by the Prescription Pricing Authority) lack of agreed routines in use of such data.
and the English NHS acute hospital reference However, standardised estimates are now pro-
costs. duced annually for these and a range of other
personnel such as district nurses and health
Health Technology plus Unit Costs visitors.39
The most readily available unit costs are prices, Survey Health Technology Unit Costs
which apply when the NHS purchases a tech-
nology from some other organisation. Prices for The annual report from the Personal Social
products purchased from the pharmaceutical Services Research Unit ‘Unit costs of com-
industry provide the obvious example. munity care’,39 provides an example of the use
Routine data on unit costs of non-drug health of surveys to establish unit costs. This important
technologies has historically been much less source contains unit costs not only for consulta-
common. The National Schedule of Reference tions but also for packages of care such as for
Costs34 provides data on cost per surgical health- mental health services. Estimated on a standard-
related group (HRG) in each acute NHS hospital ised basis, these rely heavily on surveys carried
trust in England. Since HRGs are based on the out for research projects. A major advantage is
HES data, their level of specificity is limited by its use of consistent and transparent methods for
the coding systems used there. The aim is to roll deriving unit costs.
out this approach to costing the medical special- This group of datasets might also include
ities and to reconcile the data with Trust annual individual examples of bottom-up costing to the
accounts35 – which should improve the accuracy extent that such costing is achieved without use
of HRG costs. HRGs can be divided into those of any routine data. Although such an approach
which identify a single technology, such as total is recommended in recent guidelines,40 few UK
hip replacement, and those which comprise a studies have attempted this.
group of technologies. The former fit group (a)
in being health technology plus unit cost, while
the latter make up group (b) grouped health CONCLUSIONS
technology plus unit cost.
This chapter has suggested that the potential use
Grouped Health Technology Unit Costs of routine data varies by type of HTA. While
there is scope for the use of routine data in each
HRG unit costs that refer to a group of health of the three types of HTA we have distin-
technologies, typically surgical interventions, in guished, i.e. the measurement of efficacy and
a particular specialty are similar to the average effectiveness, of equity and diffusion and of
cost per FCE per specialty which has long been costs, that scope is greater the less the focus of
available via the Trust Financial Return 2 the HTA is on efficacy. Where possible, efficacy
(TFR2). Health technology assessments have should be assessed in trials, partly because of
tended to use these in both cost per finished con- requirements in relation to measurement of
sultant episode and per in-patient day. Many but health status but particularly in relation to con-
not all of the non-surgical costed HRGs provide trolling for confounding and bias. Routine data-
unit costs for similar groups of cases (casemix), sets have an essential role in monitoring effec-
which necessitates further estimation to obtain tiveness, especially where large and long-term
the unit costs of particular interventions. pragmatic trials are not possible. To play this
146 The Advanced Handbook of Methods in Evidence Based Healthcare
role however, they need development. For differential patterns of service use. Ingenuity in
assessment of equity and diffusion and for cost- linking these to other datasets, notably to the
ing, routine data are widely used, but may also Census of Population, has indicated problems
need development. and led to policy changes (e.g. capitation formula
A major encouraging theme emerging from based on linking census to hospital use data).
the review of routine datasets was the growth of Finally, a number of issues require attention if
high-quality technology-specific clinical data- routine datasets are to fulfil the potential out-
sets in the UK. These offer scope for assessing lined above. Those databases collecting data on
the effectiveness of key technologies, as well as health technology, patients and health status
assessing diffusion and equity. Few of these need to be supported, perhaps initially by means
databases include costs, however. of a strategic overview. The scope for greater
Information technology is rapidly increasing linkage of datasets requires active exploration,
the scope for use of routine data in health including facilitation and monitoring of the use
technology assessment. The increasing data cap- of the unique patient identifiers. The gaps in
ture of details (capta) of healthcare using com- existing datasets must be addressed. These, we
puters is transforming the meaning of routine have suggested are largely around diagnostic
data. Although healthcare spending on com- activity, medical procedures in hospital and out-
puterisation has lagged other sectors,41 there are patient and ambulatory activity. Regular surveys
already signs of how healthcare is being trans- of organisation and delivery of care, such as
formed. Electronic patient records, computerised stroke units, should be seen as part of routine
work stations, bar-coding, use of smart cards, intelligence. The lack of data in the UK from the
and telemedicine are only some of the key private sector needs to be remedied. More sys-
developments. Improved methods of coding tematic surveillance of adverse events needs to
(Read) and grouping (Healthcare Resource be encouraged, including use of standardised
Groups), again relying on information technol- definitions and fuller coverage. Improved train-
ogy, may facilitate extraction and collation of ing should be provided in the use of routine
real-time data which may for the first time datasets and their potential use in the different
contain clinical health state data. The sensitivity kinds of health technology assessment.
and specificity of these data will depend on the
underlying coding systems used, which will
always be partly out of date and hence of less NOTES
relevance for new health technologies. The chal-
lenge may be around how best to absorb and
interpret the enormous amount of routine data a. The Shorter Oxford English Dictionary defines
that will shortly be available within healthcare. ‘routine’ as follows:
Health technology assessment has the oppor- ‘1.a. Routine- a regular course of procedure, a more
tunity to become routinised to the extent that it or less mechanical or unvarying performance of
rises to this challenge. certain acts or duties.
For the above reasons, the common equations 1.b. A set form of speech, a regular set or series (of
‘routine 5 rubbish’ and ‘customised 5 good’ is phrases etc)
much too simple. The scope for reliable, vali- 2. regular, unvarying or mechanical procedure or
dated routine data is increasing rapidly, and discharge of duties’.
there are examples of excellent datasets which
have already been used for assessing health b. Literature searches provided very few references
technologies such as IVF. That said, it must be under ‘routine data’ but very many under ‘admin-
noted that the bulk of routine administrative istrative data’, particularly from countries whose
datasets collected by the NHS (and other health healthcare systems generate masses of administrative
systems) are of no use to health technology data (notably the US).
assessment. Ideally to be useful for health tech- c. The term ‘Datasets’ is used to refer to excerpts
nology assessment, a dataset should identify a from databases. In the past, datasets were often pub-
health technology, the patient, the health status lished one-off analyses of paper based datasets. The
and its unit cost. One UK dataset which has long computerisation of databases means that what matters
identified unit cost – that from the Prescription in assessing their usefulness is the extent to which the
Pricing Authority – lacks any data on patients, relevant fields are included, such as use of a health
thus precluding any information on health status technology, patient identification, and outcomes.
or equity. Some of the more recent develop- d. Both the Review Body for Nursing Staff,
ments with registry-type datasets could include Midwives, Health Visitors and Professions Allied to
unit cost, but none does so at present. Medicine and the Review Body for Doctors and
For assessment of equity and diffusion, Dentists Remuneration publish annual reports which
routine datasets provide clear indications of include data on pay scales.
Potential Use of Routine Datasets in HTA 147
in Oxford hospitals. Health Trends 1997; 29: 37. Review Body for Doctors and Dentists. London:
96–9. Department of Health, annually.
34. NHS Executive. The National Schedule of Refer- 38. Graham B, McGregor K. What does a GP consulta-
ence Costs (NSRC) and National Reference Cost tion cost? Br. J. Gen. Pract. 1997; 47: 170–2.
Index (NRCI). Leeds: NHSE 1999. (available on 39. Netten A, Dennett J. Unit costs of community
http://www.doh.gov.uk/dhhome.htm) care PSSRU. Univ. of Kent, annual 1998.
35. Raftery J. Benchmarking costs in the NHS. Journal 40. Gold M, Siegel J, Russell L, Weinstein M. Cost
of Health Service Research and Policy 1999; 4: effectiveness in health and medicine. Oxford Uni-
63–4. versity Press, 1996.
36. Review Body for Nursing Staff, Midwives, Health 41. Anonymous. Healthcare lags in spending on
Visitors and Professions Allied to Medicine. information technology. Financial Times, 23
London: Department of Health, annually. November, 1998.
9
Using Routine Data to Complement and
Enhance the Results of Randomised
Controlled Trials
clinically relevant difference that the costs of a Routine data may also be used as a supple-
study would be difficult to justify in terms of its ment to clinical trials for subsequent Bayesian
potential benefits. analysis. This may be for a specific trial, when
When a trial is feasible, its design is generally the data provide prior information regarding
based on specifying the desired power and sig- the expected outcomes of one or more of the
nificance levels. Routine data may allow reliable treatment groups, or during subsequent meta-
estimates of the treatment effect likely to be analysis, when the routine data may be regarded
observed in a clinical trial, with the consequence as comprising a large single or multi-centre trial.
that power calculations will be more accurate or Such analyses take the form of a weighted aver-
realistic. Estimating the ‘patient horizon’ (the age of the information provided by the routine
number of present and future patients, including data set and that provided by the trial, with the
those taking part in the trial) can only be done weighting depending on both the numbers
realistically by using routine data; this has involved (favouring routine data) and the like-
implications not only for the estimation of lihood of bias (favouring RCT data).9
sample size but also for the method of treatment Routine data may be used before a trial begins
assignment.2 by providing estimates of likely event rates or
Even when trials have been completed, even treatment effects. They can thus be of
important questions are often left unanswered. value when performing power calculations, esti-
For example, the use of angiotensin-converting mating sample sizes or deciding when to stop a
enzyme (ACE) inhibitors following an acute trial. The decision to stop or continue a trial is
myocardial infarction is arguably one of the made by weighing the consequences of the pos-
most comprehensively researched questions in sible actions, averaging over the distribution of
clinical medicine, and yet it is still unclear future observations.10 Observational data can
whether the use of such drugs should be then be used in conjunction with trial data to
restricted to ‘high-risk’ patients. Equally, it is improve estimates of the probability that an
apparent difference occurred by chance, lead-
not clear how soon the treatment should be
ing to the trial being either stopped early or
initiated after the myocardial infarction, or for
continued until decisive evidence has been
how long it should be continued.3,4 Data also accumulated.
exist which show that in spite of overwhelming
evidence of the efficacy of these drugs, their
uptake is very low.5,6 The Impact and Uptake of New
There is thus a need to augment information Technologies
from clinical trials: this project explores the uses
of routine NHS data in this context. We examine After the initial introduction of new technology,
the use of routine data, firstly, as a way of there is a period before it becomes a recom-
‘sharpening’ the treatment comparisons derived mended treatment during which its use may be
from RCTs, either by informing the trial design, refined and clinicians gain experience with it.
or – more directly – by using routine data to Little is known about the rate of uptake of new
estimate treatment effects and, secondly, as a technologies and the factors which influence this
means of exploring the impact of RCTs on process. As an example, fewer than 10% of
clinical practice and patient outcomes. hospitals in Pennsylvania were using laparo-
scopic cholecystectomy one year after the intro-
duction of the technique; after a further year,
Sharpening Treatment Comparisons this proportion had increased to 80%.11
When an RCT recommends the uptake of
Although RCTs may be the only unambiguous a new technology, it anticipates an expected
method of equalising risk among groups of improvement in outcomes. Routine data make it
patients receiving different treatments,7 there possible to examine observed outcomes and to
may be ethical or financial constraints which evaluate whether the expected benefits have
mean that RCTs are not feasible; in these situ- been realised. This may be in terms of the
ations, forming pseudo-trials from routine data difference between alternative treatments (cur-
is an alternative. There are limitations and prob- rent and historical) or differences between
lems associated with this approach: routine data centres and subgroups of patients. If a treatment
sets tend to include limited information on is recommended for a particular subpopulation,
potential confounding variables (such as educa- then it is possible to assess the extent to which
tional status or smoking), and the allocation of appropriate patients have been targeted, whether
patients to treatments will tend to reflect clinical the benefits are also seen in other subpopula-
factors whose effects cannot be fully captured in tions and whether there has been a change in
a covariate adjustment.8 trends in outcomes (such as survival rates).
Using Routine Data to Enhance Results of RCTs 151
Large routine data sets also permit assessments data set created by the Information and Statistics
of the importance of an RCT result, particularly Division of the NHS in Scotland.12 Summaries
for rare conditions or when rare outcomes are of hospital in-patient episodes are completed on
being considered. Scottish Morbidity Record Form 1 (SMR1) for
100% of non-psychiatric and non-obstetric in-
Case Studies patient and day cases in Scotland, and provide
some information about the background of the
The two key areas in which routinely assembled patient as well as detailing case management,
data have uses in the field of clinical trials are diagnoses and surgical procedures. The data are
for improving treatment comparisons and for linked by individuals and are also linked to
assessing the generalisability and impact of the statutory death records from the Registrar Gen-
results of RCTs. The two case studies presented eral for Scotland and registrations in the Scottish
below illustrate their potential for: Cancer Register. They provide information
about hospital care and mortality for individuals
(a) sharpening treatment comparisons: as an in the whole Scottish population for the period
aid to power calculations; as the entire 1981–1995, and are unique in Britain. Because
basis of an analysis; or as a supplement to the SMR1 records and the death records both
a meta-analysis; and include the individual’s postcode sector of resi-
(b) putting trial results into context in terms of: dence, they can be related to population denomi-
the representativeness of their populations; nators and socio-demographic area descriptions
the uptake of new technologies; and the derived from the 1991 census. The content of
impact of new technologies on outcomes. these records is listed in Table 1.
Patient ‘identifier’ Not permitting the identification of individual patients, but a unique
number which permits linkage of hospital discharge, death and
cancer registry records
Continuous in-patient stay marker SMR1 forms are completed for each episode of care; a patient may
have several episodes within a period of hospital stay
Hospital code The identifier for the hospital attended
Area of residence Health Board, Local Government District, postcode sector and census
enumeration district
Age The patient’s age in months
Sex
Marital status
Admitted from Home, other NHS hospital, other unit in this hospital, other
Type of admission Deferred, waiting list/diary/booked, repeat admission, transfer,
emergency
Wait Days on the waiting list
Discharge code Irregular (e.g. self discharge), home, convalescent hospital or home,
other hospital, local authority care, transfer to other specialty in
same hospital, died
Date of admission
Date of discharge
Date of operation
Length of stay
Specialty
Diagnoses Up to six diagnoses, coded using ICD9 (WHO, 1997)38
Operation Up to four procedures, coded using OPCS4 (OPCS, 1987)22
152 The Advanced Handbook of Methods in Evidence Based Healthcare
not have been possible, for example, to use the rate that is almost three times greater than those
linked SMR1 database to investigate the uptake of other parts of the world.15
of thrombolysis for the treatment of acute myo-
cardial infarction.
Given this restriction, we chose first to study Clinical Presentation
the timing of surgery following aneurysmal sub-
arachnoid haemorrhage (SAH) as a way of With a major haemorrhage, consciousness is lost
informing the design of a trial of early versus rapidly, and death can occur very quickly.
late surgery. The second study compares per- Smaller leaks lead to a severe headache, with
cutaneous coronary angioplasty (PTCA) with a characteristic rapid onset. Despite its charac-
coronary artery bypass grafting (CABG) as teristic presentation, SAH is commonly mis-
approaches to coronary revascularisation. This diagnosed17–19 – more severe cases with head-
latter study illustrates the use of routine data ache, vomiting and neck stiffness can be
to augment a meta-analysis of RCT data, and diagnosed as meningitis, resulting in admission
to explore the uptake and impact of the two to an infectious diseases unit – but misdiagnosis
procedures. can have serious outcomes because a minor
‘warning leak’ may precede a potentially devas-
tating rupture, which could have been prevented
CASE STUDY 1: THE TIMING OF SURGERY by early intervention.14
AFTER SUBARACHNOID HAEMORRHAGE
Diagnosis
Introduction
CT scanning is the key initial investigation: if
Spontaneous aneurysmal subarachnoid haemor- the scan is performed within 24 hours of the
rhage is a common and often devastating event: onset of symptoms, then a high-density clot in
there are approximately 6000 cases each year the subarachnoid space will be identified in
in the United Kingdom. The case fatality rate over 90% of cases,14 but diagnostic sensitivity
is of the order of 40%, and there is significant declines progressively after the first day. A
residual morbidity in about 50% of survivors.13–16 negative CT scan should be followed by a
In about two-thirds of cases the haemorrhage diagnostic lumbar puncture; if this is also
results from the rupture of an aneurysm on a negative then the diagnosis of a warning leak is
major cerebral blood vessel but, even after com- effectively excluded and the prognosis is excel-
plete angiography, no source can be identified lent. If a positive diagnosis of SAH is reached,
in about 25% of cases.14 This risk is greatest in then cerebral angiography is the next standard
the few days following the initial haemorrhage investigation to identify the source of the
but operative intervention can be technically haemorrhage.
more difficult in this period, and the patient
may be less able to withstand the added trauma
of surgery. For these reasons, there is con- Timing of Surgery
troversy about the optimal timing of surgical
interventions. The only published randomised trial of the
timing of surgery20 randomised a total of 216
patients to: acute surgery (days 0–3); inter-
Epidemiology mediate surgery (days 4–7); or late surgery (day
8 or later). Only good grade patients (Grades I to
Epidemiological reports suggest that the inci- III according to the Hunt and Hess classifica-
dence of SAH is of the order of 10 to 15 cases tion21) were recruited, but in this selected sub-
per 100 000 population per year, although more group – and in what was an under-powered trial
recent reports yield estimates in the range of six – there was some evidence to support acute
to eight cases per 100 000 per year. A recent surgery. This conclusion appears to reflect
meta-analysis15 reviews this literature and shows current practice, where there has been a drift
that the apparent fall in incidence can be towards early surgery, especially for patients in
explained by more precise diagnosis, following a good clinical condition.14
the wider availability of computed tomography Endovascular techniques as an alternative to
(CT) scanning. SAH is more common in women clipping intracranial aneurysms have evolved
(unselected series are typically 60% female). over the last 10–15 years. The aneurysm is
There is strong evidence that genetic factors occluded either by inflating a detachable bal-
predispose individuals to the risk of SAH, and loon, or by releasing coils of platinum wire to
some populations have much higher incidence induce thrombosis. The ‘state of the art’ is
rates: Finland, for example, has an incidence arguably the Guglielmi Detachable Coil device
Using Routine Data to Enhance Results of RCTs 153
(GDC) which has been available in North information about the patient’s condition on
America since 1991 and in Europe since 1992. discharge from hospital or subsequently.
Uncontrolled series have provided encouraging There were other technical problems. In the
results, especially with patients in poor clinical SMR1 data, surgical interventions are coded
condition or with difficult surgical aneurysms. according to the OPCS codes with a change
Formal evaluation of the GDC device continues, from OPCS3 to OPCS4 between 1988 to 1989;22
most notably in a MRC-funded trial of coiling codes for embolisation are only included in
versus conventional surgery (ISAT – Inter- OPCS4 so that it was not possible to analyse the
national Subarachnoid Aneurysm Trial). data for earlier periods. The form of the record
makes it difficult to reconstruct sequences of
operative events and so it was not always
SMR1 Data in the Study of SAH possible to distinguish between a definitive pro-
cedure to repair a ruptured aneurysm and the
Against this background, there was a sequence surgical treatment of a later complication.
of questions that it seemed possible to explore With these qualifications, the potential of the
using the linked SMR1 data. These were, first, SMR1 data for studying subarachnoid haemor-
whether it was possible to describe the epi- rhage appeared to be:
demiology of the condition in the whole popula- (a) Epidemiology: the linked data and death
tion and thus establish a context for its manage- records provide an opportunity for estimat-
ment in specialist centres and the outcomes they ing the incidence of SAH, and the case
achieved. Second, given the problem of mis- fatality rate; the inclusion of place of resi-
diagnosis and the sequence of events needed to dence allows the use of Census-derived
establish a definitive diagnosis prior to surgical population denominators.
intervention, was it possible to describe the (b) Admissions to neurosurgical units: the
patterns of care on which the practice of linkage of ‘episode’ data allows an account
specialised centres was actually based? Third, of the sequences of patient care – for
acknowledging the difficulties in conducting example, the numbers of patients who died
formal trials of one form of intervention or before being admitted to hospital, the num-
another, could the SMR1 data be used to bers admitted to a neurosurgical unit, and
describe patients treated by the open clipping of their prior in-patient care. Doing so can pro-
aneurysms with those treated by embolisation? vide a useful account of admission policies,
What had been the pattern of uptake of the although the absence of data on the patients’
newer embolisation treatment? Finally, provid- clinical state limits this capability.
ing that satisfactory answers could be found for (c) Timing of surgery: the data report the num-
these questions, was it possible to devise a bers of patients undergoing clipping of an
pseudo-trial of early or late interventions? aneurysm, and the timing of surgery can be
It quickly became evident that the SMR1 data related to the date of first hospital admis-
presented a number of difficulties in addressing sion. Linking this information to the date
these questions. One was that of identifying of death allows the construction of best-
‘true’ cases of SAH amongst multiple patient and worst-case scenarios for early surgery
transfers either between hospitals or between versus late surgery.
specialities within a hospital. It was not uncom- (d) Other treatments: the potential of SMR1
mon, for example, to have an initial diagnosis of data to investigate other treatments is
meningitis which changed to SAH in a sub- limited because, for example, information
sequent episode record, or an initial diagnosis of about drugs is not included. The data could
SAH meriting transfer to a neurosurgical unit have uses, for example, for exploring
where the diagnosis of SAH is subsequently changes in case fatality rates following the
refuted. What were the criteria for accepting publication (in the late 1980s) of evidence
a diagnosis of SAH, and at what stage in a about the value of nimodipine in reducing
patient’s history should they be applied? This cerebral infarction.
question was complicated further by the lack of The case study we describe focussed on the uses
a record of the onset of symptoms or of the of these data as a means of assessing the optimal
patient’s condition on admission to hospital or timing of surgical interventions.
later deterioration. This meant that it was not
possible to discern what had been the intended
management of the patient: was early surgery Methods
planned or would this be the consequence of a
further haemorrhage? Although death was avail- For the reasons we describe above, subgroups
able as an outcome measure, there was no other reflecting different degrees of confidence in the
154 The Advanced Handbook of Methods in Evidence Based Healthcare
diagnosis of SAH were extracted from the com- needing review if trial populations are to be
plete SMR1 data set. The main analysis was optimised.)
restricted to records from 1989–95, when sur-
gical interventions were coded according to
OPCS4. The data were for the whole of Scot- The Timing of Surgery
land, with no attempt to compare the different
neurosurgical units involved. Patients who were treated within a neurosurgical
The first sample (Sample I) comprised all unit for presumed SAH (Sample IV) comprised
cases of suspected SAH, defined as all cases 2613 individuals, of whom 1628 (62.3%) under-
with a primary diagnosis of SAH not otherwise went surgery (when this was defined as clipping
specified (ICD9 430.9) in an SMR1 record or as of the aneurysm). Figure 1 shows the distribu-
the principal cause in a death record. A total of tion of the timing of surgery, in days from first
4903 cases were identified, comprising 1796 hospital admission, and is virtually the same as
males (37%) and 3107 females (63%); this total the corresponding results from the International
provides an incidence rate of 13.7 per 100 000 Co-operative Study on the Timing of Aneurysm
population per year. Surgery,23 although in the latter study the timing
Sample I consisted of three disjoint sub- was from onset of symptoms of SAH rather than
samples: (i) individuals with a death record but from admission to hospital.
no record of hospital admission for SAH; (ii) Table 2 reports the numbers of patients under-
individuals admitted to hospital but not to a going clipping in each of three age groups (,45
neurosurgical unit during the relevant continu- years, 45–64 years, and $65 years): about half
ous in-patient stay (CIS); and (iii) patients who of those patients aged 64 or less underwent
were admitted to a neurosurgical unit during the definitive procedures, but this proportion was
relevant CIS. We then excluded two subgroups only 32% for those at older ages. Figure 2 shows
of patients where it might be argued that the the timing of surgery for these three groups. It is
diagnosis of SAH had not been established, to clear that the tendency towards earlier surgical
obtain Sample II which can be considered as intervention decreases with increasing age;
cases of presumed SAH. For patients who were Figure 2c indicates that 15% of those aged 65
hospitalised but not admitted to a neurosurgical and over who underwent surgery were not oper-
unit, we excluded cases where the final SMR1 ated on within two weeks of admission.
record did not mention SAH as any of the One-fifth of this sample died from any cause
recorded diagnoses. When patients had been within 90 days of their first admission. Table 3
admitted to a neurosurgical unit, we excluded details 90-day survival for the three age groups,
cases where there was no primary diagnosis of when again mortality was higher (32%) in the
SAH for the records relating to the period of oldest group. There were only 61 deaths (5%) in
neurosurgical care. These two rules excluded 51 the 1252 operated cases compared with 441
male and 478 female cases, leaving Sample II deaths (32%) amongst the 1361 patients who did
with 4374 cases of presumed SAH. This total not receive clipping. Table 4 shows the survival
provides an annual incidence of 12.3 cases per time following hospital admission for the two
100 000 population. groups when 33% of cases with no operation
Sample III comprised all individuals in compared to only 11% of those who did have
Sample I who were admitted to a neurosurgical operations died within three days of admission;
unit – that is, patients with suspected SAH 30% of deaths in the operated cases and 28% in
admitted to neurosurgery. This sample was fur- the group who did not have operations occurred
ther reduced to 3091 individuals comprising between 15 and 90 days after the first admission
1163 males (38%) and 1928 females (62%). to a neurosurgical unit.
This sample provided a means of exploring the Table 5 shows the timing of the 61 operated
admission policies of the different neurosurgical deaths broken down according to the timing of
units. surgery. Collectively, the tables allow one to
The final Sample IV included individuals perform various ‘what if . . .?’ calculations.
from Sample II who were admitted to a neuro- Taking a very pessimistic view of early surgery,
surgical unit – that is, patients with a presumed we could assume that all deaths following sur-
SAH who were treated in a neurosurgical unit. gery performed within three days of admission
This sample comprised 2613 individuals (949 could have been prevented if surgery had been
males and 1664 females) and is the relevant delayed. This would have prevented 45 deaths
sample for exploring the potential for trials of (28 following surgery performed on days 0–1
the timing of surgery or of embolisation. (One and a further 17 following surgery performed on
might, however, take a wider view and regard days 2–3), reducing the overall mortality rate for
the admission policies of neurosurgical units as the whole sample of 2613 patients from 19.2%
Using Routine Data to Enhance Results of RCTs 155
to 17.5%. In contrast, by taking a very opti- achieve 90% power to detect a reduction in
mistic view of early surgery, we could assume mortality from 19.2% to 17.5% at the 5% sig-
that all deaths in unoperated patients occurring nificance level. The SMR1 data could be used to
from day 2 to day 10 could have been prevented refine these calculations – for example, by im-
by early surgery, and that deaths in patients posing an upper age limit on the trial population
operated on after day 3 could have been pre- – but they would not permit more precise
vented by early surgery. This would have pre- inclusion/exclusion criteria based on the patient’s
vented 190 deaths – 174 unoperated (Table 4) clinical state. These would be important con-
and 16 deaths among patients operated upon siderations in the design of an actual trial.
after day 3 (Table 5) – reducing the overall
mortality rate from 19.2% to 11.9%.
These are extreme scenarios; in practical CASE STUDY 2: CORONARY
terms, they mean that a comparison between
early and late surgery in an unselected neuro-
REVASCULARISATION
surgical population would require a trial with
sufficient power to detect a reduction in mortal- Introduction
ity of the order of 1.7%. This implies a trial
in which an unfeasibly large total sample size Coronary artery bypass grafting (CABG) was
of about 21 800 would be needed in order to first reported in 196824 as a technique for coron-
ary revascularisation in patients with sympto-
matic coronary artery disease. There is strong
Table 2 Number and proportion of patients evidence that the technique results in prolonged
undergoing surgery, according to age group survival and improved quality of life in specific
patient subgroups when the procedure is com-
Age group Number of Number (proportion) pared to medical therapies.25 The rate of coronary
(years) patients undergoing surgery artery surgery in the United Kingdom has risen
,45 902 493 (54.7%) steadily from 212 per million in 1986 to 341 per
45–64 1273 619 (48.6%) million in 1993–94.26
$65 438 140 (32.0%) Percutaneous transluminal coronary angio-
plasty (PTCA) was described in 197927 as an
156 The Advanced Handbook of Methods in Evidence Based Healthcare
Table 3 Number and proportion of patients dying within 90 days of their first
admission to hospital, according to age group
Age group (years) Number of patients Number (proportion) dying within 90 days of first admission
Table 4 Timing of deaths within the first 90 days following first admission to
hospital, according to operational status
Days to death 0–1 2–3 4–7 8–10 11–14 15–28 29–90 Total
Operated cases 2 (3%) 5 (8%) 13 (21%) 15 (25%) 8 (13%) 4 (7%) 14 (23%) 61 (100%)
Unoperated cases 100 (23%) 47 (11%) 62 (14%) 65 (15%) 44 (10%) 66 (15%) 57 (13%) 441 (100%)
Table 5 Timing of deaths within the first 90 days following admission to first
hospital for operated cases, according to timing of surgery
Timing of surgery Days to death (days from first admission)
(days from first
admission) 0–1 2–3 4–7 8–10 11–14 15–28 29–90 Total
0–1 2 5 7 5 4 1 4 28
2–3 – 0 5 6 2 2 2 17
4–7 – – 1 4 1 0 2 8
8–10 – – – 0 0 0 3 3
11–14 – – – – 1 0 2 3
151 – – – – – 1 1 2
All operated cases 2 (3%) 5 (8%) 13 (21%) 15 (25%) 8 (13%) 4 (7%) 14 (23%) 61 (100%)
take place worldwide in 1998. These will average follow-up of 5.4 years.31 The 5-year
comprise between 60% and 90% of all PTCA survival rate was 89.3% for the CABG group
procedures.28 and 86.3% for the PTCA group (P 5 0.19, 95%
Clearly, there are fundamental differences CI for difference 20.2% to 6.0%). Again, there
between the two techniques. With PTCA there is is no strong evidence of a difference in out-
a risk of further stenosis as a result of elastic comes, although the limited number of deaths is
recoil, or from the progression of atheroscler- reflected in a wide confidence interval which
osis. PTCA does, however, retain future options does not exclude the possibility of a clinically
for revascularisation. CABG places venous con- relevant difference. A post hoc subgroup analy-
duits which are largely immune from advancing sis suggested that CABG is to be preferred to
disease,29 although they do tend to deteriorate PTCA in patients with treated diabetes.
over time. Repeating the bypass surgery can be
compromised if the ‘prime’ veins have been
used in an initial procedure. Limitations of the Trial Data
In a commentary accompanying the Pocock
The Evidence from Trials meta-analysis, White pointed out that the trials
of CABG versus PTCA leave many questions
There are several well-defined subgroups of unanswered.32 There is, firstly, the problem of
patients where one or other procedure is clearly generalising these findings: the trials have only
indicated, but there is a large ‘grey area’ where randomised a minority of the patients who were
either procedure might be thought appropriate. eligible for revascularisation. Secondly, even
Several clinical trials comparing CABG and when taken together in a meta-analysis, trials up
PTCA as the first intervention in patients with to the present have been under-powered, and
symptomatic coronary artery disease have been their results do not exclude what may be clinic-
conducted or are on-going. In 1995, a meta- ally relevant differences in mortality and or later
analysis summarised the results of eight trials in myocardial infarction. Thirdly, there is the ques-
which a total of 3371 patients (1661 CABG and tion of the most appropriate outcome measures:
1710 PTCA) had been followed for a mean of one could, for example, consider PTCA as a
2.7 years.30 The number of deaths was 73 in the ‘package’ or a holding procedure, accepting as a
CABG group and 79 in the PTCA group, result- ‘cost’ the possibility of further intervention in
ing in a relative risk (RR) of 1.08 with a 95% return for the perceived benefit of delaying a
confidence interval (CI) of 0.79 to 1.50. This CABG procedure that it might not be possible to
number of deaths is too small for this result to repeat. For this argument, follow-up periods of
establish the equivalence of the two mortality the trials have been too short, because CABG
rates: a more sensitive comparison of the two procedures would not be expected to fail for
procedures employed the composite end-point about six years. Fourth, there is the question of
of cardiac death or non-fatal myocardial infarc- subgroup effects. Various trials have suggested
tion (MI). The results of the eight trials were that there are such effects, but none has had
relatively consistent, with a total of 127 end- sufficient power to investigate this issue. This is
points in the CABG group during the first year a crucial difficulty because the clinical question
compared to 135 end-points in the PTCA group is not so much a simple comparison of CABG
(RR 1.03, 95% CI 0.84–1.27). Over the entire with PTCA, but rather one of narrowing the
follow-up period there were 154 end-points in indeterminate area between them by improving
the CABG group and 169 in the PTCA group the identification of patients where one or other
(RR 1.10, 95% CI 0.89–1.37). In contrast to procedure is indicated. Finally, the trials report
these findings, there was evidence that CABG – outcomes that may have been overtaken by
relative to PTCA – gave better control of angina major advances in the procedures themselves.
and required less frequent subsequent interven- Anaesthetic and surgical techniques have im-
tion. A subgroup analysis based on three of the proved, and the use of stents in conjunction with
eight trials suggested that CABG was to be PTCA is becoming widespread.28 How relevant
preferred to PTCA in patients with single-vessel are comparisons of yesterday’s technology for
disease in terms of the end-point of cardiac today’s practice?
death or non-fatal MI. This is, perhaps, a Some of these problems are inherent in any
counter-intuitive finding to be interpreted with approach to health technology assessment. By
caution given the very small number of events definition, evaluation must lag behind the devel-
on which the subgroup analysis was based. opment of new technologies – in particular, and
These findings are limited by the short for questions of this kind, it takes time to
follow-up period, and so it is helpful to consider accumulate adequate long-term follow-up data.
the results of the BARI trial which had an Routine data do, however, have the potential to
Using Routine Data to Enhance Results of RCTs 159
address some of these problems. They provide a around the world each year, and so the poverty
powerful means of assessing the relationship of of trial data is an interesting commentary on the
trial populations to the ‘real world’ from which perceived importance of trials as a support for
they are drawn, and they have the potential for major invasive procedures when compared to
assembling much larger numbers of cases than those expected for pharmaceutical innovations.
can be achieved in trials. Routine data could be It is, of course, a reflection of differing regu-
used to augment a meta-analysis of trial data, latory standards and requirements.
and could be used to explore subgroup effects. Routine data do provide the potential to study
very large numbers of patients, with prolonged
SMR1 Data in the Study of Coronary follow-up. Unlike trial data, they also provide
the opportunity to study epidemiological trends
Revascularisation: Limitations in the uptake of different procedures, and to
characterise their use in the entire patient popu-
In practice, several problems limited the useful-
lation. In this case study we describe these
ness of the linked SMR1 data for analysing
epidemiological aspects of their use briefly, but
the outcomes of coronary artery revascularisa-
chiefly focus on trying to combine information
tion. The first of these was that there was no
from the routine data with the trial data in order
definitive way to identify patients who were
to make formal comparisons of outcome follow-
undergoing their first coronary revascularisation
ing the two procedures.
outwith the data themselves. The linked data had
to be used to identify records of previous pro-
cedures; doing so was further complicated by Methods
the fact that an operative code for PTCA was
not included in the OPCS3 codes (used before The initial sample comprised all patients with an
1989) so that patients who had this procedure SMR1 record of CABG or PTCA between 1989
before this date could not be confidently identi- and 1995. These records were then divided
fied. The records did not include a description between patients who had procedures for single-
of the patients’ clinical state (specifically, the or multi-vessel disease on the basis of the
severity of angina). This information would OPCS4 operative codes. A subset of this sample
have been invaluable as a way of stratifying was then generated to reflect as closely as
case-mix, and in interpreting outcome measures. possible the entry criteria for the major trials.
There was no record of the extent of the under- Specifically, patients were excluded if there was
lying pathology – particularly, whether patients a record of an earlier CABG (1981 onwards) or
had single or multiple-vessel disease, used in PTCA (1989 onwards), or if any of the follow-
many trials as an entry criterion or as a stratify- ing diagnoses were recorded as being present at
ing factor. It was possible to approximate to this the time of surgery: diseases of mitral valve
pathological classification by using procedure (ICD9 394); diseases of aortic valve (ICD9
codes which permitted a distinction between 395); diseases of mitral and aortic valves (ICD9
single- or multiple-vessel interventions, whether 396); acute myocardial infarction (ICD9 410);
for CABG or PTCA. Doing so introduced a other diseases of endocardium (ICD9 424);
degree of confounding because it was possible cardiac dysrhythmias (ICD9 427); heart failure
for patients with multi-vessel disease to have a (ICD9 428); ill-defined descriptions and compli-
first procedure performed on only one of the cations of heart disease (ICD9 429). Within the
affected vessels. limits of the SMR1 data, these were patients
undergoing their first coronary revascularisation.
This subset became the ‘Pseudo-RCT Sample’
SMR1 Data in the Study of Coronary below. It included 12 238 individuals, of whom
Revascularisation: Their Potential 8524 (70%) underwent CABG. Those failing to
meet these criteria are described as the ‘Excluded
The two main limitations with trial data in this Sample’ and comprised 2359 individuals, of
area are, first, that – even with meta-analysis – whom 1825 (77%) underwent CABG.
patient numbers are insufficient for detecting
modest but relevant differences in mortality
between the two procedures, and are inadequate Epidemiology: Coronary
for exploring subgroup effects. The length of Revascularisation in Scotland between
follow-up is unlikely to detect late failures of 1989 and 1995
CABG. A further problem is that trial subjects
are highly selected and likely to be unrep- Figure 3 plots the number of CABG and PTCA
resentative of the vast majority of patients who procedures undertaken in Scotland from 1989 to
undergo coronary revascularisation. Hundreds of 1995 for patients in the ‘Pseudo-RCT’ sample.
thousands of these operations are conducted At 33 per 100 000, the rate of CABG operations
160 The Advanced Handbook of Methods in Evidence Based Healthcare
was considerably greater in 1995 than the 9 per record of MI in the previous four weeks and
100 000 in 1989; the rate for PTCA was 14 patients with angina as a recorded co-morbidity
per 100 000 in 1995 – an increase from 7 per were more likely to receive PTCA. This pat-
100 000 in 1989. In absolute numbers, CABG tern was clearly evident in the ‘Pseudo-RCT’
was substantially greater than PTCA, and until sample, and even more apparent in the ‘Ex-
1994 the rate of increase was also greater for cluded’ sample.
CABG. This increase was mostly attributable to
the largest of the three main Scottish centres
performing these operations. Comparing RCT Subjects with
An initial analysis, based on logistic regres- ‘Routine’ CABG/PTCA Patients
sion, investigated case-mix differences between
patients undergoing the two procedures. As Table 6 compares the baseline characteristics of
expected, there was strong evidence of selection the two samples derived from the SMR1 data
bias: female patients, patients with multi-vessel with a major trial published after the Pocock
disease and patients with hypertension were meta-analysis,31 and the three largest trials
more likely to receive CABG. Patients with a from the meta-analysis.33–35 The ‘Pseudo-RCT’
Table 7 Event rates for death and/or myocardial infarction within one year
Pseudo-RCT Excluded CABRI RITA EAST GABI Toulouse MASS Lausanne ERACI
Total number
of events 377 288 72 65 57 28 12 6 8 15
patients are broadly similar to the trial popula- ‘Pseudo-RCT’ and ‘Excluded’ samples. It is
tions, although a major difference between the clear that PTCA patients are far more likely to
trials is whether or not they focussed on multi- require reintervention than CABG patients. The
vessel disease. The ‘Excluded’ sample had a BARI and EAST trials reported similar analy-
higher proportion of females, but was otherwise ses, and the similarity in the shape of the sur-
broadly similar to the other populations. vival curves from the two sources is striking.
For the PTCA patients, the rate of reintervention
Cardiac Death and/or Myocardial was higher for the ‘Pseudo-RCT’ sample than
Infarction following CABG/PTCA for the ‘Excluded’ sample.
Revascularisation following
CABG/PTCA
Figure 4, derived from the Cox regression Figure 4 Survival free from revascularisation
model, shows survival free from revascularisa- in CABG and PTCA groups for (a) the ‘Pseudo
tion in the CABG and PTCA groups, for the RCT’ sample and (b) the ‘Excluded’ sample.
162 The Advanced Handbook of Methods in Evidence Based Healthcare
Table 8 summarises mortality at one year for Continuing research on this topic is concerned
the ‘Pseudo-RCT’ and ‘Excluded’ groups, and with ways of combining the evidence on treat-
Table 9 gives the corresponding results for ment differences from the routine data set with
deaths within 5 years. The 5-year results are the evidence provided by the meta-analysis.
based on relatively small numbers of cases from Two Bayesian approaches are being developed.
early in the study period. For the ‘Pseudo-RCT’ 1 To derive a posterior predictive distribution
patients, both unadjusted and adjusted one-year for the treatment difference based on the
outcomes show that the deaths among the meta-analysis data. Compare this with the
CABG patients were significantly more frequent posterior (using a vague prior) distribution
than those in the PTCA patients. This finding for the treatment difference based on the
contrasts with the results from the randomised routine data in order to examine possible
trials where, on the basis of substantially smaller discrepancies.
numbers, no significant difference in mortality 2 To derive a posterior distribution from the
has been reported. The 1-year mortality rate of routine data and use this as a prior for the
3.2% for CABG patients is comparable to the meta-analysis. Doing so will allow invest-
rate of 2.7% reported in the CABRI trial. This igation of the extent to which beliefs based
trial reported a rate of 3.9% for the PTCA solely on the routine data change in the light
of evidence from the RCTs.
patients, however, and this is more than twice
the rate observed for the ‘Pseudo-RCT’
patients. CONCLUSIONS AND RECOMMENDATIONS
In the ‘Pseudo-RCT’ sample, neither the un-
adjusted nor the adjusted results provide evi- The Uses of Routine Data
dence of a difference in mortality at 5 years
between the CABG and PTCA groups. This is The case studies on which this chapter is based
consistent with the results from the randomised were undertaken to test the proposition that
trials, but there is a similar pattern to that routinely assembled NHS data can either sub-
observed at 1 year: the ‘Pseudo-RCT’ mortality stitute for RCTs in situations where they are not
rate for CABG was similar to that observed in feasible or can supplement the information that
the randomised trials, but the rate for PTCA (at RCTs provide in assessing the benefits ensuing
7.2%) was lower than that reported for the from new technologies or interventions. There
BARI and EAST trials. are three main ways in which routine data can
The results for the ‘Excluded’ patients show provide this supplementary function:
the expected high mortality rates in the CABG 1 In situations where formal RCT designs are
group. The low 5-year mortality rate for the inappropriate – as in the SAH study.
‘Excluded’ PTCA patients possibly reflects very 2 For evaluating outcomes over longer periods
careful selection of patients with complicated or in a wider context than is practical even
coronary disease in the early years of developing with meta-analyses – as in the CABG
this procedure. assessment.
Using Routine Data to Enhance Results of RCTs 163
3 By providing a link between the methodo- posed, given the gain in death rates suggested by
logically constrained conclusions of RCTs the analysis of the routine data.
and their later adoption in health services Two different kinds of question were posed
practice. for the second case study: firstly, were the
outcomes promised by earlier trials achieved
The investigation of surgical interventions for when these interventions were adopted in more
SAH illustrates both the strengths and weak- routine practice; and secondly – against this
nesses of routine data as an alternative to RCTs: background – was it possible to evaluate the
the major problem of the analysis we describe relative uses and benefits of CABG and PTCA
was simply the adequacy and precision of the as complementary, rather than alternative, inter-
data for the purposes of the enquiry. This is not ventions? Both questions depended on earlier
a very surprising conclusion given that routine RCT findings, but both required longer follow-
data are, at best, summaries of case records up and larger numbers than RCTs (even as
assembled for other (usually administrative) meta-analyses) were likely to provide. As with
purposes. As we are able to demonstrate, how- SAH, initial analyses of the SMR1 records
ever, the data do have value as a means of identified the lack of diagnostic precision neces-
describing the epidemiological features of the sary for formal RCT designs as a problem, but
condition, the patterns of surgical intervention the study also illustrates the diffusion of the
relating to it, and rates of survival. The lack of procedure to other patients: 77% of the patients
additional data made it difficult to go beyond excluded from the ‘Pseudo-RCT’ we devised
this stage (for example, in assessing outcomes also underwent CABG.
other than death). In other ways, the data did not This said, and with the exception of data
permit treatment comparisons that might have about multi-vessel disease, it was possible to
made it possible to address the question of the devise a sample of patients who were broadly
benefits of early or late surgery. This is not to similar to trial patients and to compare their
say, however, that the study was without value outcomes with the results of RCTs – notably the
in understanding the context of the original meta-analysis of Pocock et al. Its importance is
question. The need to develop case-definitions not (or not only) that it provides information that
through different stages is illustrative of the can be set alongside earlier RCT findings when
ways in which RCT populations may not be these procedures are undertaken in the less pre-
typical of the complete diagnostic entity, and cisely defined context of surgical practice. It is
thus of the extent to which RCT findings can be also an answer to the question of whether the
generalised to expectations for the healthcare potential for ‘health gain’ implicit in RCT
system as a whole. Whilst it may appear at first results finds realisation in the wider healthcare
sight that there were 4903 instances of SAH, system.
based upon the principal cause of death or the
recorded primary diagnosis in hospital, further
evidence available from routine data suggests
The Limitations of Routine Data
that this may be an over-count. In 11% of such
The link between the benefits to clinical practice
cases the diagnosis was altered in neurosurgery that accrue from RCTs and the gain to the health
or prior to discharge from hospital if the patient of the population that might then be expected is
was never admitted to neurosurgery, so only important for considering the future content and
89% of all the ‘suspected’ cases of SAH could uses of routinely assembled data. It forms part
be considered ‘presumed’ SAH. If the interest is of the rationale for evidence based practice
in surgery, as in this case, the numbers are because routine data will continue to provide the
further limited since only 63% of all ‘suspected’ basis for the epidemiological monitoring of mor-
cases of SAH are admitted to neurosurgery, and bidity and its outcomes. Present data have a
just 53% of all ‘suspected’ cases may be con- number of deficiencies for these purposes. The
sidered ‘presumed’ SAH and are admitted to most significant is the lack of diagnostic pre-
neurosurgery. Such a comment has a bearing cision in regard to both their content (considered
on other possible uses of RCT findings – for as diagnostic-mix) and the slow response of ICD
example, as baseline standards for audit studies The difficulty of sequencing of events and the
or as anticipated measures of hospital perform- lack of information about medication and non-
ance. Other uses of routine data are illustrated surgical interventions are other constraints as is
by the ‘what if . . .’ speculation we include as an the limited reporting of a patient’s condition at
indication of the trial conditions necessary to the end of an episode of care. Case-severity and
answer certain kinds of question. In this example, case-complexity are perhaps the most significant
an almost impossibly large trial would be neces- deficiency, even though scoring systems such as
sary in order to answer the question we first POSSUM36 and APACHE III37 are increasingly
164 The Advanced Handbook of Methods in Evidence Based Healthcare
22. Office of Population Censuses and Surveys. Clas- randomised trials comparing coronary angioplasty
sification of surgical operations and procedures. with bypass surgery. Lancet 1995; 346: 1184–9.
Fourth revision. London: OPCS, 1987. 31. BARI Investigators. Comparison of coronary
23. Kassell NF, Torner JC, Haley EC, et al. The bypass surgery with angioplasty in patients with
international co-operative study on the timing of multivessel disease. N. Eng. J. Med. 1996; 335:
aneurysm surgery. Part 1: Overall management 217–25.
results. J. Neurosurg. 1990; 73: 18–36. 32. White HD. Angioplasty versus bypass surgery.
24. Favaloro RG. Saphenous vein autograft replace- Lancet 1995; 346: 1174–5.
ment of severe segmental coronary artery occlu- 33. CABRI Trial Participants. First year results of
sion. Operative technique. Ann. Thorac. Surg. CABRI (Coronary Angioplasty versus Bypass
1968; 5: 334–9. Revascularization Investigation). Lancet 1995;
25. Killip T, Ravi K. Coronary Bypass Surgery. In: 346: 1179–83.
Pitt B, Julian D, Pocock S (eds). Clinical Trials in 34. RITA Trial Participants. Coronary angioplasty
Cardiology. Saunders, London, 1997, pp. 173–
versus coronary artery bypass surgery: the Ran-
84.
domised Intervention of Angina (RITA) trial.
26. Black N, Langham S, Coshall C, Parker J. Impact
Lancet 1993; 341: 573–80.
of the 1991 NHS reforms on the availability and
35. King SB III, Lembo NJ, Weintraub WS, et al. A
use of coronary revascularisation in the UK
randomised trial comparison of coronary angio-
(1987–1995). Heart 1996; 76(Suppl. 4): 1–30.
27. Grüntzig AR, Senning A, Siegenhaler WE. Non- plasty with coronary bypass surgery. N. Eng. J.
operative dilatation of coronary-artery stenosis. Med. 1994; 331: 1050–5.
N. Eng. J. Med. 1979; 301: 61–8. 36. Copeland GP, Jones D, Walters M. POSSUM: a
28. EPISTENT Investigators. Randomised placebo- scoring system for surgical audit. Br. J. Surg.
controlled and balloon-angioplasty-controlled trial 1991; 78: 355–60.
to assess safety of coronary stenting with use of 37. Knaus WA, Wagner DP, Draper EA, et al. The
platelet glycoprotein-IIb/IIIa blockade. Lancet APACHE III prognostic system: risk prediction
1998; 352: 87–92. of hospital mortality for critically ill hospitalized
29. Knudtson ML. Commentary. Evidence-Based adults. Chest 1991; 100: 1619–36.
Medicine 1996; 1: 83–4. 38. World Health Organization. Manual of the
30. Pocock SJ, Henderson RA, Rickards AF, international statistical classification of diseases,
Hampton JR, King SB III, Hamm CW, Puel J, injuries and causes of death. Ninth revision.
Hueb W, Goy J, Rodriguez A. Meta-analysis of Geneva: WHO, 1977.
10
Qualitative Methods in Health Technology
Assessment
E L I Z A B E T H M U R P H Y and R O B E R T D I N G W A L L
quantitative method. This applies particularly or dismissed en bloc. This obscures the con-
when examining the delivery and organisational tested nature of qualitative research, and the
contexts of new technologies – how can a extent to which different versions exist which
national roll-out achieve the same results as the may be more or less appropriate to the goals of
demonstration projects? – or some of the inter- HTA. There is little agreement among practi-
personal dimensions of the technology. New tioners of qualitative research about what it is,
drugs depend for their effectiveness on patient what it is for, how it should be done or how it
concordance with specific regimens which, in should be evaluated. Given this dissensus, it is
turn, seems to be affected by the interactional meaningless to argue for or against qualitative
environment within which prescription and dis- research in general. Whether for the prosecution
pensing occur and the ways in which infor- or the defence, the case needs to be made in
mation and advice about appropriate use are relation to particular versions in particular con-
delivered. Some new technologies, of course, texts for particular purposes.
are non-material – the rise of counselling in At the most fundamental level, the methodo-
primary care would be a good example. Even a logical literature on qualitative research reflects
material technology may have some dimensions two conflicting sets of basic philosophical
that imply qualitative research. How can con- assumptions. Some writers1,2,3 argue that qual-
sistent reading of DNA maps be achieved? It itative and quantitative research methodologies
seems to be difficult enough with the established derive from radically incompatible paradigms or
technology of cervical smears! world-views.4 The proponents of this ‘two para-
However, there is also considerable disagree- digms’ approach hold that the differences are so
ment amongst users of qualitative methods about pervasive and ineluctable5,6 that it is impossible
their nature, status and employment. This chap- to ‘mix and match’ between them.
ter attempts to demonstrate this diversity of The separatist position is consistently argued
view, while indicating which positions can make by Smith, for example.1,2,7–10 The fundamental
particular contributions to HTA. For this reason philosophical divide between qualitative and
it is not based on a conventional systematic quantitative research reflects the former’s
review, which was both materially and intellec- grounding in Idealism and the latter’s in Scien-
tually impracticable in this context. The lit- tific Realism. (The opposition between these
erature to be searched was too large, and the two positions has a long history in philosophy,
search engines inadequate for the task. More- which will not be reviewed here.) Smith8 sums
over, in the absence of an agreement about how up Realism, which he sees as antithetical to
to resolve disagreements about the methodo- qualitative research, as ‘the idea that reality
logical foundations of qualitative research, it is exists independent of us. Independent means
impossible to produce an incontestable ‘Best that this reality exists whether or not we are
Buy’. The most that one can achieve is to review aware of it or take any interest in it.’ (p. 8).
the range of options and to assess their fit- Idealism, on the other hand, is the view that
ness for the purposes of HTA. The test of this what we take to be the external world is merely
chapter is the representativeness of the range of a creation of our minds.11 Idealists do not neces-
positions reviewed rather than the exhaustive- sarily deny the possibility that a material world
ness of the reading on which it is based. Never- exists, they hold that it is impossible to ‘know’
theless, if the objectives of HTA are taken as such a world. While Realists pursue what Smith9
given, it is possible to make some fairly firm has called a ‘God’s eye point of view’ (p. 381),
assessments of the value of particular positions or one single ‘true’ reality, Idealists hold that
for this programme. there are multiple realities or ‘as many realities
This chapter does not provide a ‘how to do it’ as there are persons’9 (p. 386).
guide. There are numerous excellent textbooks The Constructivism of some qualitative meth-
and courses available for that.a However, it does odologists leads to similar conclusions. Guba
attempt to explain why qualitative research has a and Lincoln,12 for example, argue that ‘there
place in the HTA armamentarium, how to iden- exist multiple, socially constructed realities,
tify when it is appropriate, how to evaluate it ungoverned by laws, natural or otherwise . . .
and how to think sensibly about it. these constructions are devised by individuals as
they attempt to make sense of their experiences
. . . constructions can be and usually are shared
. . . this does not make them more real but simply
KEY ISSUES AND CONTROVERSIES more commonly assented to’ (p. 86, original
emphasis). The quotation highlights the prob-
Debates in HTA, as in other fields, about the use lem of this position for HTA research. HTA is
of qualitative research often homogenise this committed to pursuing truth as a secure basis for
intellectual tradition as if it were to be adopted action. Idealists and Radical Constructivists hold
168 The Advanced Handbook of Methods in Evidence Based Healthcare
that multiple, even contradictory, true versions than so-called ‘objective reality’ in determining
of reality can exist simultaneously. There is no how they respond.20 People act on the basis of
ultimate basis on which we can choose between what they believe to be true21 rather than what
them. If this is the case, and researchers simply might be ‘objectively true’. This has important
produce yet another version that has no better implications for the successful implementation of
claim to truth, then one wonders about the technologies in health, as in other fields, in that
justification for spending public money on their advocates for change will need to engage the
work.13 Radical relativism has deeply negative beliefs of intended users and beneficiaries rather
consequences for the value of research findings than relying purely on rationalist justifications.
in practice and/or policy making.14 It leads either A number of methodologists have moved from
to a ‘debilitating nihilism’15 (p. 252) or the this uncontroversial assertion of the importance
chaos of ‘anything goes’. of interpretation in human action and interaction
In rejecting Idealism and Radical Constructiv- to argue that the central role of qualitative
ism as a basis for HTA research, it should not be research should be to uncover the meanings
thought that the pursuit of ‘truth’ in any research behind such action.22–31 This is sometimes
(qualitative or quantitative) is unproblematic. presented as a defining difference between qual-
There are profound difficulties with the concep- itative and quantitative research. Henwood and
tion of truth in conventional (or as Hammersley13 Pidgeon29 argue, for example, that qualitative
terms it ‘naı̈ve’) Realism. Researchers cannot researchers seek to avoid quantitative research’s
escape the social world to make their observa- tendency to ‘fix meanings’ without reference to
tions.16 A researcher’s observation is always and the meanings employed by the participants in
inevitably shaped by the ‘biological-cultural the context being studied. The understanding of
lenses through which it is seen’17 (p. 157). As a participants’ meanings is associated with a pref-
result we can never know, in an absolute sense, erence for methods that allow the researcher to
that our research findings are true. Likewise, any ‘get close’ to those being studied. These include
given reality can be represented from a range of prolonged periods of observation,32 which allow
different perspectives and it is futile to search researchers to put themselves ‘in the shoes’ of
for ‘a body of data uncontaminated by the their subjects, and/or in-depth interview tech-
researcher’16 (p. 16). It is possible to have mul- niques, which are thought to encourage people
tiple descriptions or explanations of the same to express their understandings and perspectives
behaviour or event that are all valid, as long as in their own words according to their own
they are compatible with each other. What is not priorities.
possible is to have multiple contradictory ver- The identification of qualitative research with
sions that are all true. As Phillips18 suggests, we uncovering participants’ meanings is flawed in a
can combine an acceptance that multiple per- number of respects. First, it is inaccurate to
spectives are possible without abandoning truth suggest that a focus on participant meanings
as a ‘regulatory ideal’. defines qualitative research.33–35 Much of this
This position (sometimes known as ‘Subtle type of work has been less concerned with
Realism’) is equally appropriate for qualitative participants’ perspectives than with their prac-
and quantitative research. It is an alternative tices.35 It focuses not on what people think, but
to the polarising rhetoric of some qualitative upon what they do. These researchers are more
researchers which is neither an accurate analysis concerned with studying the function rather than
of the current state of qualitative and quantita- the meaning of such practices.35 For example,
tive research, nor a necessary or helpful distinc- Silverman et al. studied the organisation and
tion between these two traditions. It also allows reception of advice giving in HIV counselling
that qualitative and quantitative research meth- sessions in ten medical centres. The analysis
ods may be used in combination. was based on audio-tapes of sessions.36 Silver-
A second controversy within qualitative man et al. found that such advice giving was
research, which is of fundamental importance in typically truncated and non-personalised –
relation to HTA, concerns the frequent assertion features which may be expected to be related to
that a defining characteristic of qualitative low patient uptake of advice. They considered
research is its commitment to uncovering the the functions that such demonstrably ineffective
meanings that underpin human action. Health practices might serve within the context in
technologies are applied by people (be they which they were performed. They showed how
doctors, nurses or technicians) to people (usually such truncated and non-personalised advice giv-
patients). One of the distinctive features of ing can be effective in handling the delicate
human action is that it is endowed with mean- issues which arise in relation to sexual practices.
ing.19 This has the very practical consequence Generalised advice allowed counsellors to avoid
that what people think each other’s actions mean singling out particular patients’ sexual activities
(their interpretation) may be more important for discussion. It also meant that patients were
Qualitative Methods in HTA 169
not required to discuss their own behaviour. Qualitative researchers are again divided on
Such practices also avoided the embarrassment this issue. Some appear to reject generalisability
that might have arisen if the counsellors ap- as an appropriate goal43,44 as being ‘unimport-
peared to be telling comparative strangers what ant, unachievable or both’45 (p. 201). In HTA
they should be doing in their intimate lives. research, of course, generalisability is extremely
Finally, these practices minimised the likelihood important, since research commissioners are
of any overt conflict within the consultation. normally less interested in particular settings for
Overall, Silverman et al. concluded that such their own sake than in the extent to which
truncated and non-personalised advice-giving findings from one can be extrapolated to others.
was functional, not least because it enabled If qualitative research is to be acceptable to
counsellors to limit the amount of time such practitioners and policy-makers, then its ap-
counselling sessions require. proaches to sample selection, representativeness
The second problem raised by the identifica- and generalisability must be understood.46
tion of qualitative research with the search for Some qualitative methodologists, such as
participants’ meanings is that there are serious Stake,44 argue that concern with representative-
questions about the feasibility of ‘uncovering’ ness may have a distorting effect. He suggests
such meanings.33 How are we to access these that the researcher’s object should be to under-
meanings? Homo sapiens is not known to be a stand the uniqueness of individual cases and to
telepathic species. Participants can be asked to offer the reader a vicarious experience of each.
explain the meanings that underlie their actions; Readers are invited to engage in ‘naturalistic
indeed, this is the justification for much qual- generalization’ as they relate these case studies
itative interviewing. However, this immediately to their existing knowledge of other settings. If
presents two problems. Firstly, there is the pos- the primary function of qualitative research is to
sibility that participants have knowledge and provide ‘vicarious experience’ and ‘experiential
experience which are relevant, but which they understanding’ in order to allow readers to make
are unable to articulate, leading to distortion and ‘naturalistic generalizations’ then probability
over-simplification. Secondly, even more prob- sampling does appear to be both inappropriate
lematically, asking participants to explain why and impractical. However, this position places
they behave in certain ways leads them to pro-
qualitative research closer to an exercise in
duce ‘accounts’.37–39 Their answers are not
communication than to a public science. As
straightforward reproductions of the mental
Kennedy41 has wryly remarked, if this is the
states which lay behind the actions they are
case then the researcher would need ‘the talent
intended to explain. They are versions of the
respondent’s experience specifically designed to of Tolstoi to be able to describe . . . events in the
make their behaviour seem meaningful and con- way that allow the reader to draw the appro-
textually appropriate to their interlocutor. As priate inference’ (p. 664).
such they have, as Gould and his co-workers24 There are, in fact, no fundamental reasons
have observed, an indeterminate relationship to why probabilistic sampling methods could not
the original decision to act in one way rather be adopted in qualitative research.43,47–51 Prac-
than another. tical and resource constraints, however, mean
that they are normally unlikely to be efficient
or cost-effective. Because of the intensive and
time-consuming nature of qualitative research,
QUALITATIVE RESEARCH IN HTA the ratio of settings studied to the number in the
population of interest can rarely be raised to a
A frequently expressed reservation about qual- sufficient level to allow statistical inference.52,53
itative research is its alleged failure to meet the In qualitative research the general use of non-
criterion of generalisability, often held to be the probability sampling methods is best understood
hallmark of science.40,41 Quantitative research as a trade-off between depth and breadth.
usually pursues generalisability by sample to However, the impracticality of probability
population inference, using probabilistic sam- sampling need not, and should not, lead qual-
pling methods. Much qualitative research is itative researchers to abandon concerns for gen-
carried out in a single setting or with a small eralisability. The basis for sampling decisions in
sample of informants, and does not meet the qualitative research has too often been pure
requirements for statistical inference.42 The rela- opportunism. Pragmatic considerations such as
tive absence of probability sampling generates ease of access (both geographical and organisa-
concerns that qualitative work is merely anec- tional) are, of course, important in sampling
dotal and therefore of no practical use to policy- decisions for both qualitative and quantitative
makers or practitioners. research. In some cases, particularly where the
170 The Advanced Handbook of Methods in Evidence Based Healthcare
research topic is highly sensitive or the popula- This approach can be applied at the beginning
tion of interest is mobile or inaccessible, oppor- of the research in identifying the particular set-
tunistic sampling may, of course, be the only tings, groups or individuals to be studied, or in
option. If, though, the role of research in HTA is the course of a project, as recurrent decisions are
to provide cumulative knowledge about tech- made about who, what and where to study with-
nologies and their implementation, then oppor- in the setting(s) chosen. A range of sampling
tunistic sampling will be the method of last resort techniques can be used, including the study
in anything but the most exploratory research. In of negative, critical, discrepant and deviant
all other qualitative research, samples should be cases, to explore, modify and extend existing
drawn in a systematic and principled way.45 theory.51,53,57
The research objective will determine which The relevance of theoretical sampling to
of two such systematic approaches is appro- research in HTA may be less immediately obvi-
priate in a particular study. Where researchers ous than approaches that emphasise sample
are primarily concerned with the representative- to population inference. However, it allows
ness or typicality of their findings, they will researchers to take findings that have been estab-
adopt purposive sampling strategies that max- lished in one context or field and test their
imise the empirical generalisability of their find- applicability in another. Researchers can exploit
ings. Researchers will examine people or set- and build upon existing knowledge from related,
tings which are thought to be typical in some or sometimes quite different, fields. Silverman,35
way of the aggregate to which they are seeking for example, reports how his observational study
to generalise. For example, in reviewing an of NHS and private oncology clinics was con-
innovative technology, researchers might study ceived as a test of the generalisability of
its use in a hospital department that is believed Strong’s findings in paediatric clinics to other
to be ‘typical’ in significant respects of the medical settings.58 A study of the general practi-
settings in which it could be applied. They tioner’s role in giving lifestyle advice, reported
would consciously exclude ‘state-of-the-art’ by Parry and Pill,59 took theoretical generalisa-
centres or locations known to be particularly tions developed by Bernstein in the context of
poorly resourced or staffed. In other cases, education and tested them in medical settings.
though, researchers may deliberately set out to Qualitative researchers take different views
study settings or groups that are in some way about the relative merits of these sampling
atypical. For example, they might choose to approaches, emphasising empirical generalis-
study an unusually successful programme in ability or theoretical generalisability. However,
order to identify the characteristics of settings these strategies are not incompatible. It is pos-
that facilitate achievement. In both cases, the sible, in making sampling decisions, to combine
researchers are making principled selection a concern with the empirical generalisability of
decisions that define the generalisability and findings with a commitment to sampling in ways
utility of their findings. This is not, of course, which are theoretically informed and which seek
entirely unproblematic: the validity of purposive to test, modify and extend existing knowledge.
sampling is dependent upon the researcher’s In judging the quality of sampling, the main
ability to identify which variables or character- concern should be with its fitness for the specific
istics are likely to be significant. A ward or purpose of a particular study and the system-
clinic which is ‘typical’ on one dimension may aticity with which it has been carried out.
not be typical on another.45 Nevertheless, purpos-
ive sampling makes an important contribution to
the generalisability of qualitative research. JUDGING THE QUALITY OF QUALITATIVE
Other authors54–56 suggest that sampling should
be directed towards making possible generalisa-
RESEARCH
tions to theoretical propositions. This approach,
sometimes known as theoretical sampling, in- The contested nature of qualitative research is
volves directing case selection towards the testing most clearly evident in debates about its evalu-
of some developing theory rather than towards ation. Those who adopt the ‘two paradigms’
representativeness. Theoretical sampling uses approach outlined above, tend to argue that
existing theory to make predictions and then qualitative and quantitative research are so fun-
seeks out cases which allow the researcher damentally different that they cannot be evalu-
to test the robustness of these predictions ated by the same criteria.5,29,48,60 In particular,
under different conditions.53 As such, it parallels these authors reject the application of conven-
the use of critical experiments in laboratory tional criteria such as validity to a research
science. paradigm that, they claim, allows the possibility
Qualitative Methods in HTA 171
of multiple and conflicting truths. The altern- frequent candidates are respondent validation
ative position, adopted here, is that all HTA and triangulation. However, as Bloor in particu-
research, whether qualitative or quantitative, lar has shown, both of these have substantial
should be evaluated in terms of the robustness of problems when used as validity tests.64
its truth claims, on the one hand, and its useful- A number of authors recommend respond-
ness in informing policy-making and practice, ent validation (sometimes known as member
on the other. If research findings are to inform checking, host recognition or informant valida-
the development and use of health technologies tion) as a means of establishing the validity of
we must be confident that their findings are qualitative research findings.12,48,65–69 Respond-
sufficiently true for all practical purposes. Their ent validation can be carried out in several ways
relevance to the concerns identified by commis- (for details see Bloor70), but generally involves
sioners must also be clear. the presentation of findings (either orally or in
Whilst it is impossible, in either qualitative or writing) to the researched. If they accept the
quantitative research, to provide absolute proof findings as true, then validity is taken to have
of validity, researchers have a number of ways been demonstrated.
to ‘limit the likelihood of error’.61 The commit- There are four main difficulties with respond-
ment should be to what Dewey has termed ent validation. First, there are practical problems
‘warranted assertability’.62 Given the differences in asking research participants to comment on
between qualitative and quantitative research, in findings. Participants cannot be relied upon to
terms both of the problems which each addres- read or listen to research reports with the critical
ses and of their research practices, there will be attention required by this exercise. Researchers
different means by which warranted assertability who have used the technique report that they
and the researchers’ success in limiting error can have been both criticised and praised for things
be judged. The ways in which such judgements which their reports do not contain.71 Second,
can be made about qualitative research will now even where participants have clearly read and
be considered. fully understood reports, their responses cannot
It has become fashionable to call for the be assumed to represent ‘immaculately pro-
development of check-lists which can be used to duced’70 (p. 171) evaluations of the findings.
judge the quality of qualitative research reports. Responses are inevitably shaped and constrained
These calls often reflect the anxiety of journal by the context in which they are produced.
editors and systematic reviewers, who are more Validation interviews are reported to be marked
familiar with the outputs of quantitative research, by consensus seeking64 and deference to the
about their ability to sift wheat from chaff in rules of polite behaviour, which normally char-
qualitative research. While the commitment to acterise our social interactions. Participants are
quality control is commendable, and it is un- reluctant to challenge researchers’ findings and
doubtedly important to clarify the methods of eager, as Emerson and Pollner have put it, to ‘do
assessing truth claims from qualitative research, friendship’ with the researcher rather than to be
the ‘check-list’ approach seems unlikely to be a directly critical of his or her work.71 Third,
profitable strategy. As Hammersley argues, participants cannot be assumed to be entirely
no simple algorithmic criteria can be applied disinterested evaluators of findings. As members
mechanically to establish the quality of qual- of the setting studied they have their own micro-
itative research.63 The assessment of research political agendas. Both of these latter points
always involves judgement and it is impossible further exemplify the naivete of treating inter-
to side-step this by means of check-list criteria. view responses as straightforward reproductions
Check-lists risk becoming rigid constraints that of respondents’ mental states.37–39 Finally, as
are an end in themselves, rather than enhancing Bloor found, participants’ responses may not be
the validity of a study.27 The outcome is the consistent either within validation interviews or
research equivalent of defensive medicine in across time.64
which researchers produce findings which con- Taken together, these problems undermine
form to the check-list but are insignificant. any claim that respondent validity can be used
Below, we present an alternative to such mech- as a means of establishing the validity of
anical check-lists by suggesting a number of research findings. This procedure might reduce
questions which evaluators of research findings error. Discussions with participants about find-
might reasonably ask in relation to any piece of ings may highlight aspects of the setting that the
qualitative research. researcher has overlooked or offer an alternative
Even if a check-list approach to the validity of interpretation of data. In this way, they can
qualitative research could be established, it is contribute additional information that enhances
not clear what items it would include. Two the comprehensiveness of the study. However,
172 The Advanced Handbook of Methods in Evidence Based Healthcare
they cannot prove or disprove the validity of the the cultural basis of the doctor–patient relation-
findings.33,63,64,70,71 ship. Although the particular analysis which
The second validation technique often advo- they offered has since been a matter of some
cated in qualitative research is triangulation. debate, the methodological lesson has been
This rests on a metaphor drawn from military, widely learned.
navigational or surveying contexts.16,72–75 Ham- A fundamental problem with triangulation
mersley and Atkinson16 describe how triangula- lies in the asymmetry of the conclusions that can
tion is carried out in surveying: be drawn from it.64 If we assume that there is
For someone wanting to locate their position on a
one best method of investigating any research
map, a single landmark can only provide the infor-
topic, triangulation necessarily involves putting
mation that they are situated somewhere along a the findings of a superior method with those of
line in a particular direction from that landmark. an inferior. This is not a problem where the
With two landmarks, however, one’s exact position findings of the two methods converge, and the
can be pinpointed by taking bearings on both; one findings of the inferior method confirm those of
is at the point on the map where the two lines cross. the superior. However, where they diverge, we
(p. 231) do not know whether this arises because the
findings of the superior method were indeed
When applied to the evaluation of qualitative mistaken or because the inferior method is
research, triangulation involves the combination inadequate. In other words, triangulation can
of two or more methodologies in the study of only corroborate findings, it can never refute
the same phenomenon.76,77 This can involve them. As such it cannot be a validity test.64,81
using different methods, data sources, invest- Neither respondent validation nor triangula-
igators, disciplinary perspectives or theoretical tion can be used mechanically to test the validity
models.75,76,78 In all these cases, the underlying of the findings of qualitative research. The same
argument is that where the findings of two or is true of all the other measures that might be
more approaches agree, the validity of findings proposed for inclusion on a validity check-list.
is confirmed. However, this is not to suggest that we must
Just as with respondent validation, multi- abandon the pursuit of standards in qualitative
method approaches can enhance the compre- research. Qualitative research should be sub-
hensiveness of a study and reduce the likelihood jected to rigorous evaluation, and there are key
of error. However, there are again significant issues towards which evaluation can appropri-
problems if triangulation is elevated to the status ately be directed. However, these necessarily
of a validity test. As Silverman has argued,35,78 a involve judgement and cannot be achieved
strength of qualitative method is its recognition through the simple application of rigid algorith-
of the context-boundedness of data. Used as mic criteria. Some of the questions which evalu-
a test of validity, triangulation involves de- ators might usefully ask are discussed below.
contextualising data, leaving researchers unable
to relate their findings to the circumstances
of their production. Triangulation encourages What is the Quality of the Evidence
researchers to focus on identifying a single real-
ity rather than understanding the way partici-
upon which the Researcher’s Claims
pants use different versions of that reality in are Based?
different contexts.79 The resulting limitations are
well illustrated by the study of Welsh general The credibility of research findings depends
practice carried out by Stimson and Webb.80 crucially on the nature and quality of the evi-
They combined participant observation of gen- dence upon which they are based. To judge the
eral practice consultations and interviews with extent to which claims are justified, readers
patients. There were striking inconsistencies require information about the process by which
between patients’ reports of their interactions data were collected. Altheide and Johnson have
with doctors and the researchers’ own observa- suggested that there is a minimal set of issues
tions of such interactions. If Stimson and Webb that are encountered in almost all studies and
had used triangulation as a validity check, they should be detailed in research reports, so that
would have been forced to adjudicate between readers can assess the likely degree of error.82
their observations and the accounts given by These include: how access was negotiated; how
patients in interviews, presumably judging the the researchers presented themselves; the role
latter to be invalid. They chose instead to ana- the researchers adopted in the setting or inter-
lyse the interviews as data on the ways in which views; any ‘mistakes’ made by the researcher;
patients represented their interactions with doc- the types of data collected; and how such data
tors and to consider what this might imply for were recorded. On such a basis, readers can
Qualitative Methods in HTA 173
exercise what Glaser and Strauss have described considered alternative plausible explanations of
as their ‘joint responsibility’ to judge the evi- their data.27,61,84
dence alongside the researcher.83
Does the Research Build upon Existing racists, disability activists, gay and lesbian
Knowledge in the Field? polemicists, etc. – who have committed them-
selves to using research as a means of advocacy
The importance of attention to negative evi- on behalf of oppressed groups. This ‘underdog’
dence highlights a shortcoming of much current perspective has a long, though never unchal-
qualitative research.56 The difficulty of funding lenged, history in qualitative research.35,91
large-scale, multi-site studies means that the Research is always and inevitably shaped by
capacity to search comprehensively for negative values, but this does not oblige researchers to
evidence within a single study is inevitably adopt a partisan position. In seeking to limit the
constrained. To test the limits of findings emerg- possibility of error, researchers can plan to
ing from studies in a range of settings and under include the perspectives of people at different
various conditions, earlier work must be used as status levels within a group or setting. This
a foundation for subsequent studies. Qualitative commitment to ‘fair dealing’46 means that the
researchers have been curiously reluctant to perspectives of both the privileged and the less
work in this way. To some extent this reflects a privileged will be incorporated in the data col-
justifiable concern to avoid imposing ‘theories lection and analysis. It is the concern for even-
that have dubious fit and working capacity’88 handedness that distinguishes research from
(p. 4) upon data. It also reflects a widespread sensationalist or investigative journalism.46,92
misunderstanding of Glaser and Strauss’s pro- One of the ways in which the truth claims of a
gramme for grounded theory. When this was research report can be evaluated is the extent to
laid out in the 1960s, there was little to build on, which it offers understanding of the behaviour
and they emphasised the elements of discovery of people at all levels within an organisation
and theory-creation in qualitative research. In or setting. If the privileged and powerful are
the 1990s, the stress on innovation is less rele- presented as mere cardboard cutouts who are
vant and was greatly modified by Strauss, either ‘misguided or wilfully putting their own
although not by Glaser.89 As Strauss’s own late interests first’92 (p. 61) readers should be con-
work showed, it is certainly possible to avoid cerned about the value of the analysis in help-
the arbitrary imposition of theory while building ing them fully to grasp the complexity of the
on what is known from previous work. Our phenomenon being studied.
confidence in research findings will always be
increased where researchers show how their
findings articulate with other studies. A commit- WHEN TO USE QUALITATIVE RESEARCH
ment to cumulation increases opportunities for
the systematic search for negative evidence that
we have already discussed. There are at least four occasions on which HTA
research commissioners should consider using
qualitative research.
Is there Evidence of Partisanship in the Firstly, qualitative research can be an essen-
Research Report? tial precursor to quantitative work. It is a tool
for establishing that the right things are being
The recognition that any phenomenon may be counted using the right kind of operational
understood from a number of different stand- definitions. This is particularly important with
points has serious implications for the conduct ‘soft technologies’ like counselling. One of the
of qualitative research. In particular, researchers reasons why so many trials of counselling,
should be wary of presenting the perspective of psychotherapy and similar interventions have
one group as if this defined objective truth about been so inconclusive is their failure to standardise
the phenomenon and paying scant regard to for the intervention experience itself. The things
other perspectives. This kind of one-sided analy- that are easy to control, like the length of the
sis pervades research in the health field where session, the environment in which it is delivered,
doctors and senior managers are easily cast in or the certification of the counsellor, bear an
the role of villains. unknown relationship to what the counsellor
The debate over partisanship in research has a actually does. Qualitative work would make it
long history. There are those, such as Becker,90 possible to construct more homogeneous and
who have argued that research is always morally commensurable categories of intervention of the
and politically partisan and that the obligation kind required by experimental logic. However,
for researchers is not to seek neutrality, but to the application of qualitative work does go
make a moral decision about ‘whose side we are wider, particularly in understanding the way
on’ (p. 239). This commitment to principled people interact with ‘hard technologies’ as much
partisanship is characteristic of the work of a of the work on information technologies has
range of writers – Marxists, feminists, anti- shown.
Qualitative Methods in HTA 175
This spills into the second area, which is was supposed to: even more valuably, they are
where qualitative methods are used to explain the potential resource for finding out how to
unanticipated or inconclusive findings from improve the chances of implementing the next
quantitative studies, or from programme roll- piece of technology in a way that will achieve
outs. It is not uncommon to find that tech- its intended impact. How can it be brought into
nologies, which work in the design laboratory or use in ways that work with the grain of an
in carefully managed pilots, are far less success- organisation and its stakeholders rather than
ful in general use. This is a well-recognised being imposed in a fashion that invites sub-
problem in surgery, for instance, where a new version and misuse? These are problems that are
technique may be established by its pioneers to common to all human institutions and where the
the point where its mortality and postoperative generic theoretical tools of social science may
morbidity rates fall to acceptable levels, but point to a range of solutions from entirely un-
which then sees a marked jump in risk when it is expected quarters.
generally adopted. Although the normal pro-
cesses of professional trial-and-error will usually
correct the problem in time, the application of
qualitative methods may help to identify crucial CONCLUSIONS
differences between the innovation environment
and routine practice. This should reduce the Qualitative research has an important contribu-
amount of learning from adverse outcomes as tion to make to HTA. Decisions about whether
opposed to systematic proactive inquiry! qualitative or quantitative research (or some
A third area is that of hypothesis generation combination of the two) are more appropriate
or need identification. Before mounting a large for addressing a particular HTA problem should
and expensive RCT to establish whether a tech- be made in terms of which is most likely to
nology is an appropriate solution to some prob- resolve the problem in an efficient and effective
lem, it may be useful to get some sense of manner, rather than on philosophical or ideo-
whether it is reasonable to expect to find an logical grounds.
effect and, if so, what would be relevant dimen- Qualitative methods are particularly suited to
sions to incorporate into a trial design. If, for answering ‘how does it come to happen?’ ques-
example, we want to evaluate a new drug for the tions, rather than those concerned with ‘how
treatment of schizophrenia, it may be useful to many?’, ‘how much?’ or ‘how often?’. They have
carry out some qualitative studies of small an important contribution to make wherever the
patient groups at an early stage of clinical trial- context in which a health technology is applied
ing in order to identify possible areas of impact could be expected to have an impact upon
on interpersonal behaviour and social function- outcome. In some situations, particularly where
ing and to consider whether there are broad the groups being researched are involved in
variations in response by age, gender or eth- sensitive or stigmatised activities, qualitative
nicity. Should these emerge, they may then research may be the only practical, and therefore
inform the design of a major trial – for example the preferred, option. In others, a combination of
by suggesting that the costs and benefits for qualitative and quantitative methods will offer
women need to be considered separately from the optimum combination.
those of men in certain specific ways. In some Given the lack of standardisation in qual-
cases, of course, the hypotheses generated in itative research, a degree of caution is required
this way may not be amenable to testing using from both commissioners and users. Funda-
quantitative methods. In such cases, the re- mentally different objectives and commitments
searcher may use qualitative methods for hypoth- can drive research on the same substantive topic,
esis testing as well as hypothesis generation. using similar research techniques. The products
But the most important use of qualitative of research carried out by those who are com-
research is fundamentally to answer the crucial mitted primarily to advocacy on behalf of the
question for any manager, planner or policy- oppressed are likely to be substantially different
maker – what is going on here? It is not from those of researchers committed to the pur-
designed to deal with questions of number – suit of warrantable knowledge through scientific
how often?, how many?, how much? However, rigour. The validity of research findings and
these are a very particular and limited subset of their usefulness in informing both policy and
the important questions that anybody working in practice will depend upon the agenda which the
health technology needs to answer. Qualitative researcher brings to the research, as well as upon
methods are the tools for finding out why this the professional judgement, skills, competence
piece of technology is not having the impact it and systematicity with which it is carried out.
176 The Advanced Handbook of Methods in Evidence Based Healthcare
29. Henwood K, Pidgeon N. Qualitative research and 47. Honigman J. Sampling in ethnographic fieldwork.
psychological theorising. Br. J. Psychol. 1992; 83: In: Burgess RG (ed.). Field Research: A Source-
97–111. book and Field Manual. London: George, Allen
30. Oiler Boyd C. Combining qualitative and quan- and Unwin, 1982, pp. 79–90.
titative approaches. In: Munhall PL, Oiler Boyd C 48. Sandelowski M. The problem of rigor in qual-
(eds). Nursing Research: a Qualitative Perspec- itative research. Ans-Advances in Nursing Science
tive, 2nd edn. New York: National League for 1986; 8: 27–37
Nursing Press, 1993, pp. 454–75. 49. Johnson J. Selecting ethnographic informants. In:
31. Lindlof TR. Qualitative Communication Research Miller ML, Manning PK, Van Maanen J (eds).
Methods. London: Sage, 1995. Qualitative Research Methods. Newbury Park:
32. Foster P. Unit 12 Observational Research: In Sage, 1990.
DEH313 Course Team, Principles of Social and 50. Hammersley M. Some questions about theory in
Educational Research. Milton Keynes: The Open ethnography and history. In: Hammersley M
University, 1993, pp. 37–73. (ed.). What’s Wrong with Ethnography? London:
33. Emerson R. Observational fieldwork. Annu. Rev. Routledge, 1992, pp. 32–42.
Sociol. 1981; 7: 351–78. 51. LeCompte M, Preissle J. Ethnography and Qual-
34. Hammersley M. Deconstructing the qualitative- itative Design in Educational Research, 2nd edn.
quantitative divide. In: Hammersley M (ed.). New York: Academic Press Inc., 1993.
What’s Wrong with Ethnography? London: Rout- 52. Hammersley M. The generalisability of ethno-
ledge, 1992, pp. 159–82. graphy. In: Hammersley M (ed.). What’s Wrong
35. Silverman D. Interpreting Qualitative Data: with Ethnography? London: Routledge, 1992,
Methods for Analysing Talk, Text and Interaction. pp. 85–95.
London: Sage, 1993. 53. Firestone W. Alternative arguments for general-
36. Silverman D, Bor R, Miller R, Goldman E. ising from data as applied to qualitative research.
‘Obviously the advice then is to keep to safe sex’: Educational Researcher 1993; 22: 16–23.
Advice giving and advice reception in AIDS 54. Mitchell JC. Case and situational analysis. Socio-
counselling. In: Aggleton P, Davies P, Hart G logical Rev. 1983; 31: 187–211.
(eds). AIDS: Rights, Risks and Reason. London: 55. Bryman A. Quality and Quantity in Social
Falmer, pp. 174–91. Research. London: Unwin Hyman, 1988.
37. Scott M, Lyman S. Accounts. American Socio- 56. Silverman D. Telling convincing stories: a
logical Review 1968; 33: 46–62. plea for cautious positivism in case studies. In:
38. Scott M, Lyman S. Accounts, Deviance and the Glassner B, Moreno JD (eds). The Qualitative-
Social Order. In: Douglas J (ed.). Deviance and Quantitative Distinction in the Social Sciences.
Respectability. New York: Basic Books, 1970, Dordrecht: Kluwer Academic Publishers, 1989,
pp. 89–119. pp. 57–77.
39. Lyman S, Scott M. A Sociology of the Absurd, 57. Gubrium J, Buckholdt D. Describing Care: Image
2nd edn. New York: General Hall Inc., 1989. and Practice in Rehabilitation. Cambridge MA:
40. Smith H. Strategies of Social Research: The Oelgeschlager, Gunn and Hain, 1982.
Methodological Imagination, London: Prentice 58. Strong P. The Ceremonial Order of the Clinic.
Hall, 1975. London: Routledge and Kegan Paul, 1979.
41. Kennedy M. Generalizing from single case studies. 59. Parry O, Pill R. ‘I don’t tell him how to live his
Evaluation Quarterly 1979; 3: 661–78. life’: the doctor/patient encounter as an educa-
42. Miles MB, Huberman MA. Qualitative Data tional context. In: Bloor M, Taraborelli P (eds).
Analysis: An expanded sourcebook. 2nd edn. Qualitative Studies in Health and Medicine.
Thousand Oaks, CA: Sage, 1994. Aldershot: Avebury, 1994, pp. 5–21.
43. Denzin N. Interpretive Interactionism. In Morgan 60. Leininger M. Current issues, problems and trends
D (ed.). Beyond Method: Strategies for Social to advance qualitative paradigmatic research
Research. Beverly Hills: Sage, 1983, pp. 129– methods for the future. Qual. Health Res. 1992;
46. 12: 392–415.
44. Stake R. The Art of Case Study Research, 61. Hammersley M. Reading Ethnographic Research.
Thousand Oaks, CA: Sage. New York: Longman, 1990.
45. Schofield J. Increasing the generalizability of 62. Phillips D. Validity in qualitative research: why
qualitative research. In: Hammersley M (ed.). the worry about warrant will not wane. Education
Social Research: Philosophy, Politics and Prac- and Urban Society 1987; 20: 9–24.
tice. London: Sage, 1993, pp. 200–25. 63. Hammersley M. By what criteria should ethno-
46. Dingwall R. Don’t mind him – he’s from graphic research be judged? In: Hammersley M
Barcelona: qualitative methods in health studies. (ed.). What’s Wrong with Ethnography? London:
In: Daly J, McDonald I, Wilks E (eds). Research- Routledge, 1992, pp. 57–82.
ing Health Care. London: Tavistock/Routledge, 64. Bloor M. Techniques of validation in qualitative
1992, pp. 161–75. research: a critical commentary. In: Miller G,
178 The Advanced Handbook of Methods in Evidence Based Healthcare
Dingwall R (eds). Context and Method in 80. Stimson G, Webb B. Going to See the Doctor.
Qualitative Research. London: Sage, 1997, London: Routledge, 1975.
pp. 37–50. 81. Trend M. On the reconciliation of qualitative and
65. Frake C. How to ask for a drink in Subanun. quantitative analyses. In: Reichardt C, Cook T
American Anthropologist 1964; 66: 127–32. (eds). Qualitative and Quantitative Methods in
66. Frake C. How to enter a Yakan house. In: Evaluation Research. Beverly Hills: Sage, 1979,
Sanchez M, Blount B (eds). Socio-cultural pp. 68–86.
Dimensions of Language Use. New York: Aca- 82. Altheide D, Johnson J. Criteria for assessing
demic, 1975, pp. 25–40. interpretive validity in qualitative research.
67. Goodwin L, Goodwin H. Qualitative vs quantita- In: Denzin N, Lincoln Y (eds). Handbook of
tive or qualitative and quantitative research? Qualitative Research. Thousand Oaks: Sage,
Nursing Res. 1984; 33: 378–80. 1994, pp. 485–99.
68. Walker M. Analysing qualitative data: ethno- 83. Glaser BG, Strauss A. The discovery of sub-
graphy and the evaluation of medical education. stantive theory: a basic strategy underlying
Medical Education 1989; 23: 498–503. qualitative research. The American Behavioural
69. Beck CT. Qualitative research: the evaluation of Scientist 1965; 8: 5–12.
its credibility, fittingness and auditability. West. J. 84. Waitzkin H. On studying the discourse of medical
Nurs. Res. 1993; 15: 263–6. encounters. A critique of quantitative and qual-
70. Bloor M. Notes on member validation. In: itative methods and a proposal for a reasonable
Emerson RM (ed.). Contemporary Field Research: compromise. Medical Care 1990; 28: 473–88.
a collection of readings. Boston: Little, Brown, 85. Graffam Walker A. The verbatim record: the
pp. 156–72.
myth and the reality. In: Fisher S, Todd A (eds).
71. Emerson RM, Pollner M. On members’ responses
Discourse and Institutional Authority: Medicine,
to researchers’ accounts. Human Organization
Education and the Law. Norwood, NJ: Ablex,
1988; 47: 189–98.
1986, pp. 205–22.
72. Jick TD. Mixing qualitative and quantitative
86. Athens L. Scientific criteria for evaluating qual-
methods: triangulation in action. Administrative
itative studies. In: Denzin NK (ed.). Studies in
Science Quarterly 1979; 24: 602–11.
Symbolic Interaction 5. Greenwich, CT: JAI,
73. Flick U. Triangulation revisited: strategy of vali-
dation or alternative. J. Theor. Soc. Behav. 1992; 1984, pp. 259–68.
22: 175–97. 87. Secker J, Wimbush E, Watson J, Milburn K.
74. Nolan M, Behi R. Triangulation: the best of all Qualitative methods in health promotion research:
worlds? Br. J. Nurs. 1995; 4: 829–32 some criteria for quality. Health Education J.
75. Janesick V. The dance of qualitative research 1995; 54: 74–87.
design: metaphor, methodolatry and meaning. 88. Glaser BG, Strauss AL. The Discovery of
In: Denzin N, Lincoln Y (eds). A Handbook of Grounded Theory: Strategies for Qualitative
Qualitative Research. Thousand Oaks: Sage, Research. New York: Aldine, 1967.
1994, pp. 209–19. 89. Melia KM. Producing ‘plausible stories’: inter-
76. Denzin N. The Research Act, 2nd edn. New York: viewing student nurses. In: Miller G, Dingwall
McGraw-Hill, 1978. R (eds). Context and Method in Qualitative
77. Krefting L. Rigor in qualitative research: the Research. London: Sage, 1997, pp. 26–36.
assessment of trustworthiness. Am. J. Occup. 90. Becker HS. Whose side are we on? Social Prob-
Ther. 1991; 45: 214–22. lems 1967; 14: 239–48.
78. Silverman D. Qualitative Methodology and 91. Fielding N. Ethnography. In: Gilbert N (ed.).
Sociology. 1st edn. Aldershot: Gower, 1985. Researching Social Life. London: Sage, 1993,
79. Dingwall R. The ethnomethodological movement. pp. 154–71.
In: Payne G, Dingwall R, Payne J, Carter M 92. Voysey M. A Constant Burden: the Reconstitution
(eds). Sociology and Social Research. London: of Family Life. London: Routledge and Kegan
Croom Helm, 1981, pp. 124–38. Paul, 1975.
Part III
MEASUREMENT OF
BENEFIT AND COST
I N T R O D U C T I O N by J O H N B R A Z I E R
The acquisition of data on the benefits and costs represents a key set of methodological questions
of healthcare interventions forms a core activity in the design of HTA studies.
in Health Technology Assessment (HTA). Con- In HTA it has now become widely accepted
siderable resources are expended on gathering that conventional clinical measures should be
and processing routine administrative data sets supplemented by more patient-focussed meas-
and obtaining further data from patients, carers ures. The past 20 years has seen a rapid develop-
and professionals by questionnaire and other ment of patient based measures of outcome, and
methods. The resulting data form the foundation a plethora of instruments for collecting data on
for any assessment of the clinical or cost- health-related quality of life. A researcher con-
effectiveness of a new healthcare intervention. It sidering which instruments to use can consult
is imperative to collect the correct data for the one of the excellent reviews, but these quickly
assessment task, and to do so as efficiently as become out of date.1, 2 Furthermore, existing
possible. reviews may not address the specific condition
Assessing the benefits and costs in HTA or likely outcomes of the intervention being
involves three stages: the identification of what assessed. Researchers may need to undertake
benefits and resource costs to measure; the their own reviews of instruments. Fitzpatrick
measurement of those benefits and the use of and colleagues provide a non-technical and up-
resources; and finally, the valuation of the bene- to-date review of the key concepts for judging
fits and resources. The chapters in this section patient based measures.3 In addition to the usual
focus on key aspects of these three stages of concerns about the reliability and validity of a
assessment: Fitzpatrick and colleagues consider measure, they have brought together a range of
the identification and measurement of benefits other criteria in the design of a HTA study,
from the perspective of the patient; Brazier and including practicality and interpretability.
colleagues consider the implied values of the It is becoming increasingly common to
patient based measures of health status used in attempt to undertake economic evaluations
HTA and whether they are suitable for use in alongside clinical studies, since they provide an
economic evaluation; Cairns and van der Pol opportunity to collect patient specific data on
consider the valuation of health benefits over resource use and outcomes. This has implica-
time, and Johnston the identification and meas- tions for the type of data to collect. There are
urement of resource use in clinical trials. currently some excellent texts on the methods of
McColl reviews the use of questionnaire survey economic evaluation,4 but there has been rather
methods for measuring the benefits and costs of less written on the practical issues of data col-
healthcare. This is not an exhaustive list, but lection. One misconception has been the view
180 The Advanced Handbook of Methods in Evidence Based Healthcare
that an economic evaluation can always be address a range of theoretical and practical ques-
undertaken retrospectively after the trial has tions about the conduct of questionnaire surveys
been completed. This may be appropriate in in the health field, not in the ideal, but in the
some circumstances but often it leaves the analyst resource constrained world of research studies.
without the means to perform an economic An important area of controversy in HTA has
evaluation. The precise data requirements of been how to treat benefits which are expected to
economic evaluation will vary depending on the occur sometime in the future. Previous practice
technology and the decision-making context. has been to apply the same positive rate of
What additional data are required on costs and discount to health benefits as to costs (i.e. this
benefits must be carefully considered at the assumes that health gains today are preferred to
design of HTA studies. equivalent gains later), but this has been ques-
The chapter by Brazier and colleagues exam- tioned in recent years. The chapter by Cairns
ines the use of patient based measures of health and van der Pol reviews the rationale for dis-
status in economic evaluation. Many of these counting health benefits,7 looking at past studies,
measures have not been designed for the pur- estimating the appropriate rate for health benefits
pose of undertaking economic evaluation, and of saving lives or improving health status, and at
this review identifies their limited role in econ- the design of a study addressing some of the key
omic evaluation. The chapter goes on to review questions in estimating time preference rates for
the different economic measures designed to use in HTA.
estimate a single index value for health-related Although these reviews are not comprehen-
quality of life which is used in the estimation of sive in their coverage of the issues in measuring
quality adjusted life years (QALYs). The differ- the benefits and costs of health technologies,
ent multi-attribute utility scales and techniques they do provide a unique collection of reviews
for health state valuation are reviewed, and of a set of fundamental issues in HTA. For an
advice offered to researchers contemplating overview of cost and benefit measurement,
undertaking an economic evaluation alongside a readers are advised to refer to established text,
clinical study. Readers interested in non-QALY such as that by Drummond et al.4
economic measures of benefits are referred to an
overview by Johannesson et al.5
The next chapter examines the somewhat
neglected topic of collecting resource use data
REFERENCES
for costing purposes in clinical trials. It covers
many important practical issues often over- 1. Bowling, A. Measuring health: a review of quality
looked in the economics literature. Johnston and of life and measurement scales. Milton Keynes:
her colleagues consider what resource items to Open University Press, 1991.
include, how to measure them, and how to value 2. Wilkin D, Hallam L, Doggett MA. Measures of
them. They examine the roles of economic need and outcome for primary health care. Oxford:
theory and perspective in determining what to Oxford Medical Press, 1992.
measure, and provide practical advice on how to 3. Streiner DL, Norman GR. Health Measurement
collect data on the economically significant Scales: a practical guide to their development and
resource items in clinical trials and from routine use. Oxford: Oxford University Press, 1989.
administrative data sets. 4. Drummond MF, O’Brien B, Stoddart GL, Torrance
Whether data are being collected for use in a GW. Methods for the economic evaluation of
clinical or economic evaluation, an important health care programmes. Oxford: Oxford Medical
method of collecting data is the questionnaire Publications, 1997.
survey. There is a very large literature on the 5. Johannesson M, Jonsson B, Karlsson G. Outcome
conduct and design of questionnaire surveys measurement in economic evaluation. Health
in the social sciences, much of it outside the Economics 1996; 5: 279–98.
field of health. McColl and colleagues supple- 6. Sudman S, Bradburn N. Asking questions: a prac-
ment existing textbooks (such as Sudman and tical guide to questionnaire design. San Francisco:
Bradburn6) by bringing together this very Jossey-Bass, 1993.
diverse literature on the evidence and expert 7. Parsonage M, Neuberger H. Discounting and
opinion found in more theoretical texts. They health benefits. Health Economics 1992; 1: 71–6.
11
Criteria for Assessing Patient Based
Outcome Measures for Use in
Clinical Trials
SUMMARY INTRODUCTION
enormous array of such measures that can be status’ and ‘quality of life’ are used interchange-
used in clinical trials. The purpose of the chapter ably to such an extent that they lack real
is to make clear the criteria investigators should descriptive value. The term ‘patient based out-
have in mind when they select patient based come measure’ is here used wherever possible
outcome measures at the stage of designing a as the most all-embracing term to encompass all
clinical trial. of the types of instruments conveyed by other
This chapter is based on a structured review terms such as ‘health status’, or ‘quality of
carried out for the NHSE R&D programme.1 life’.
The review used a combination of electronic and There are circumstances where patients are
hand searching. A total of 5621 abstracts and unable to provide their unique report of their
references were identified by initial searches as perceptions, due to ill-health, physical or cogni-
potentially relevant to the review, but only 404 tive problems, or some other incapacity. In these
references were actively used in the final draft cases, proxy reports may be necessary because
review. An initial first draft was written by one of the need for some assessment to inform a
of the present authors (R.F.). A second draft, clinical trial. Because there is consistent evi-
revised in the light of discussion with the other dence of lack of agreement of patients’ judge-
three authors, was then sent out to a panel of ten ments of their quality of life with assessments
experts from the fields of clinical medicine, made by observers such as health professionals,
clinical trials, health economics, health services informal carers, and other so-called proxy judges,
research, psychology, sociology and statistics. this is increasingly considered a second-best solu-
They were invited to comment both on broad tion to be used only when the patient cannot
and strategic aspects of the approach of the contribute.2
review, as well as on whatever specific details The content of the instruments with which we
they wished. The literature review was then
are concerned varies enormously, and in general
revised considerably in the light of this set of
a researcher will usually be able to find an
comments. The final draft, of which this is a
instrument with questionnaire items that at least
summary, is a quasi-consensus view of a wide
range of experts as to key issues surrounding approximate to the issues of concern to his or
selection of patient based outcome measures. her research question. Every instrument attempts
to provide an assessment of at least one dimen-
sion of health status, either the respondent’s
Concepts and Definitions global assessment of health, or more specific
dimensions such as mobility, pain or psycho-
This is a review of a field in which there is no logical well-being. Instruments range in content
precise definition of or agreement about subject from items which are most obviously related to
matter. We are concerned with questionnaires a patient’s health status, such as the patient’s
and related instruments that ask patients about global view of their health, experiences of
their health. However, with regard to more pre- symptoms or psychological illness, through to
cise definitions of what such instruments are dimensions that increasingly reflect the broader
intended to assess, there is no agreed terminol- impact of illness on the individual’s life such as
ogy, and reviews variously refer to instruments social function, role activities and impact on
as being concerned with ‘quality of life’ (QoL), paid income.
‘health-related quality of life’ (HRQoL), ‘health
status’, ‘functional status’, ‘performance status’,
‘subjective health status’, ‘disability’ and ‘func- Types of Instruments
tional well-being’. To some extent this diversity
reflects real differences of emphasis between One of the main decisions to be made in select-
instruments. Some questionnaires focus exclus- ing an instrument for a clinical trial is to choose
ively upon physical function; for example, among the different kinds of instrument that
assessing mobility and activities of daily living exist. They differ in content and also in the
without reference to social and psychological primary intended purpose.
factors, and might appropriately be described as
functional status instruments. Other instruments Disease/condition-specific
may ask simple global questions about the indi-
vidual’s health. Other instruments again are con- As the title implies, these instruments have been
cerned with the impact of health on a broad developed in order to provide the patient’s per-
spectrum of the individual’s life, for example, ception of a specific disease or health problem.
family life and life satisfaction, and might An example of such a questionnaire is the
reasonably be considered to assess quality of Asthma Quality of Life Questionnaire.3 This
life. In reality, the various terms such as ‘health contains 32 questions assessing four dimensions
Assessing Patient Based Outcome Measures in Clinical Trials 183
(activity limitations, symptoms, emotional func- domains or concerns that are of personal con-
tion and exposure to environmental stimuli). cern that are not predetermined by the invest-
igator’s list of questionnaire items. By a variety
of means the respondent is encouraged to iden-
Dimension-specific tify those aspects of life that are personally
Dimension-specific instruments assess one spe- affected by health, without imposing any stan-
cific aspect of health status. By far the most dardised list of potential answers. Individualised
common type of dimension-specific measure is measures are still in their infancy, but have
one that assesses aspects of psychological well- attracted interest precisely because they appear
being. An example is the Beck Depression to offer considerable scope for eliciting respond-
Inventory.4 This contains 21 items that address ents’ own concerns and perceptions. One ex-
symptoms of depression, the scores for items ample is the Schedule for the Evaluation of
being summed to produce a total score. It Individual Quality of Life (SEIQoL).6 This is
was largely developed for use in psychiatric completed in three phases by semi-structured
patients, but is increasingly used more widely to interview in order to produce an overall quality
assess depression as an outcome in physically ill of life score for sick or healthy people. The first
populations. stage asks the individual (with structured inter-
viewer prompting when necessary) to list five
areas of life most important to their quality of
Generic instruments life. Secondly, each of the nominated five areas
Generic instruments are intended to capture a is rated on a visual–analogue scale from ‘as
very broad range of aspects of health status and good as it could be’ to ‘as bad as it could be’.
the consequences of illness, and therefore to be The individual patient also rates overall quality
relevant to a wide range of patient groups. The of life. The last stage uses 30 hypothetical case
content of such questionnaires has been deliber- vignettes which vary systematically in terms of
ately designed to be widely appropriate. They the properties respondents have already identi-
may provide assessments of the health status of fied as important to them. Judgement analysis of
samples of individuals not recruited because respondents’ ratings of these vignettes allows
they have a specific disease, for example, from the investigator to produce weights for the five
primary care or the community as a whole. One chosen aspects of life, and an index score is
of the most widely used of such instruments is calculated between 0 and 100. This exercise can
the SF-36.5 This is a 36-item questionnaire then be repeated at subsequent assessments.
which measures health status in eight dimen-
sions: physical functioning, role limitations due Utility measures
to physical problems, role limitations due to
emotional problems, social functioning, mental Utility measures are distinctive because they
health, energy/vitality, pain and general percep- yield a single number to summarise individuals’
tions of health. Item responses are summed to overall preferences regarding health status. The
give a score for each dimension. most familiar example in Europe is the EuroQol
EQ-5D.7 The part of the EQ-5D questionnaire to
elicit health status comprises five questions,
Summary items each of which has three alternative response
Single questionnaire items have an important categories. The five items assess mobility, self-
role in healthcare research. They invite respond- care, usual activity, pain/discomfort and anxiety/
ents to summarise diverse aspects of their health depression. These items can be used by them-
status by means of one or a very small number selves as descriptions of respondents’ health
of questions. The General Household Survey for states. Responses are also scored by means of
England and Wales has, in annual surveys since weights obtained from the valuations that other
1974, used two questions that together provide samples from the general population have
an assessment of chronic illness and disability: assigned to health states using visual–analogue
‘Do you have any long-standing illness or dis- scales or time–trade-off methods (time–trade-off
ability?’ and ‘Does this illness or disability limit is discussed in Chapter 12). These weightings
your activities in any way?’ A positive answer combine to provide utilities.
to the two questions provide an indication of
chronic disability.
CRITERIA FOR ASSESSING INSTRUMENTS
Individualised Measures
Individualised measures are instruments in which There are eight issues that need to be examined
the respondent is allowed to select issues, when selecting a patient based outcome measure
184 The Advanced Handbook of Methods in Evidence Based Healthcare
for a trial: appropriateness, reliability, validity, often not clear with investigators uncritically
responsiveness, precision, interpretability, ac- inserting questionnaires into their trials without
ceptability and feasibility. These criteria are careful consideration of content and relevance to
discussed in turn. the purpose of the trial.10
Reliability
Appropriateness
Reliability is concerned with the reproducibility
The first and most fundamental consideration to and internal consistency of a measuring instru-
be faced when selecting a patient based outcome ment. It assesses the extent to which the instru-
measure is how to identify one that is most ment is free from random error, and may be
appropriate to the aims of the particular trial. considered as the proportion of a score that is
This requires careful consideration of the aims signal rather than noise. It is a very important
of the trial, with reference to the quality of life property of any patient based outcome measure
research questions, i.e. which dimensions will be in a clinical trial because it is essential to
primary and secondary end-points, the nature of establish that any changes observed in a trial are
the study intervention and of the patient group, due to the intervention and not to problems in
and about the content of possible candidate the measuring instrument. As the random error
instruments. For this reason it is particularly of such a measure increases, so the size of the
difficult to give specific recommendations about sample required to obtain a precise estimate of
what in general makes an outcome measure effects in a trial will increase. An unreliable
appropriate to a trial, because this is ultimately measure may therefore imprecisely estimate the
a judgement of the fit between investigators’ size of benefit obtained from an intervention.
specific trial questions and content of instru- The reliability of a particular measure is not a
ments. However, it is clear from a number of fixed property, but is dependent upon the con-
reviews already carried out in this field that it is text and population studied.11
an absolutely fundamental issue. In practice, the evaluation of reliability is
There have been several previous reviews that made in terms of two different aspects of a
have discussed appropriateness of outcome meas- measure: internal consistency and reproducibil-
ures in clinical trials in general terms. Some of the ity. The two measures derive from classical
early reviews are mainly critical of clinical trials measurement theory which regards any observa-
for failing to use any kind of patient based tion as the sum of two components, a true score
outcome measure where the subject matter and an error term.12
seemed to indicate that such an approach was
needed. Thus, Brook and Kamberg8 concluded
that, from a sample of 73 clinical trials in which
Internal consistency
they considered health status or quality of life Normally, more than one questionnaire item is
was likely to be a major issue, in only two trials used to measure a dimension or construct. This
was an appropriate patient based outcome meas- follows from a basic principle of measurement
ure used. that several related observations will produce a
A more formal evaluation of outcome meas- more reliable estimate than one. For this to be
urement in trials is reported by Guyatt and true, the items all need to be homogeneous, that
colleagues.9 In their study, two raters independ- is all measuring aspects of a single attribute or
ently examined all clinical trials published in a construct rather than different constructs.11 The
range of journals in 1986. Of the 75 trials they practical consequence of this expectation is that
evaluated, they considered quality of life as individual items should highly correlate with
crucial or important in 55 (73%) of trials. How- each other and with the summed score of the
ever, in 44% of this subgroup of trials no effort total of items in the same scale.
was made to assess this dimension of outcome. Internal consistency can be measured in a
In a further 38% of the 55 trials an untested number of different ways. One approach – split-
measure was used that the reviewers considered half reliability – is randomly to divide the items
inappropriate. They concluded that appropriate in a scale into two groups and to assess the
measures would have considerably strengthened degree of agreement between the two halves.
the basis for recommendations emerging from The two halves should correlate highly. An
the trial. extension of this principle is Coefficient alpha,
Most obviously an instrument needs to fit the usually referred to as Cronbach’s alpha, which
purpose of a trial. This purpose needs to be essentially estimates the average level of agree-
specified as precisely as is reasonable, and ment of all the possible ways of performing
outcome measures selected accordingly. The split-half tests.13 The higher the alpha, the higher
rationale for selection of outcome measures is the internal consistency. However, it is also
Assessing Patient Based Outcome Measures in Clinical Trials 185
possible to increase Cronbach’s alpha by Commonly cited minimal standards for reli-
increasing the number of items, even if the ability coefficients are 0.7 for group data,
average level of correlation does not change.11 although some experts set much higher require-
Also, if the items of a scale correlate perfectly ments.17 It can also be argued that absolute mini-
with each other, it is likely that there is some mally acceptable coefficients are not meaning-
redundancy among items, and also a possibility ful, since larger sample sizes for a trial permit
that the items together are addressing a rather more measurement error in an instrument.
narrow aspect of an attribute. For these reasons
it is suggested that Cronbach’s alpha should be
above 0.70 but not higher than 0.90.11,14 Validity
It has been argued that excessive attention to
internal reliability can result in the omission of The validity of a measure is an assessment of
important items, particularly those that reflect the extent to which it measures what it purports
the complexity and diversity of a phenomenon.15 to measure. There are a number of different
Certainly obtaining the highest possible reliabil- ways of establishing the validity of a measure.
ity coefficient should not be the sole objective in As with reliability, it is not a fixed property of a
developing or selecting an instrument, because measure; its validity is assessed in relation to a
the reductio ad absurdum of this principle would specific purpose and setting.14,18 It is therefore
be an instrument with high reliability produced meaningless to refer to a validated measure as
by virtually identical items. such; it should be considered a measure vali-
dated for use in relation to a specific purpose or
Reproducibility set of purposes. For example, a valid measure
of disability for patients with arthritis cannot
Reproducibility more directly evaluates whether automatically be considered valid for use for
an instrument yields the same results on patients with multiple sclerosis; a measure con-
repeated applications, when respondents have sidered validated for individuals with mild
not changed on the domain being measured. impairment may not be valid for those with
This is assessed by test–retest reliability. The severe impairments.
degree of agreement is examined between scores Some types of validity are not relevant to
at a first assessment and when reassessed. There this field. Criterion validity is concerned with
is no exact agreement about the length of time whether a new measure correlates with an
that should elapse between test and retest; it existing measure generally accepted as a more
needs to be a sufficient length of time that accurate or criterion variable. However, in the
respondents are unlikely to recall their previous field of application of health status measures
answers, but not so long that actual changes in with which we are concerned, as outcome meas-
the underlying dimension of health have occur- ures in clinical trials, rarely if ever does a
red. Streiner and Norman11 suggest that the usual perfect ‘gold-standard’ measure exist against
range of time elapsed between assessments which to test the validity of new health status
tends to be between 2 and 14 days. measure, and a number of different and more
Test–retest reliability is commonly examined indirect approaches are recommended to judge
by means of a correlation coefficient. This is instruments’ validity.
often the Pearson product moment correlation
coefficient. This approach is limited and may Face and content validity
exaggerate reproducibility because results from
two administrations of a test may correlate Face, content and (below) construct validity are
highly but be systematically different. The sec- far the most relevant issues for the use of patient
ond test may result in every respondent having a based outcome measures in trials. It is vital to
lower score than their first response, yet the inspect the content of a measure in relation to
correlation could be 1.0. For this reason, an its intended purpose. This inspection largely
intra-class correlation coefficient is advocated. involves qualitative matters of judgement that
This uses analysis of variance to determine how contrast with more statistical criteria that also
much of the total variability in scores is due need to be considered in the context of construct
to true differences between individuals and validity. Judgement of the content of an instru-
how much due to variability in measurement. ment contributes to what has been termed face
Alternatively, Bland and Altman16 advocate validity and content validity. The two terms are
graphically plotting scores for the two admin- related but have been distinct. Guyatt and col-
istrations of a test, so that, for example, it is leagues19 make the distinction thus: ‘Face valid-
possible to identify areas in the range of scores ity examines whether an instrument appears to
of an instrument which are less reproducible. be measuring what it is intended to measure, and
186 The Advanced Handbook of Methods in Evidence Based Healthcare
One common form of standardised expression methods they used in a sample of patients before
of responsiveness is the effect size. The basic and after total hip replacement surgery.
approach to calculation of the effect size is to
calculate the size of change on a measure that Ceiling and floor effects
occurs to a group between assessments (for
example, before and after treatment), compared The actual form of questionnaire items in an
with the variability of scores of that measure.26 instrument may reduce the likelihood of further
Most commonly this is calculated as the differ- improvement or deterioration being recorded
ence between mean scores at assessments, beyond a certain point. Put another way, the
divided by the standard deviation of baseline wording of questionnaire items may not make it
scores. The effect size is then expressed in possible to report most favourable or worst
standardised units that permit comparisons health states. The terms ‘ceiling’ and ‘floor’
between instruments. effects are usually used to refer to the two forms
of this problem. A study administered the MOS-
20 scale to patients in hospital at baseline and
Standardised response mean again six months later.33 At the follow-up survey
An alternative measure is the standardised respondents also completed a ‘transition ques-
response mean (SRM). This only differs from an tion’ in which they assessed whether their health
effect size in that the denominator is the stand- was better, the same or worse than at baseline
ard deviation of change scores in the group in assessment. A number of respondents who
order to take account of variability in change reported the worst possible scores for the MOS-
rather than baseline scores.27 Because the 20 at baseline reported further deterioration in
denominator in the SRM examines response their follow-up assessment in their answers to
variance in an instrument, whereas the effect the transition question. It was clearly not pos-
size does not, Katz and colleagues28 consider sible for such respondents to report lower scores
that the SRM approach is more informative. on the MOS-20 than at baseline.
health status scores could be related to other Patients may also be unable to complete ques-
human experiences that have clear and intuitive tionnaires because of their health status or other
meaning. disabilities, particularly cognitive or visual.
Another approach to interpreting results is More general features of the layout, appear-
to identify a plausible range within which a ance and legibility of a questionnaire are
minimal clinically important difference (MCID) thought to have a strong influence on accept-
falls.41,42 Jaeschke et al. define a MCID as ‘the ability. Some instruments have deliberately
smallest difference in score in the domain of included extremely simple and short forms of
interest which patients perceive as beneficial and wording of questions together with pictorial
which would mandate, in the absence of trouble- representations to add to ease and acceptability
some side effects and excessive costs, a change of use. A rare experimental study to test the
in the patient’s management’ (p. 408). They benefit of pictorial representation in a quality of
examined this concept in relation to patients life study showed that cartoon figures to depict
completing at baseline and follow-up either the degrees of illness severity improved test–retest
Chronic Respiratory Questionnaire in a drug reliability compared with responses to conven-
trial for asthma or the Chronic Heart Failure tional formatting.44
Questionnaire in a drug trial for patients with There is only limited evidence available com-
heart failure. Changes between baseline and paring the response rates of different health
follow-up were examined in relation to their status instruments, rather than the method of
benchmark for a MCID which was the patient’s their administration. In a series of older patients
follow-up assessment in a transition item of who had undergone total hip replacement sur-
whether they were worse, better or the same com- gery, higher completion rates were obtained
pared with baseline assessment. They showed from a 12-item condition-specific questionnaire
that a mean change of 0.5 for a seven-point scale compared to a longer generic instrument, the
was the minimal change amongst patients SF-36.45
reporting a change. Other methods of under- Another form of evidence comes from the
standing clinically important changes require the differential responses to different subject matters
selection of other external benchmarks such as in surveys of health status. Guyatt and col-
the global judgement of the clinician or lab- leagues46 found that a sample of elderly respond-
oratory tests or reference to distribution based ents were somewhat more likely to complete the
interpretations, such as using effect size.27,43 section of a questionnaire concerned with phys-
ical rather than with emotional items, suggesting
Acceptability differential acceptability of topics depending on
how personal they were. By contrast, in a qual-
It is essential that instruments be acceptable to itative study of patients with small-cell lung
patients. This is clearly desirable to minimise cancer47 it was reported that patients found ques-
avoidable distress to patients already coping tions about their psychological well-being more
with health problems. It is also essential in order tolerable than questions about tumour-related
to obtain high response rates to questionnaires in symptoms.
order to make results of trials more easy to
interpret, more generalisable, and less prone to Time to complete
bias from non-response. The acceptability of
patient based outcome measures has far less It is often assumed that one aspect or deter-
frequently been examined than issues such as minant of the acceptability of a questionnaire is
reliability and validity. its length; the longer it takes to complete, the
less acceptable is the instrument. Many instru-
ments are published with claims by their devel-
Reasons for non-completion opers about the length of time required to com-
Patients may either not return a whole assess- plete them. Far less commonly is this property
ment, or may omit some items in an assessment. independently assessed or instruments’ time to
If patients either do not attempt to complete an complete measured comparatively. Amongst
instrument at all or omit particular items fre- instruments requiring the least time to complete
quently, this is potentially a sign that a ques- are the self-completed COOP charts which have
tionnaire is difficult to understand, distressing, been estimated to take 2–3 min.48 Similarly,
or in some other way unacceptable. However, Wolfe and colleagues49 directly assessed the
there may be other reasons for non-completion mean length of time required to complete one of
such as the method of delivery of the ques- the most commonly used of instruments for
tionnaire. Patients may not receive a mailed arthritis – the Health Assessment Questionnaire
questionnaire in the first place or may not have a (HAQ) – as 3 min. Most health status instru-
telephone in order to be contacted in this way. ments are longer than these two examples,
190 The Advanced Handbook of Methods in Evidence Based Healthcare
and probably require more time to complete. annoying or distressing, or whether issues were
Aaronson and colleagues50 directly measured omitted. When the EORTC QLQ-C30 was
time to complete the EORTC QLQ-C30 on two assessed in this way, 10% of patients reported
separate occasions, before and during active one or more items was confusing or difficult to
treatment (12 and 11 min, respectively). The answer, and less than 3% that an item was up-
time required may depend upon the character- setting, whilst more generally patients welcomed
istics of respondents. the opportunity to report their experiences.50
A smaller number of studies have examined Another formal evaluation of acceptability of a
comparatively the time required for various questionnaire found that 89% enjoyed the task
instruments or methods of administration. Wein- of completing the COOP instrument, and 97%
berger and colleagues51 assessed the time reported understanding the questions.48
required for SF-36 to be completed by two In general, users should expect to see evi-
different methods of administration; when self- dence of acceptability being examined at the
completed, the instrument required 12.7 min, design stage. Subsequently, the most direct and
compared to 9.6 min for face-to-face interviews. easy to assess evidence are the length and
In an elderly group of patients the SF-36 took response rates of questionnaires.
14 min by personal interview and 10.2 min
by telephone administration.52 Read and col- Translation and cultural applicability
leagues53 compared the time to administer of the
General Health Rating Index, the Quality of One basic way in which a questionnaire may fail
Well-Being Scale and the Sickness Impact Pro- to be acceptable is if it is expressed in a lan-
file, which required 11.4, 18.2 and 22.4 min, guage unfamiliar to respondents. This issue has
respectively. Generally such evidence is not received a large amount of attention in recent
available. In a comparative study of health status literature on patient based outcomes, mainly
measures of outcomes, Bombardier and col- because of the increasing need for clinical trials
leagues54 estimated that the HAQ required 5 min incorporating quality of life measures to be
to complete, compared to three different utility conducted on a multi-national basis, especially in
measures that required administration by inter- Europe.58,59 As a result there are quite elaborate
view and between 30 and 60 min to complete. guidelines available intended to ensure high
Independent assessment of timing is important standards of translation of questionnaires.60,61
because developers of instruments may be over- Amongst procedures to improve translation,
optimistic in their estimates. according to such guidelines, are: use of several
It is increasingly argued that, if there are no or independent translations that are compared;
minimal costs in terms of validity, responsive- back-translation; testing of the acceptability of
ness and other key components of instruments, translations to respondents. Less attention has
then instruments should be reduced in terms of been to cultural and linguistic variations within
length and number of items in order to increase national boundaries, but it would seem that
acceptability.55 Some comparative studies have similar principles could be applied to increase
shown no loss of responsiveness when such cultural applicability. Presently, few patient
shorter instruments are used.28,56 based outcome measures have been translated
Evidence of difficulties with an instrument is into the languages of ethnic minorities in the
provided by Guyatt and colleagues,35 who com- UK.
pared the measurement properties of Likert and
visual–analogue forms of response categories to Feasibility
a health-related quality of life instrument. In
explaining the two forms of task to patients they In addition to patient burden and acceptability, it
found that patients viewed visual–analogue is important to evaluate the impact of different
scales as harder to understand. In specific terms, patient based outcome measures upon staff and
they report that it took up to twice as long to researchers in collecting and processing infor-
explain. mation.62,63 Data from patients for clinical trials
are often gathered in the context of regular
clinical patient care, and excessive burden to
Direct assessment of acceptability staff may jeopardise trial conduct and disrupt
It is preferable directly to assess patients’ views clinical care. An obvious example is the addi-
about a new questionnaire. Sprangers and col- tional staff effort and costs involved in person-
leagues57 argue that patients’ views should be ally administering questionnaires compared to
obtained at the pre-testing phase prior to formal postal delivery. To a lesser extent the length and
tests for reliability, etc., by means of a struc- complexity of instrument are an additional com-
tured interview in which they are asked whether ponent. Certainly it may require additional staff
they found any questionnaire items difficult, time to assist and explain how more complex
Assessing Patient Based Outcome Measures in Clinical Trials 191
in Medical Sociology. JAI Press Inc.: London, 49. Wolfe F, Kleinheksel SM, Cathey MA, Hawley
1994, pp. 255–80. DJ, Spitz PW, Fries JF. The clinical value of the
35. Guyatt GH, Townsend M, Berman LB, Keller JL. Stanford Health Assessment Questionnaire Func-
A comparison of Likert and visual analogue tional Disability Index in patients with rheuma-
scales for measuring change in function. J. toid arthritis. J. Rheumatol. 1988; 15: 1480–8.
Chronic. Dis. 1987; 40: 1129–33. 50. Aaronson NK, Ahmedzai S, Bergman B, et al.
36. Jaeschke R, Singer J, Guyatt GH. A comparison The European Organization for Research and
of seven-point and visual analogue scales. Data Treatment of Cancer QLQ-C30: a quality-of-life
from a randomized trial. Controlled Clinical instrument for use in international clinical trials
Trials 1990; 11: 43–51. in oncology. J. Natl Cancer Inst. 1993; 85:
37. Jenkinson C. Why are we weighting? A critical 365–76.
examination of the use of item weights in a 51. Weinberger M, Oddone EZ, Samsa GP, Lands-
health status measure. Soc. Sci. Med. 1991; 32: man PB. Are health-related quality-of-life meas-
1413–16. ures affected by the mode of administration? J.
38. Stucki G, Daltroy L, Katz JN, Johannesson M, Clin. Epidemiol. 1996; 49: 135–40.
and Liang MH. Interpretation of change scores in 52. Weinberger M, Nagle B, Hanlon JT, Samsa GP,
ordinal clinical scales and health status measures: Schmader K, Landsman PB, Uttech KM, Cowper
the whole may not equal the sum of the parts. J. PA, Cohen HJ, Feussner JR. Assessing health-
Clin. Epidemiol. 1996; 49: 711–17. related quality of life in elderly outpatients: tele-
39. Deyo RA, Patrick D L. Barriers to the use of phone versus face-to-face administration. J. Am.
health status measures in clinical investigation, Geriatr. Soc. 1994; 42: 1295–99.
patient care, and policy research. Med. Care 53. Read JL, Quinn RJ, Hoefer MA. Measuring
1989; 27: S254–68. overall health: an evaluation of three important
40. Testa MA, Anderson RB, Nackley JF, Hollenberg approaches. J. Chron. Dis. 1987; 40(Suppl. 1):
NK. Quality of life and antihypertensive therapy 7S–26S.
in men. A comparison of captopril with enalapril. 54. Bombardier C, Raboud J and The Auranofin
The Quality-of-Life Hypertension Study Group.
Cooperating Group. A comparison of health-
N. Engl. J. Med. 1993; 328: 907–13.
related quality-of-life measures for rheumatoid
41. Jaeschke R, Singer J, Guyatt GH. Measurement
arthritis research. Controlled Clinical Trials 1991;
of health status. Ascertaining the minimal clinic-
12: 243S–256S.
ally important difference. Controlled Clinical
55. Burisch M. You don’t always get what you pay
Trials 1989; 10: 407–15.
for: measuring depression with short and simple
42. Jaeschke R, Guyatt GH, Keller JL, and Singer J.
versus long and sophisticated scales. J. Res.
Interpreting changes in quality-of-life score in n
Personality 1984; 18: 81–98.
of 1 randomized trials. Controlled Clinical Trials
1991; 12: 226S–233S. 56. Fitzpatrick R, Newman S, Lamb R, Shipley M. A
43. Deyo RA, Patrick DL. The significance of treat- comparison of measures of health status in rheu-
ment effects: the clinical perspective. Med. Care matoid arthritis. Br. J. Rheumatol. 1989; 28:
1995; 33: AS286–91. 201–6.
44. Hadorn DC, Hays RD, Uebersax J, Hauber T. 57. Sprangers MA, Cull A, Bjordal K, Groenvold M,
Improving task comprehension in the measure- Aaronson NK. The European Organization for
ment of health state preferences. A trial of infor- Research and Treatment of Cancer. Approach to
mational cartoon figures and a paired-comparison quality of life assessment: guidelines for develop-
task. J. Clin. Epidemiol. 1992; 45: 233–43. ing questionnaire modules. EORTC Study Group
45. Dawson J, Fitzpatrick R, Murray D, Carr A. on Quality of Life. Qual. Life Res. 1993; 2:
Comparison of measures to assess outcomes in 287–95.
total hip replacement surgery. Qual. Health Care 58. Orley J, Kuyken W. Quality of Life Assess-
1996; 5: 81–8. ment: International Perspectives. Springer-
46. Guyatt GH, Eagle DJ, Sackett B, Willan A, Verlag, Berlin, 1994.
Griffith LE, McIlroy W, Patterson CJ, Turpie I. 59. Shumaker S, Berzon RA. The International
Measuring quality of life in the frail elderly. J. Assessment of Health-Related Quality of Life:
Clin. Epidemiol. 1993; 46: 1433–44. Theory, Translation, Measurement and Analysis.
47. Bernhard J, Gusset H, Hurny C. Quality-of-life Rapid Communications of Oxford, Oxford, 1995.
assessment in cancer clinical trials: an inter- 60. Bullinger M, Anderson R, Cella D, Aaronson N.
vention by itself? Support. Care Cancer 1995; 3: Developing and evaluating cross-cultural instru-
66–71. ments from minimum requirements to optimal
48. Nelson EC, Landgraf JM, Hays RD, Wasson JH, models. Qual. Life Res. 1993; 2: 451–9.
Kirk JW. The functional status of patients. How 61. Leplege A, Verdier A. The adaptation of health
can it be measured in physicians’ offices? Med. status measures: methodological aspects of the
Care 1990; 28: 1111–26. translation procedure. In: Shumaker S, Berzon R
194 The Advanced Handbook of Methods in Evidence Based Healthcare
(eds). The International Assessment of Health- lems and common sense solutions. Eur. J. Cancer
Related Quality of Life: Theory, Translation, 1992; 28A: 1304–7.
Measurement and Analysis. Rapid Communica- 63. Erickson P, Taeuber RC, Scott J. Operational
tions of Oxford: Oxford, 1995, pp. 93–101. aspects of quality-of-life assessment: choosing
62. Aaronson NK. Assessing the quality of life of the right instrument: review article. Pharmaco-
patients in cancer clinical trials: common prob- Economics 1995; 7: 39–48.
12
The Use of Health Status Measures in
Economic Evaluation
J O H N B R A Z I E R , M A R K D E V E R I L L and C O L I N G R E E N
such as the chronic respiratory questionnaire,4 or weights. Each health state has a value from 0 to
more generic, such as the SF-36. 1, where death is equivalent to 0 and full health
The patient focus of these measures has made is 1. States of health regarded as worse than
them popular amongst health services research- death can have a value of less than 0. These
ers and clinicians in clinical research (see Chap- single index scores, sometimes referred to by
ter 11). One of the questions addressed in this economists as health state utilities, are used to
chapter is whether such measures can be used in adjust survival to calculate quality adjusted life
economic evaluation even if they have not been years (QALYs).5 The five most widely used
designed for the purpose. multi-attribute utility scales are reviewed in this
A small subset of health status measures have chapter.
been specially developed for use in economic
evaluation known as multi-attribute utility scales. Health Status Measures in Economic
They comprise a health state classification sys- Evaluation
tem with multi-level dimensions or attributes
(e.g. the EQ-5D). Health states are constructed Economic evaluation is the comparative assess-
from these classifications by selecting one level ment of the costs and benefits of alternative
from each dimension (and in this way it is healthcare interventions. The measure of benefit
possible to define 243 health states with the EQ- is the key feature that distinguishes the different
5D). A special feature of multi-attribute utility techniques of economic evaluation.5
scales is that a single index score is calculated Measures of health status are not used in cost-
for each state of health using preference effectiveness analysis, which typically employs
Health Status Measures in Economic Evaluation 197
‘natural’ units such as life years or cancer 1 How to judge the appropriateness of health
cases detected, or cost–benefit analysis, where status measures for use in economic evalu-
all the benefits of an intervention are valued in ation. There have been a number of published
monetary terms. reviews of measures of health status.2,10–12
A technique of economic evaluation which The absence of economic considerations
does use measures of health status is cost–utility from these criteria has often resulted in
analysis, where benefits are measured in terms economic measures of health-related quality
of health gain.6 The most widely used measure of life being neglected and portrayed as
is years in full health or the QALY. The number ‘invalid’ or irrelevant in the assessment of
of QALYs is calculated by multiplying the value health benefits. This is an important omis-
of each health state, which can be obtained by sion given the role of assessing efficiency in
using multi-attribute utility scales, by the length modern health services research. We have
of time spent in each health state. QALYs can developed, therefore, a check-list for judging
be used to undertake comparisons between the merits of health status measures for use
in economic evaluation by adapting the cri-
healthcare interventions in terms of cost per
teria used by psychometricians to judge the
QALY at the margin. performance of non-preference based meas-
Beyond the usual suite of economic evalu- ures of health status.
ation techniques is cost–consequences analysis, 2 A comparison of the techniques for valuing
where there is no attempt to combine multiple health states. The relative merits of these
outcomes into a single indicator of value (such techniques has been a subject for debate for
as the QALYs or £s).7 The decision-maker is left many years. There have been a number of
with the task of weighing-up the costs and the informative reviews of the techniques,6,13–15
multiple outcomes, which may include measures and our review builds upon this earlier work,
of health status such as the SF-36. There is no in particular using a review by Froberg and
formal theoretical basis for the outcome meas- Kane13 published in 1989 as a point of
ures used, and cost–consequences analysis is departure. It focuses on the following tech-
not strictly one of the techniques of econ- niques for the elicitation of preferences (or
omic evaluation but it retains the discipline of quality weights) for use in the valuation of
the framework of economic evaluation. health states: standard gamble (SG), time
trade-off (TTO), visual–analogue or rating
scales (VAS), magnitude estimation (ME)
Scope of the Reviews and person trade-off (PTO).
3 A comparison of multi-attribute utility scales.
The scope of the reviews reported in this chapter This review examines the five most widely
is limited to the use of health status measures in used multi-attribute utility scales: Quality
economic evaluation. These measures are not of Well-Being (QWB) scale,16 Rosser’s
used in CEA and CBA, and hence these tech- disability/distress classification,17 the Health
niques of economic evaluation are excluded Utility Index versions one, two and three
from this review. The review also excludes a (HUI-I, HUI-II and HUI-III),18 the EQ-5D19
number of wider issues. For example, there have and the 15D.20 These multi-attribute utility
been criticisms of using the QALY as a measure scales are reviewed against the criteria of
of benefit, since it makes a number of limit- practicality, reliability and validity using the
ing assumptions about people’s preferences for check-list developed in the first review.
4 The use of non-preference based measures in
health, including the assumption that health state
economic evaluation. Non-preference based
values are independent of how long they last, measures of health status, such as the SF-36,
when they occur and their context (e.g. whether are used in clinical trials to assess the effi-
they occur at the start or the end of a poor cacy and benefits of healthcare interventions
sequence of states).8,9 This review also excludes in terms of patient-perceived health. These
consideration of the important debate about measures may be intended to assess patient-
whose values to elicit in the valuation of health perceived health, but there are considerable
states. There are also other methods of measur- doubts as to whether the scoring systems
ing benefit, including health year equivalents, they employ reflect patient’s views. They
willingness to pay and conjoint analysis. These have been used in economic evaluations21,22
too are excluded. and are likely to continue to be so.23 This
This leaves a large area of issues and con- review examines the use and abuse of these
troversies to be tackled by this review. These measures in economic evaluation.
were identified at the start of the project and 5 The practice of using health status measures
include the following: in economic evaluations in studies published
198 The Advanced Handbook of Methods in Evidence Based Healthcare
in 1995. This final review uses the criteria from a collaboration between economists (J.B.,
developed in reviews one and four to assess M.D. and C.G.) and an expert in searching
the appropriateness of the measures of health literature databases and are fully documented in
status measures used in these published the main report.1
studies, and whether they are correctly em-
ployed in the economic evaluation.
Issues in using health status measures in
We provide here an overview of these reviews economic evaluation
of the use of health status measures in economic
This search yielded over 1300 papers, of which
evaluation. Detailed findings have been reported
154 were found to be relevant for the review.
elsewhere.1
severity levels, and see whether the scales are researchers are recommended to use when select-
able to reproduce the orderings. Some econo- ing a measure for use in economic evaluation
mists believe social values are different from (Table 2).
individually based ones and hence would want a
different test. Nord et al.31 have proposed that
‘. . . the validity of the values obtained from Techniques for Health State Valuation
different scaling techniques may be tested by
asking whether the people from whom the Visual–analogue scale (VAS)
values were elicited actually agree with the
consequences in terms of the implied priorities The VAS (sometimes referred to as the category
for different health programs’. Different per- rating scale or just rating scale) is simply a line,
spectives imply different tests of validity. usually with well-defined end-points, on which
respondents are asked to indicate the desirability
of a health state. There are many variants of the
A check-list technique. The lines can vary in length, be
We have developed a check-list of questions vertical or horizontal, and may or may not
for examining these three aspects of validity, have intervals marked out with different num-
along with reliability and practicality, which bers. An example is shown in Figure 1 (i.e. the
Health Status Measures in Economic Evaluation 201
VAS techniques do not present respondents with However, there are theoretical arguments against
a choice, they have no element of sacrifice or the preferred use of SG in health state valu-
opportunity cost, and hence have no basis in ation14 and limited empirical support for assump-
economic theory. With respect to ME, assump- tions underpinning expected utility theory;8
tions surrounding the ratio level properties of therefore we are unable to support the ‘gold
the technique remain unsupported and it is not standard’ status which some attach to the SG.
related to economic theory. The empirical literature to inform on the
PTO is a choice-based technique, but it relates extent to which the assumptions underlying the
to social choice and the opportunity cost is not different techniques correctly describe individ-
directly borne by the individual; therefore, con- ual preferences is not extensive. We find that
sumer theory is not applicable. However, the PTO and ME lack theoretical support; therefore,
PTO is intuitively appealing and its approach we do not comment further on theoretical valid-
is considered by some to be more meaningful ity in terms of empirical evidence. When con-
to decision-makers.14 The TTO technique has sidering VAS, TTO and SG, the empirical evi-
merit, in that it is a choice based technique, but dence casts doubt on the theoretical basis of all
its theoretical underpinnings are not well devel- of these techniques.
oped.15,40 Of all the techniques, SG has the most In consideration of the theoretical support
rigorous theoretical foundation. SG is supported offered for VAS,37 we find evidence to suggest
by expected utility theory which has been that a measurable value function does not under-
the dominant theory of decision-making under lie the technique, and that VAS health state
uncertainty since the 1950s. Through its asso- valuations are influenced by the severity of the
ciation with expected utility theory, SG has other health states being valued and a spreading
often been referred to as the ‘gold standard’, or effect.41–44 Empirical evidence has shown that
reference method, of health state valuation.6,41 duration and time preference have an impact on
204 The Advanced Handbook of Methods in Evidence Based Healthcare
the elicitation of TTO values.45,46 Given the marginally favours TTO. However, there is not
absence of risk in the TTO, it is possible to sufficient evidence to say that one technique is
adjust TTO-elicited values to incorporate un- more valid than another on empirical grounds.
certainty, yet this is rarely done in practice.47,48
The SG is the only technique with clear theoret-
ical foundations, but the evidence suggests that Mapping from VAS to SG or TTO
SG values can be strongly influenced by the Valuations
outcomes used in the task and by the manner in
which the task is presented.49,50 There is also There would be significant practical advantages
evidence to suggest that attitude to risk is not to being able to map from VAS to one of the
constant.8,51 Further to such findings, there is choice based techniques. Studies mapping the
evidence that respondents systematically violate relationship between VAS and the choice based
the axioms of expected utility theory in the methods have used regression analyses. Some
health context15,52 and more generally.53,54 use aggregate level data sets, where the regres-
On the basis of theoretical validity, we con- sion analyses have been undertaken on mean
clude that only choice based techniques should health state values, and others have used indi-
be used. The comparison between SG, TTO and vidual level data.
PTO depends on the perspective employed. The amount of evidence available to address
The debate surrounding SG versus TTO is un- the question was limited. There were only seven
resolved and depends on ones belief in the studies in total reporting estimated relation-
empirical support for expected utility theory or ships.35,47,63–67 At both levels of analyses the
its normative basis. relationships were not stable between studies in
terms of the form of the relationship and size of
Empirical validity model parameters. At the individual level, there
were also significant misspecification problems.
There is not a large literature reporting on the The models performed better at the group level,
empirical validity of valuation techniques. We and were able to explain a majority of the
find evidence reported of a poor to moderate variation, but existing studies were based on
correlation between VAS and the choice based small numbers in some instances, and none
valuation techniques.36,55–57 These findings, to- reported diagnostic test results for evaluating the
gether with significant evidence to suggest a appropriateness of different specifications. There
strong correlation between VAS and non- has also been no attempt to test the ability of
preference based measures of health status24 these models to predict directly obtained mean
raise further concerns over the ability of VAS SG values on an independent data set. This lack
methods to elicit strength of preference for of evidence would seem to provide an insecure
health states. Such concerns are also supported foundation on which to conclude that there is a
by qualitative data reported in the study under- unique and stable relationship between VAS and
taken by Robinson and colleagues,58 where SG or TTO.
respondents indicated that their VAS responses
were not intended to reflect their preferences.
There is empirical evidence which reports that Review of Five Multi-Attribute Utility
SG and TTO values correlate reasonably well Scales
with one another.6,59 Whilst we have found no
direct evidence to inform on the performance of The five multi-attribute utility scales differ con-
SG against stated ordered preferences, TTO siderably in terms of the characteristics of their
mean values have been found to be consistent health state classification and methods of valu-
with respondent rank ordering of health states,60 ation (Table 3). There is little guidance in the
and Robinson et al.58 present qualitative evi- literature on which multi-attribute utility scale to
dence to indicate TTO responses are intended to use and to the best of our knowledge, there has
reflect the preferences of individuals. been no systematic review undertaken of exist-
SG, TTO and VAS have demonstrated good ing scales.
levels of consistency with hypothesised prefer-
ences,15,61 whilst a lack of evidence surrounding
PTO and ME leaves the consistency of these
Practicality
methods unproven. A number of studies have There is little to choose between the ques-
reported the performance of SG and TTO against tionnaires on the basis of practicality except in
hypothesised (or expected) preferences15,50,62 so far as the QWB does not have an accepted
offering some positive insights to their empirical method of self-completion. The other four use a
validity. In terms of consistency with multi- reasonably brief self-completed questionnaire.
attribute utility scales findings, the evidence The QWB has an interview schedule involving
Health Status Measures in Economic Evaluation 205
detailed probing of the respondents taking up is described in terms of capacity (e.g. HUI) or
to 20 min.68 All multi-attribute utility scales actual behaviour and performance (e.g. QWB).
have achieved high levels of response and The coverage of symptoms, mental health and
completion. social health is variable. The QWB explicitly
excludes mental health as a separate dimension,
Reliability but has a long list of symptoms and problems.
The HUI-III covers many of the symptoms or
There is evidence of differences between the
health problems, but does not examine role or
assessment by patients of their own health com-
social function, since these are regarded as ‘out
pared to that of health professionals using the
of skin’ and not appropriate in a measure of
Rosser69 and HUI.70,71 This implies that the
individual health preferences. The EQ-5D has
method of administering these instruments must
dimensions for role and social function, and pain
be standardised. There is evidence for the test–
and mood, but not for many other symptoms and
re-test reliability in patient populations of the
health problems.
EQ-5D72,73 and 15-D,20 but this property has not
The Rosser disability and distress classifica-
been adequately investigated in any of the five
tion is inferior to the others in terms of coverage.
measures.
The choice from the remaining four will depend
on what aspects of health the potential user
Descriptive validity wishes to cover. Their relevance may therefore
The dimensions and their items were selected vary depending on the disease group and age of
from reviews of the literature on health status the patients being evaluated. The HUI measures
measures for the QWB, EQ-5D, HUIs and 15-D. (particularly HUI-II) may be better suited to a
Items for the Rosser were based on interviews younger population than the EQ-5D, for ex-
with professionals. The HUIs and 15-D en- ample, though this has not been empirically
hanced their expert approach to item generation demonstrated. There are also issues about per-
with surveys of patients. spective and whether or not social health is
The scales do not cover the same aspects of relevant.
health. They all include physical functioning, There is evidence that all measures can detect
but there are differences in whether the concept differences in group comparisons and the scores
206 The Advanced Handbook of Methods in Evidence Based Healthcare
were significantly correlated with other meas- second or subsequent severe problem. The abil-
ures of the health. However, it is difficult to ity of these two estimation procedures to predict
compare their performance owing to the absence health state values has not been compared, and
of studies using the measures on the same popu- hence there is no means of determining which is
lations. Most of the evidence on the QWB best.
scale was limited to correlations with related
health status measures,74 with very little detailed Empirical validity
scrutiny of the descriptive classification,
whereas evidence for the HUI-II was limited Evidence on empirical validity was found to be
to survivors of childhood cancer.75 There was very limited from both an individual and a social
some suggestion of insensitivity in all meas- perspective. The QWB has been shown to corre-
ures,71,72,76,77 except the 15D where there had late with direct preference elicitation from
been insufficient published studies at the time of patients,78 but such evidence has not been
the review. published for the EQ-5D and HUI-I. There is
evidence of the EQ-5D correlating with patient
perception of health change in one study,79 but
Valuation not another,72 whilst no evidence was found of
The QWB, Rosser and the 15D can be regarded the correlation of the HUIs scores with indi-
as inferior to the other two multi-attribute utility vidual stated preferences. The scores were also
scales owing to their use of VAS and ME to found to reflect hypothesised preferences
value the health descriptions. HUI-II and III between patient groups,75,80–82 but the evidence
might be preferred to the EQ-5D by those who was too limited to draw firm conclusions.
regard the SG as the ‘gold standard’. However, Nord and his colleagues31 compared the
the SG utilities for the HUIs have been derived values generated by the QWB and HUI-I, and
from VAS values using a power transformation Rosser with stated social preferences according
criticised earlier in this chapter. The valuation of to the person trade-off method. The Rosser
the HUI has also been obtained from a smaller approach was found to reflect social preference
and less representative sample of the general better than the other two, but this study was too
population than the MVH survey. crude to come to any definite conclusions.1
A further difference between them is in the
methods of estimating values for scoring the
states defined by the EQ-5D and HUIs. The The Use of Non-Preference Based
HUIs use multi-attribute utility theory which Measures of Health in Economic
substantially reduces the valuation task by Evaluation
making simplifying assumptions about the rela-
tionship between dimensions and health state Economic criticisms of non-preference
values. The simple additive functional form
assumes dimensions to be independent and
based measures
hence permits no interaction. The multiplicative The simple scoring algorithms used by most of
function permits a very limited form of inter- these measures assume equal intervals between
action between dimensions which assume the the response choices, and that the items are of
inter-dependency to be the same between all equal importance.26 The SF-36 physical func-
dimensions and for all levels of each dimension. tioning dimension, for example, assumes that
To apply this approach, respondents are asked to being limited in walking has the same import-
value the level of each dimension separately to ance as being limited in climbing flights of
estimate single attribute utility functions, and stairs, but there is no reason for this to be the
then a sample of (multi-attribute) health states. case. The theoretical concerns with simple
The overall function is calculated by solving a aggregation has been confirmed by the evidence.
system of simultaneous functions. By contrast, Studies found in the search reported only low to
the EQ-5D TTO tariffs were estimated by multi- moderate correlations between health status
variate statistical techniques. A simple additive measures than preference based measures. A
model was chosen which contains decrements small minority of measures have used weights
for each of the moderate and severe dysfunc- obtained from patients or professionals using
tional categories of the five dimensions, a con- VAS,83 but these are also unlikely to reflect car-
stant for any kind of dysfunction and the term dinal preferences given the criticisms presented
‘N3’ for whenever any of the dimensions are above.
severe. This last term was a crude way of For measures presenting results as a profile,
allowing for interactions, since it implies that there are additional problems with conflicts
the impact of a severe problem is less if it is a between the dimensions, such as in a study
Health Status Measures in Economic Evaluation 207
Table 4 Assessing the relative efficiency of two interventions given different cost
and health outcome scenarios
Scenario Cost Health status measure Can relative efficiency be evaluated?
comparing surgical with non-invasive interven- taken in the interpretation. Our review of the
tions, where surgery may be more effective at evidence found that health status measures do
reducing symptoms but is associated with a not possess the interval properties required to
range of complications. These measures also do undertake such comparisons. Furthermore, it is
not incorporate death as an outcome measure, the incremental cost-effectiveness ratio which
and hence there is no easy solution where there is important for resource allocation purposes.
is a conflict between the impact on survival and Therefore, where the least cost-effective inter-
health related quality between different inter- vention costs more and yields a higher benefit,
ventions (see Chapter 20). then the greater benefit could be worth the extra
cost.
The use and abuse of health status For multiple outcomes, the recommended
approach is to present the costs and benefits of
measures in economic evaluation the alternatives in a cost consequences analysis.7
The usefulness of health status measures in This type of presentation uses an economic
assessing the relative efficiency of interventions framework, but is unlikely to be helpful since
depends on the results of the study. In Table 4 health status measures scores have no obvious
we present seven scenarios of costs and out- intuitive meaning. Score differences cannot be
comes in a comparison of two interventions, and compared between dimensions, nor can health
consider whether it is possible to assess their status measures scores be compared to other
relative efficiency. The first scenario is a case of outcomes, such as survival, or cost. Non-
dominance where one treatment is cheaper and preference based health status measures cannot
better on at least one of the dimensions of the be used to assess the efficiency of interventions
health status measures, whilst being no worse on in such circumstances.
any other. In the second scenario it is also
straightforward to assess relative efficiency, Developing non-preference based measures
since it is simply a question of choosing the
treatment with the better health status measure
for use in economic evaluation
scores, as the two have been found to cost the There are four methods for developing non-
same. The third scenario is the same across all preference based measures of health status
dimensions of the health status measures and measures. One is to assign responses to the
hence is a cost-minimisation analysis. Even for non-preference based measures of health to the
these three scenarios it is necessary to demon- classification of a multi-attribute utility scale,
strate the ordinality of the scale of the health but this is difficult and often arbitrary in prac-
status measure scores in relation to preferences. tice.69,84 Furthermore, the predictive ability of
The result is less straightforward for scenarios any explicit mapping rules have not been tested.
4 to 7. In scenarios 4 and 5, one treatment Another is to derive exchange rates between
performs better on same dimensions but worse health status measures and preferences, but
on others, and hence one treatment could have a this does not look promising given the weak
lower cost per unit of health gain on some relationship. A more radical way forward is
dimensions but higher on others. Even where to obtain preference weights using preference
one treatment is apparently more cost-effective elicitation techniques, which has been success-
across all the health dimensions, care must be fully undertaken on the SF-36,85 but this is a
208 The Advanced Handbook of Methods in Evidence Based Healthcare
complex and expensive task and is only likely to Questions to be applied to published
be undertaken on the most widely used measure. economic evaluations
A final way is to generate health scenarios from
the data, but there were no examples of this in 1 Did the paper discuss or address the issue of
the literature. the validity and/or suitability of a health
status measures for the particular study
question?
The Use of QALY and Non-QALY 2 Was the health status measures compatible
Measures of Health Status: Assessing with the method of economic evaluation
chosen?
the State of Art in 1995 3 Given the health status measures and
economic evaluation framework used, were
The literature any conclusions presented in the paper
legitimate?
To be included in the review each paper had to
meet the following criteria:
Studies using QALY measures
1 The study had to be a ‘full’ economic evalu-
Four studies were identified using QALY meas-
ation: this implies the following:
ures and which met our inclusion criteria. Table
(a) the technology in question had to have 5 summarises the performance of these in terms
a comparator (before and after studies of the three questions outlined above. Two of
were thus excluded); the studies failed against the first criteria.86,88
(b) both costs and consequences (out- Kennedy et al.87 make no attempt to qualify their
comes) of the technology had to be use of the TTO method within their analysis,
identified and primary data gathered relying on general references relating to health
accordingly. state valuation. Mark et al.89 do not address
2 The outcomes or consequences of technolo- issues of validity or suitability with respect to
gies had to be assessed using a recognised their use of the TTO method in their calculation
QALY or non-preference based health status of QALYs.
measures. For question 2, some ambiguity is present.
Whilst the health status measures and form of
A total of 42 papers (approximately 2.5% of evaluation can be seen as compatible, the way in
the original number) was initially identified as which the health status measures are used to
economic evaluations in terms of using either obtain QALY values casts doubt on the ‘study-
QALY-type measures (hence cost–utility analy- specific compatibility’. In Kerridge et al.88 the
ses) or non-QALY based health status measures QALY values used in the analysis are deter-
(hence ‘cost–consequences’ analyses). On receipt mined through an assessed survival effect linked
of the 42 papers the inclusion criteria outlined with valuations from the Rosser Matrix. Gournay
above were once again strictly applied. It was et al.86 derived QALY values by mapping
found that 31/42 (74%) could not be included; patient characteristics onto the Rosser matrix.
the main reason for such a high exclusion rate Mark et al.89 used the TTO technique only at
was an inability at the abstract stage to apply year 1 post-treatment, they did not use it to
the inclusion criteria. Just 11 papers met our generate both pre- and post-treatment values.
strict entry criteria, and these are reported on For question 3, there is initial support for a
below. positive response in two studies,87,88 but there
Q1
Health status (Validity/ Q2 Q3
Article authors Clinical area measure Suitability) (Compatability) (Legitimacy)
BDI 5 Beck Depression Inventory; GHQ 5 General Health Questionnaire; TTO 5 time trade-off.
Health Status Measures in Economic Evaluation 209
Q1
Health status (Validity/ Q2 Q3
Article authors Clinical area measure Suitability) (Compatability) (Legitimacy)
90
Cottrell et al. Respiratory tract SIP ✖ ✓ ✓
Johnson et al.91 Limb ischaemia HADS, Barthel, FAI ✓ ✓ ✓
Knobbe et al.92 Mental illness RLI, SNLAF ✓ ✓ ✓–✖
Lawrence et al.93 Hernia repair SF-36,EQ-5D (VAS only) ✖ ✓ ✓
Prince et al.94 Tetraplegic care/ RAND-36, LSI-A, Chart ✓ ✓ ✓
rehabilitation
Chemotherapy NHP, RSC, KPI ✖ ✓ ✓
Uyl de Groot et al.95 Non-Hodgkin
lymphoma
Wimo et al.96 Mental illness GDS, IWB ✖ ✓–✖ ✓–✖
CES-D 5 Centre for Epidemiological Studies Depression Scale; CHART 5 Craig Handicap Assessment and
Reporting Technique; FAI 5 Frenchay Activities Index; GDS 5 Global Deterioration Scale; HADS 5 Hospital
Anxiety and Depression Scale; IWB 5 Index of Well Being; KPI 5 Karnofsky Performance Index; LSI 5 Life
Satisfaction Index-A; NHP 5 Nottingham Health Profile; RLI 5 Resident Lifestyle Inventory; MOS SF GHS &
RAND-36 5 Medical Outcomes Study Short Form General Health Survey; RSC 5 Rotterdam Symptom
Checklist; SIP 5 Sickness Impact Profile; SNLAF 5 Social Network Lifestyle Analysis Form.
are some problems. In Kennedy et al.87 the however, that the GDS was suitable for a cost–
results of the TTO came from a small sample consequences analysis.
(n 5 9) of clinicians. In Kerridge et al.88 expert For the third question there are problems in
opinion was used to gauge the survival effect three papers. In the paper by Johnson et al.,91
of intensive care. The conclusions from the surgery for limb-threatening ischaemia, the
remaining two studies86,89 are in more doubt. authors failed to ascertain the quality of life
Gournay et al.86 used data covering absence scores prior to the onset of limb-threatening
from work as a proxy for health status; thus ischaemia, whilst then claiming that limb sal-
they only loosely address the valuation of the vage ‘can be the most cost-effective way of
benefits attributable to the intervention being managing limb-threatening ischaemia’. This is
assessed consequently. Although Mark et al.89 doubtful given that limb salvage does not appear
present a cost per QALY ratio in their study, the to dominate the alternative (limb revascularisa-
methods employed in determining such a ratio tion) on all the health status dimensions recorded.
would not substantiate their findings. They do Knobbe et al.92 failed to state the health status
not present their conclusions in a cost per measures values found in the study, thus making
QALY framework, relying instead on ‘cost per the paper difficult to use in any form of
life year saved’ in their conclusions. decision-making. Wimo et al.96 presented dis-
aggregated cost and outcome results but wisely
(considering the mapping of GDS values) made
Studies using non-preference based no attempt at presenting a cost per QALY
measures result.
Three of the seven studies reviewed included
little detail on the validity and/or suitability of
the health status measures used (Table 6).91,92,94 RECOMMENDATIONS ON USING HEALTH
Six of the studies used the health status meas- STATUS MEASURES IN ECONOMIC
ures correctly within the cost consequences EVALUATION
analysis framework.90–95 Wimo et al.96 mapped
Index of Well Being (IWB) values from the Selection of Health Status Measure
Global Deterioration Scale (GDS) but had not
presented any cost per QALY calculations. The We recommend that researchers planning to
UK NHS Centre for Reviews and Dissemination conduct economic evaluations alongside clinical
database entry for this study commented that the trials pay careful attention to the check-list in
actual process of mapping GDS values onto the Table 2 when selecting health status measures.
IWB was of dubious merit. It can be said, The purpose of the check-list approach is to
210 The Advanced Handbook of Methods in Evidence Based Healthcare
provide guidance rather than rules, and to Comparing the performance of EQ-5D and
acknowledge disagreements where they exist. HUI-III Research is required to compare the
The precedent in health economics is the widely EQ-5D and HUI-III head-to-head in terms of the
used check-list for economic evaluation devel- criteria in the check-list in Table 2.
oped by Drummond et al.,5 which was compiled
despite areas of disagreement in the literature. Studies to re-value existing multi-attribute
The check-list we propose is likely to need utility scales There is a strong case for obtain-
updating given further theoretical development ing weights in other countries for the HUI-II and
and the accumulation of more evidence. the HUI-III if the HUI-III is found to perform
well against the EQ-5D. There could also be a
case for obtaining weights for the QWB using a
Selection of Valuation Technique choice based technique if this continues to be
widely used.
We recommend only choice based techniques,
either SG or TTO, be used to value health states. The development of new multi-attribute utility
SG and TTO values should be obtained directly, scales The EQ-5D and HUI may not be suit-
rather than estimating them from VAS values able for many patient groups, and hence another
via a mapping function. measure, perhaps based on a more sensitive
classification, may be required. This could either
be done by: (i) developing new multi-attribute
Selection of Multi-Attribute Utility utility scales de novo, such as the Australian
quality of life measure;97 or (ii) estimating pref-
Scales erence weights for an existing measure of health
status such as the SF-36,85 though this latter
We conclude that, at the present time, the best of option is only advisable for the most widely
the five multi-attribute utility scales reviewed used measures.
are the EQ-5D and HUI-II and III. The HUI-II is
recommended for children, but in the absence of The valuation of scenarios constructed from
published weights for HUI-III, the EQ-5D is health status measures data This approach
preferred for adult patients. This will have to be has the advantage of being able to focus on
re-appraised when (Canadian) weights become those aspects of health most relevant to the
available for the HUI-III. treatment being evaluated. It is also possible to
incorporate the time profile of health, with tim-
ing, duration and sequence of states. The meth-
Using Non-Preference Based Health odology for constructing the vignettes from the
Status Measures In Economic potentially large volume of health status meas-
Evaluation ures (and other) data needs to be developed.
to recommend any one measure of health (such 10. Streiner DL, Norman GR. Health Measurement
as the EQ-5D) for use in economic evaluation. Scales: a practical guide to their development
and use. Oxford: Oxford University Press,
1989.
ACKNOWLEDGEMENTS 11. McDowell I, Newell C. Measuring Health: A
Guide to rating scales and questionnaire.
We would like to thank Andrew Booth and Oxford: Oxford University Press, 1996.
Rosemary Harper for their help in the literature 12. Wilkin D, Hallam L, Doggett MA. Measures of
searches. This research project was funded by need and outcome for primary health care.
the NHS HTA programme. Oxford: Oxford Medical Press, 1992.
13. Froberg DG, Kane RL. Methodology for meas-
uring health-state preferences. II: Scaling meth-
NOTE ods. J. Clin. Epidemiol. 1989; 42: 459–71.
14. Richardson J. Cost–utility analysis – what
a. TTO and SG have been adapted for valuing should be measured. Soc. Sci. Med. 1994; 39:
health states regarded as worse than death. For TTO 7–21.
alternative 1 involves dying immediately. Alternative 15. Dolan P, Gudex C, Kind P, Williams A. Valuing
2 involves x years in the health states regarded as health states: a comparison of methods. J.
worse than death followed by (t 2 x) years in perfect Health Econ. 1996; 15: 209–31.
health. Again, duration x is varied until the respond- 16. Kaplan RM, Anderson JP. A general health
ent is indifferent between the two alternatives. The policy model: update and application. Health
formula for calculating the health state value becomes Services Research 1988; 23: 203–35.
2X/(t 2 x). 17. Kind P, Rosser R, Williams A. Valuation of
Quality of Life: Some Psychometric Evidence.
In: Jones-Lee MW (ed.). The Value of Life and
REFERENCES Safety. North Holland, 1982.
18. Torrance GW, Furlong W, Feeny D, Boyle
1. Brazier JE, Deverill M, Green C, Harper R, M. Multi-attribute preference functions. Health
Booth A. A review of the use of health status Utilities Index. PharmacoEconomics 1995; 7:
measures in economic evaluation. NHS Health 503–20.
Technology Assessment 1999; 3(9). 19. Dolan P, Gudex C, Kind P, Williams A. A social
2. Bowling A. Measuring health: a review of tariff for Euroqol: Results from a UK general
quality of life and measurement scales. Milton population survey, Centre for Health Economics
Keynes: Open University Press, 1991. Discussion Paper 138, University of York,
3. Ware JE, Snow KK, Kolinski M, Gandeck B. 1995.
SF-36 Health Survey manual and interpretation 20. Sintonen H. The 15D measure of HRQoL: reli-
guide. Boston: The Health Institute, New England ability, validity, and the sensitivity of it’s health
Medical Centre, Boston, MA, 1993. state descriptive system. NCFPE Working
4. Guyatt GH, Thompson PJ, Bernam LB, et al. paper 41, Monash University/The University of
How should we measure function in patients Melbourne, 1994.
with chronic lung disease? J. Chron. Dis. 1985; 21. Buxton M, Acheson R, et al. Costs and Benefits
38: 517–24. of the Heart Transplant programmes at Hare-
5. Drummond MF, Stoddart GL, Torrance GW.
field and Papworth hospitals, DHSS Office of
Methods for the economic evaluation of health
the Chief Scientist Research Report No. 12,
care programmes. Oxford: Oxford Medical Pub-
London: HMSO, 1985.
lications, 1987.
22. Nicholl J, Brazier JE, Milner PC, et al. Random-
6. Torrance GW. Measurement of health state
utilities for economic appraisal: a review. J. ised controlled trial of cost-effectiveness of
Health Econ. 1986; 5: 1–30. lithotripsy and open cholecystectomy as treat-
7. Drummond MF. Economic analysis alongside ments for gallbladder stones. Lancet 1992; 340:
controlled trials. London: Department of Health, 801–7.
1994. 23. Drummond MF, Davies L. Economic Analysis
8. Loomes G, McKenzie L. The use of QALYs in alongside clinical trials: Revisiting the methodo-
health care decision making. Soc. Sci. Med. logical issues. Int. J. Technol. Assess. Health
1989; 28: 299–308. Care 1991; 7: 561–73.
9. Gafni A, Birch S. Preferences for outcomes in 24. Revicki DA, Kaplan RM. Relationship between
economic evaluation: an economic approach to psychometric and utility-based approaches to the
addressing economic problems. Soc. Sci. Med. measurement of health-related quality of life.
1995; 40: 767–76. Quality in Life Research 1993; 2: 477–87.
212 The Advanced Handbook of Methods in Evidence Based Healthcare
25. Williams A. Measuring functioning and well- 41. Gafni A. The standard gamble method: what is
being, by Stewart and Ware. Review article. being measured and how it is interpreted. Health
Health Economics 1992; 1: 255–8. Serv. Res. 1994; 29: 207–24.
26. Culyer AJ. Measuring Health: Lessons for 42. Parducci A, Carterette E, Friedman M (eds).
Ontario. Toronto, University of Toronto Press, Handbook of Perception, Vol II. Contextual
1978. effects. A range-frequency analysis. New York:
27. Arrow KJ. Uncertainty and the Welfare Econ- Academic Press, 1974.
omics of Medical Care. American Economic 43. Nord E. The validity of a visual analogue scale
Review 1963; 53: 941–73. in determining social utility weights for health
28. Culyer AJ. The nature of the commodity health states. Int. J. Health Policy Management 1991;
care and its efficient allocation. Oxford Econ- 6: 234–42.
omic Papers 1971; 24: 189–211. 44. Kuppermann M, Shiboski S, Feeny D, Elkin EP,
29. Donaldson C, Gerard K. Economics of Health Washington AE. Can preference scores for dis-
Care Financing: the Visible Hand. Macmillan, crete states be used to derive preference scores
London, 1993. for an entire path of events? An application to
30. Williams A. The role of the Euroqol instrument prenatal diagnosis. Med. Decision Making 1997;
in QALY calculations. Discussion paper 130. 17: 42–55.
Centre for Health Economics, University of 45. Sutherland HJ, Llewellyn TH, Boyd D, Till JE.
York, 1995. Attitudes towards quality of survival: The con-
31. Nord E, Richardson J, Macarounds-Kichnamm cept of maximum endurable time. Med. Decision
K. Social evaluation of health care versus per- Making 1982; 2: 299–309.
sonal evaluation of health states: evidence on the 46. Dolan P, Gudex C. Time preference, duration
validity of four health-state instruments using and health state valuations. Health Econ. 1995;
4: 289–99.
Norwegian and Australian surveys. Int. J. Tech-
47. Stiggelbout AM, Kiebert GM, Kievit J, Leer
nol. Assess. Health Care 1993; 9: 463–78.
JW, Stoter G, de Haes JC. Utility assessment in
32. Euroqol group Euroqol – a new facility for the
cancer patients: adjustment of time tradeoff
measurement of health-related quality-of-life.
scores for the utility of life years and compar-
Health Policy 1990; 16: 199–208.
ison with standard gamble scores. Med. Decision
33. von Neumann J, Morgenstern O. Theory of
Making 1994; 14: 82–90.
Games and Economic Behaviour. Princeton:
48. Cher DJ, Miyamoto J, Lenert LA. Incorporat-
Princeton University Press, 1944.
ing risk attitude into Markov-process decision
34. Torrance GW, Thomas WH, Sackett DL. A util-
models: Importance for individual decision mak-
ity maximisation model for evaluation of health
ing. Med. Decision Making 1997; 17: 340–50.
care programmes. Health Services Research 49. Llewellyn Thomas H, Sutherland HJ, Tibshirani
1972; 7: 118–33. R, Ciampi A, Till JE, Boyd NF. The measure-
35. Reed WW, Herbers JE, Jr, Noel GL. ment of patients’ values in medicine. Med.
Cholesterol-lowering therapy: what patients Decision Making 1982; 2: 449–62.
expect in return. J. Gen. Intern. Med. 1993; 8: 50. Gage BF, Cardinalli AB, Owens DK. The effect
591–6. of stroke and stroke prophylaxis with aspirin or
36. van der Donk J, Levendag PC, Kuijpers AJ, warfarin on quality of life. Arch. Intern. Med.
Roest FH, Habbema JD, Meeuwis CA, Schmitz 1996; 156: 1829–36.
PI. Patient participation in clinical decision- 51. Kahneman D, Tversky A. Prospect theory: an
making for treatment of T3 laryngeal cancer: a analysis of decision under risk. Econometrica
comparison of state and process utilities. J. Clin. 1979; 47: 263–91.
Oncol. 1995; 13: 2369–78. 52. Read JL, Quinn RJ, Berwick DM, Fineberg HV,
37. Dyer JS, Sarin RK. Relative risk aversion. Man- Weinstein MC. Preferences for health out-
agement Science 1982; 28: 875–86. comes. Comparison of assessment methods.
38. Bleichrodt H, Johannesson M. An experimental Med. Decision Making 1984; 4: 315–29.
test of a theoretical foundation for rating-scale 53. Hershey JC, Kunrather HG, Schoemaker PJH.
valuations. Med. Decis. Making 1997; 17: Sources of bias in assessment procedures for
208–16. utility functions. Management Science 1981; 28:
39. Loomes G, Jones M, Lee M, Robinson A. What 936–54.
do visual analogue scales actually mean? Paper 54. Schoemaker PJH. The expected utility model: its
presented to HESG Conference, Newcastle, variants, purposes, evidence and limitations. J.
1994. Econ. Lit. 1982; 529–63.
40. Johannesson M. QALYs, HYEs and individual 55. Rutten van Molken MP, Bakker CH, van
preferences – a graphical illustration. Soc. Sci. Doorslaer EK, van der Linden S. Methodo-
Med. 1994; 39: 1623–32. logical issues of patient utility measurement.
Health Status Measures in Economic Evaluation 213
Experience from two clinical trials. Med. Care 69. Bryan S, Parkin D, Donaldson C. Chiropody and
1995; 33: 922–37. the QALY – a case-study in assigning categories
56. Clarke AE, Goldstein MK, Michelson D, Garber of disability and distress to patients. Health
AM, Lenert LA. The effect of assessment Policy 1991; 18: 169–85.
method and respondent population on utilities 70. Feeny D, Leiper A, Barr RD, Furlong W,
elicited for Gaucher disease. Qual. Life Res. Torrance GW, Rosenbaum P, Weitzman S. The
1997; 6: 169–84. comprehensive assessment of health status in
57. Zug KA, Littenberg B, Baughman RD, et al. survivors of childhood cancer: application to
Assessing the preferences of patients with high-risk acute lymphoblastic leukaemia. Br. J.
psoriasis. A quantitative, utility approach. Arch. Cancer 1993; 67: 1047–52.
Dermatol. 1995; 131: 561–8. 71. Barr RD, Pai MKR, Weitzman S, Feeny D,
58. Robinson A, Dolan P, Williams A. Valuing Furlong W, Rosenbaum P, Torrance GW. A
health states using VAS and TTO: what lies multi-attribute approach to health status meas-
behind the numbers? Soc. Sci. Med. 1997; 45:
urement and clinical management – illustrated
1289–97.
by an application to brain tumors in childhood.
59. Dolan P, Gudex C, Kind P, Williams A. The
Int. J. Oncol. 1994; 4: 639–48.
time trade-off method: results from a general
72. Harper R, Brazier JE,Waterhouse JC, Walters S,
population study. Health Econ. 1996; 5: 141–54.
60. Ashby J, O’Hanlon M, Buxton MJ. The time Jones N, Howard P. A comparison of outcome
trade-off technique: how do the valuations of measures for patients with chronic obstructive
breast cancer patients compare to those of other pulmonary disease in an outpatient setting.
groups? Qual. Life Res. 1994; 3: 257–65. Thorax 1997; 52: 879–87.
61. Gudex C, Dolan P, Kind P, Williams A. Health 73. Hurst NP, Jobanputra P, Hunter M, Lambert M,
state valuations from the general public using Lochhead A, Brown H. Validity of Euroqol – a
the visual analogue scale. Qual. Life Res. 1996; generic health status instrument – in patients
5: 521–31. with rheumatoid arthritis. Economic and Health
62. Churchill DN, Torrance GW, Taylor DW, Outcomes Research Group. Br. J. Rheumatol
Barnes CC, Ludwin D, Shimizu A, Smith EK. 1994; 33: 655–62.
Measurement of quality of life in end-stage renal 74. Kaplan RM, Anderson JP, Patterson TL, et al.
disease: the time trade-off approach. Clin. Validity of the Quality of Well-Being Scale for
Invest. Med. 1987; 10: 14–20. persons with human immunodeficiency virus
63. Torrance GW, Zhang Y, Feeny D, Furlong W, infection. HNRC Group. HIV Neurobehavioral
Barr R. Multi-attribute preference functions for Research Center. Psychosom. Med. 1995; 57:
a comprehensive health status classification 138–47.
system. Hamilton, Ontario: Centre for Health 75. Feeny D, Furlong W, Boyle M, Torrance
Economics and Policy Analysis, McMaster Uni- GW. Multi-attribute health status classification
versity, 1992. systems. Health Utilities Index. Pharmaco-
64. Bombardier C, Wolfson AD, Sinclair AJ, et al. Economics 1995; 7: 490–502.
Choices in Health Care: Decision Making and 76. Hollingworth W, Mackenzie R, Todd CJ, Dixon
Evaluation of Effectiveness. Comparison of three AK. Measuring changes in quality-of-life fol-
measurement technologies in the evaluation of a lowing magnetic-resonance-imaging of the knee
functional status index. Toronto: University of – SF-36, Euroqol or Rosser index. Qual. Life
Toronto, 1982.
Res. 1995; 4: 325–34.
65. Torrance GW, Feeny DH, Furlong WJ, Barr RD,
77. Liang MH, Fossel AH, Larson MG. Com-
Zhang Y, Wang Q. Multiattribute utility func-
parisons of five health status instruments for
tion for a comprehensive health status classifica-
orthopaedic evaluation. Medical Care 1990; 28:
tion system. Health Utilities Index Mark 2. Med.
632–42.
Care 1996; 34: 702–22.
66. Dolan P, Sutton M. Mapping visual analogue 78. Fryback DG, Dasbach EJ, Klein R, Klein BE,
scale health state valuations onto standard gamble Dorn N, Peterson K, Martin PA. The Beaver
and time trade-off values. Soc. Sci. Med. 1997; Dam Health Outcomes Study: initial catalog
44: 1519–30. of health-state quality factors. Med. Decision
67. Torrance GW. Social preferences for health Making 1993; 13: 89–102.
states: an empirical evaluation of three measure- 79. Hurst NP, Kind P, Ruta D, Hunter M, Stubbings
ment techniques. Socio-Econ. Plan. Sci. 1976; A. Measuring health-related quality of life in
10: 129–36. rheumatoid arthritis: validity, responsiveness and
68. Bombardier C, Ware J, Russell I, Larson MG, reliability of EuroQol (EQ-5D). Br. J. Rheumatol.
Chalmers A, Leighton Read J. Auranofin ther- 1997; 36: 551–9.
apy and quality of life in patients with rheuma- 80. Kallis P, Unsworth White J, Munsch C, et al.
toid arthritis. Am. J. Med. 1986; 81: 565–78. Disability and distress following cardiac surgery
214 The Advanced Handbook of Methods in Evidence Based Healthcare
in patients over 70 years of age. Eur. J. Cardio- 92. Knobbe CA, Carey SP, Rhodes L, Horner RH.
thorac. Surg. 1993; 7: 306–11. Benefit-cost analysis of community residential
81. Rissanen P, Aro S, Slatis P, Sintonen H, versus institutional services for adults with severe
Paavolainen P. Health and quality of life before mental retardation and challenging behaviors.
and after hip or knee arthroplasty. J. Arthro- Am. J. Ment. Retard. 1995; 99: 533–41.
plasty 1995; 10: 169–75. 93. Lawrence K, McWhinnie D, Goodwin A, Doll
82. Essink-Bot ML, Vanroyen L, Krabbe P, Bonsel H, Gordon A, Gray A, Britton J, Collin J.
GJ, Rutten FFH. The impact of migraine on Randomised controlled trial of laparoscopic ver-
health-status. Headache 1995; 35: 200–6. sus open repair of inguinal hernia: early results.
83. Jones PW. Quality of life measurement for Br. Med. J. 1995; 311: 981–5.
patients with diseases of the airways. Thorax 94. Prince JM, Manley MS, Whiteneck GG.
1991; 46: 676–89. Self-managed versus agency-provided personal
84. Coast J. Reprocessing data to form QALYs. Br. assistance care for individuals with high level
Med. J. 1992; 305: 87–90. tetraplegia. Arch. Phys. Med. Rehab. 1995; 76:
85. Brazier JE, Usherwood TP, Harper R, Jones 919–23.
NMB, Thomas K. Deriving a preference based 95. Uyl de Groot CA, Hagenbeek A, Verdonck LF,
single index measure for health from the SF-36. Lowenberg B, Rutten FFH. Cost-effectiveness
J. Clinical Epidemiol. 1998; 51: 1115–28. of abmt in comparison with chop chemotherapy
86. Gournay K, Brooking J. The community psy- in patients with intermediate- and high-grade
chiatric nurse in primary care: an economic analy- malignant non-Hodgkin’s lymphoma (nhl). Bone
sis. J. Advanced Nursing 1995; 22: 769–78. Marrow Transplantation 1995; 16: 463–70.
87. Kennedy W, Reinharz D, Tessier G, et al. Cost 96. Wimo A, Mattson B, Krakau I, Eriksson T,
utility of chemotherapy and best supportive Nelvig A, Karlsson G. Cost-utility analysis of
care in non-small cell lung cancer. Pharmaco- group living in dementia care. Int. J. Technol.
Economics 1995; 8: 316–23. Assess. Health Care 1995; 11: 49–65.
88. Kerridge RK, Glasziou PP, Hillman KM. The 97. Hawthorne G, Richardson J, Osborne R, McNeil
use of ‘quality-adjusted life years’ (qalys) to H. The Australian Quality of Life (AQoL) Instru-
evaluate treatment in intensive care. Anaesth. ment. Monash University Working Paper 66,
Intensive Care 1995; 23: 322–31. 1997.
89. Mark DB, Hlatky MA, Califf RM, et al. Cost 98. Ministry of Health (Ontario) Ontario guidelines
effectiveness of thrombolytic therapy with tissue for the economic evaluation of pharmaceutical
plasminogen activator as compared with strepto- products. Toronto: Ministry of Health, 1994.
kinase for acute myocardial infarction. N. Engl. 99. Commonwealth Department of Health, Housing
J. Med. 1995; 332: 1418–24. and Community Service Guidelines for the
90. Cottrell JJ, Openbrier D, Lave JR, Paul C, pharmaceutical industry on the submission to
Garland JL. Home oxygen therapy: a com- the pharmaceutical benefits advisory committee.
parison of 2- vs. 6-month patient reevaluation. Canberra: Australian Government Publishing
Chest 1995; 107: 358–61. Service, 1992.
91. Johnson BF, Evans L, Drury R, Datta D, Morris- 100. Department of Health and the Association of
Jones W, Beard JD. Surgery for limb threaten- the British Pharmaceutical Industry Guidelines
ing ischaemia: a reappraisal of the costs and for the economic evaluation of pharmaceuticals
benefits. Eur. J. Vasc. Endovasc. Surg. 1995; 9: [Press release]. London: Department of Health,
181–8. 1994.
13
Collecting Resource Use Data for
Costing in Clinical Trials
METHODS
CONTEXT
This chapter is based on a systematic review of
The relatively recent growth of serious interest the methodological issues to be addressed when
in formal economic evaluation of health tech- assessing the costs of healthcare technologies.1
216 The Advanced Handbook of Methods in Evidence Based Healthcare
The literature was reviewed with the aim of papers discussing all the issues. The number of
identifying methodological principles and issues papers retrieved at each stage is reported in
of method relating to the collection and analysis Table 1.
of cost data within the context of clinical trials. From this literature, this chapter identifies and
Although the focus of the review was on costing comments on the various issues that commonly
alongside clinical trials, the methodological arise in designing studies such as the types of
issues were derived from five strands of relevant cost for inclusion; determining the sampling
literature. Firstly, reviews of the specific topic approach and sample size; how to measure
(costing alongside clinical trials) were included, resource use data; how to analyse cost data; and
as well as review papers addressing single meth- how to present results.
odological issues. Secondly, empirical papers
conducting economic evaluations alongside clin-
ical trials were included as they might raise
additional methodological issues. Thirdly, the
NATURE OF EVIDENCE
literature on costing methods in economic evalu-
ation in other contexts was included, since this The evidence can be classified into three broad
forms the basis for the design of many economic categories:
evaluations. Fourthly, guidelines recently pro-
duced for various countries which attempt to 1 Statements of method for conducting econ-
standardise methodologies employed for econ- omic evaluation both in general, and in
omic evaluation were also included, since they relation to clinical trials.2–11 This category
make recommendations about methods. Finally, also includes context-specific guidelines.12–15
guidelines for authors intending to publish econ- All these works were important as state-
omic evaluations in peer-reviewed journals were ments of agreed principle and as highlight-
also included in the review since, by implica- ing points of current disagreement, but did
tion, they comment on methodological issues. not readily provide practical advice to those
The initial stages of the review consisted of: a designing clinical trials to provide economic
definition of inclusion criteria; a search of in- information.
house bibliographic data bases; and a manual 2 Published studies on economic evaluations
search of key journals. As a result of this the alongside clinical trials, of which there are
the search strategy and inclusion criteria were a growing number.16–24 These studies often
refined. This was then followed by elec- adopted, developed and justified a particular
tronic searches; issue-specific searches; citation strategy relating to these methodological
searches of key articles; and searches of refer- issues.
ence lists of the identified literature. Finally, 3 Experimental studies testing the relative per-
experts commenting on the review were asked formance of different methods. Overall, the
to identify any relevant literature that they felt review found few examples of this type. It
were missing. The aim was not to record the appears from the review that the need to
frequencies with which issues were raised, but improve our understanding of, for example,
to be comprehensive in identifying the methodo- the relative merits of alternative methods of
logical issues. The final literature included data collection, have largely been ignored in
therefore covers all the issues rather than all the literature. Thus, the relative merits of
many of the approaches to handling meth-
odological issues remain untested.
Table 1 Number of papers retrieved at each
stage
Number Number METHODOLOGICAL ISSUES
of papers of papers
Stage identified included Study Design
In-house search 174 24
Manual search n/a 81 Types of cost
Main electronic search 397 76 Cost is the product of two elements, the quantity
Issue searches (all) 143 18 of resources consumed, and the unit cost of
Citation searches (all) 161 16 the resources. Thus when estimating costs, the
Reference lists n/a 24 correct approach is to identify and measure
Experts n/a 14 separately these two elements. Traditionally,
Total 253 costs were classified into direct and indirect but
recently, partly as a result of the confusion
n/a, not applicable. regarding the interpretation of indirect costs,
Collecting Resource Use Data for Costing in Clinical Trials 217
Table 2 Types of cost events for which the timing (and/or frequency)
is set by the trial protocol and hence does not
Health service costs vary between the arms of the trial, but in a real
Direct costs of the whole intervention world situation may be required to vary. Trial
General illness costs costs should only be ignored if the resource use
Trial costs they represent does not affect outcome. Trial
Future costs costs can be addressed, for example, by collect-
Non-health service costs ing additional information on the number of
Costs incurred by other public sector budgets tests or visits carried out in usual practice and
Informal care costs re-analysing the costs accordingly.
Patients’ travel costs Future health service costs are the additional
Other out-of-pocket expenses incurred by the patient costs of treatment for diseases occurring either
Patients’ time costs incurred in receiving treatment in life years gained because of the initial inter-
Productivity costs vention or in years of life that would have been
Future costs (food and shelter costs) lived anyway.4,28 These costs can be further
classified by whether they are related or un-
Non-resource costs related diseases or treatments.4,28 Future costs in
Transfer payments years of life lived anyway are similar to general
illness costs introduced above, except that they
there has been a move away from this dichot- occur in the future periods rather than current
omous classification.2,4 A more useful approach periods. Future costs in years of life gained
is to classify costs into three broad categories: occur commonly in prevention programmes
health service costs; non-health service costs; where costs arise in the years gained from the
and non-resource costs and within each category prevention programme. It is usual to include all
to include a more detailed categorisation of the future effects of the prevention programme
costs (Table 2). in the economic evaluation and therefore, for
consistency, it could be argued that future costs
Health service costs These include the direct should be included.2 There is no consensus in
costs of the whole intervention such as hospital the literature, however, as to whether all types of
resources, for example, the number of treat- future costs should be included, and there is
ments, bed days, as well as out-patient, GP and particular disagreement about whether future
nurse visits. They include resource use asso- costs – particularly in life years gained – should
ciated with the main intervention, as well as any be included. Given the disagreement concerning
resource use arising from treatment of complica- how to handle future health service costs, ana-
tions or side-effects. They also include the use lysts should consider the impact on the results of
of buildings, other capital and equipment, and the inclusion or exclusion of future costs
overheads, such as heating and lighting, arising through methods such as sensitivity analysis.4
from the health service intervention.2,6 The term An applied example of the different types of
‘whole intervention’ is used to stress the fact health service cost is presented in Table 3 and
that the costs of the intervention should include illustrates their complex nature.
the broader health service costs.
General illness costs are the costs of being Non-health service costs These include the
treated for other illnesses whilst being treated costs incurred by other public sector budgets,
for the intervention in question. These may be such as social service costs.5 The sector incur-
either illnesses arising from the intervention or ring costs may change as budgetary arrange-
may be illnesses unrelated to the intervention.2,4 ments change. If it is unclear which budget costs
Whether a cost is attributable to an intervention fall under, the implication is that cost shifting
is discussed below. General illness costs occur may be possible; that is, costs incurred by one
during the trial period rather than in the future. sector may be shifted to another. This is a strong
A particular issue concerns trial costs which, argument for including these costs in a study.
as well as referring to the costs of doing the Informal care costs are the costs incurred by
research, are the costs of procedures in the trial family or friends in caring for patients, usually
protocol that are required solely for the purposes unpaid. These costs include the financial outlays
of the trial.25–27 These types of event may arise incurred by the carer, but also the time spent by
because patients are more closely monitored in the informal carer in providing care. Studies
the trial, or because of the necessity to preserve attempting to include these costs have used
blinding.9 The more pragmatic the trial design, inconsistent methodologies.29 Again, it is import-
the more the resource use estimates are likely to ant to include this type of cost if an intervention
reflect actual practice. A related type of cost is has the potential to shift the burden of costs to
protocol-driven costs, which usually refer to informal carers.
218 The Advanced Handbook of Methods in Evidence Based Healthcare
Patients may incur non-health service costs issues surrounding productivity costs is an area
including travel costs, other out-of-pocket of costing where there is little consensus.
expenses, and the time costs they incur when A final type of non-health service cost is
receiving treatment or attending for screen- future costs incurred, for example food and
ing.30,31 The opportunity cost of patients’ time shelter costs, as a result of additional life years
associated with receiving treatment is a further added because of the intervention.4 As with
type of non-health service cost. These time costs productivity costs discussed above, future
have often, confusingly, been referred to as non-health service costs may be excluded if
‘indirect’ costs.3 If a societal perspective is the measure of effect is restricted to health
adopted, time costs incurred by individuals in outcomes.
receiving treatment reflect the loss of production
to society. Unlike the productivity costs asso- Non-resource costs These include transfer
ciated with morbidity discussed below, there is payments which are flows of money from one
agreement that patient’s time costs associated group in society to another, such as social
with treatment should be included in the numer- security benefits. They are a loss to the payer
ator (costs) of a cost–effectiveness ratio. and a gain to the recipient, and since they
Productivity costs may also be incurred by involve no resource consumption they are
society, and these include the costs (in terms of usually excluded.2,4
time and lost production) associated with the
patient taking time off work. Productivity costs Factors influencing the types of cost
may be separated into three phases: treatment
(as just discussed); morbidity (incurred as a included
result of patients being ill); and mortality (incur- For each type of cost, a decision has to be made
red as a result of death).4 There are several as to whether its inclusion is appropriate and this
methodological issues surrounding productivity should be based on the factors listed in Table 4.
costs, including the scope of their measurement; Feasibility of measurement should not be a
whether they should be included as a cost or an criterion for deciding whether costs should be
effect; and how they should be valued. Several included,4 although it does have implications for
authors have rehearsed these issues.32–36 The the methods of resource use measurement.
Collecting Resource Use Data for Costing in Clinical Trials 219
especially if a disproportionate amount of and a unit cost attached. For example, the num-
resource consumption is made near death.47 This ber of hospital days could be measured and
is particularly relevant in advanced cancer clin- valued by the unit cost of a hospital day. Micro
ical trials where, generally, high costs are incur- costing is where resource use is identified at a
red during the terminal phase of the illness.11 detailed level and a unit cost attached to each
Rather than collecting data over a longer period, resource. For example, staff time administering
an alternative approach is to model results a drug could be measured and valued by a staff
beyond the end-point of the trial.48 cost per hour.
can be performed.11,57 There are, however, limita- them with published unit cost estimates for
tions of using pilot studies to determine sample non-key resources. The uncertainty in unit cost
sizes such as their limited length of follow-up, estimates can be addressed using sensitivity
small sample sizes and low power, and these analysis.
need to be recognised and understood.60 Secondly, a commonly cited limitation of
Beyond sample size, there are additional clinical trials is their low external validity.
sampling issues. One is that it may not be Although this usually refers to the distinction
necessary to collect all resource use data from between efficacy and effectiveness, it applies
all patients and that some resource use data equally to cost data. Factors limiting external
could be collected from a subsample of patients validity include the fact that the clinical out-
only. Another is that, in multi-centre studies, the comes realised in a non-trial setting may be
selection of centres is important. Differences different due to different locations, differences
among centres (heterogeneity) may exist and be between the trial protocol and routine practice
considerable.11,58 Heterogeneity in resource use and atypical patient compliance within clinical
arises because of different practices and policies, trials.9 Additional data are therefore required to
whilst heterogeneity in unit cost data is a result generalise results. Information can be collected
of different cost structures at centres.11 The on normal clinical practice at the trial site so that
choice of centres for inclusion are often deter- the user of the study can translate results to their
mined by the investigators on the basis of where own circumstances.27
data are easiest to collect, and the representa- Thirdly, clinical trials are usually conducted
tiveness of the sample. Disagreement exists as to for a limited period of time, yet the real interest
the extent to which unit cost data need be may be in examining the costs and benefits over
collected from all centres or countries participat- a longer period.52 Consequently, some form of
ing in a multi-centre trial.18,61 The use of centre- modelling may be required to extrapolate beyond
specific unit cost data may conceal differences the end-point of the clinical trial and to adjust or
in resource use; use of a single set of unit cost supplement the original data.2,48 There are sev-
estimates may not reflect the true range in unit eral methods available to address this, including
costs. The former approach is preferred but this Markov models and survival analysis.48
is an issue that can be addressed with sensitivity
analysis, and is discussed below. Analysis plan
An analysis plan for costing is useful in finalising
Data required from outside the trial study design. Analysis plans should include
Data from outside the trial are usually required details of the descriptive and multi-variate ana-
to perform the economic evaluation.62 Firstly, lyses planned and the modelling approach, if
unit cost estimates are required to value resource any, to be adopted.64 They should also explain
use and special unit costing studies are usually how resource use and unit cost data are to be
required, particularly in order to match the unit synthesised and how missing data are to be
cost data with the resource use quantities. The handled.57 It is not always possible to pre-
quality of the unit cost data is important, and determine the precise form of economic analysis
biases in unit cost measurement exist including to be conducted in the analysis plan, since the
centre selection bias, methods bias and case- final form of analysis is conditional on whether
mix bias.61 In practice, unit costs of healthcare the clinical effect is the same, better or worse.2,65
resources can be estimated from, for example, The criteria used to choose between the forms of
data from local hospital finance departments. economic evaluation should, however, be stated
For some resource use, published unit cost data in the analysis plan.
is available. The sources of unit cost data in the
UK differ from other countries, in particular the Data Collection
USA, where charge data are readily available
and valuation issues focus on the relationship
between costs and charges. In the UK, most
Patient-specific resource use from patients
hospitals have extra contractual referral (ECR) Patient-specific resource use data can be col-
tariffs which might be seen as corresponding to lected from patients by one of the following
US charges, but do not always reflect the costs methods: interviews; questionnaires; case record
of the bulk of patients covered under large forms; diary cards; or standardised instruments.
contracts. There is also a growing database Interviews are usually face to face, but can also
of costs per HRG (healthcare resource group) be carried out by telephone.66 Questionnaires are
in the UK.63 The preferred approach to unit administered by post or given to patients to
cost estimation is to perform specific unit cost complete when they attend their appointments.
studies for the key resources, supplementing Postal questionnaires are a cheaper method data
222 The Advanced Handbook of Methods in Evidence Based Healthcare
collection compared to interviews, but may In selecting a method of data collection from
result in a low response rate.67 Patient’s travel patients, potential sources of bias to be ad-
and time costs are often collected by ques- dressed are non-response bias, evasive answer
tionnaire,30,31 but there is no standardised, vali- bias, selection bias, question format and recall
dated questionnaire available to use to collect bias. Bias may be in either direction. There is no
these costs. Case record forms are often used in clear evidence regarding the appropriate recall
clinical trials to collect clinical data, and can be interval, and recall has been shown to depend on
adapted to include both resource use and quality the level of utilisation of resources, the fre-
of life questions. Adding questions into case quency of use, type of resource and severity of
record forms may, however, lead to lower over- the illness.72,73
all completion rates as a result of increased
burden. Patient-specific resource use from existing
Diary cards can also be used to collect
patient-specific resource use data. For example,
records and administrative databases
patients are asked to record resource use in a It is possible to derive patient-specific resource
diary at home when care is received, or returned use information from routine health service
on a daily or weekly basis.64 Two important records, such as patient notes, hospital records,
considerations when using diary cards are GP records or laboratory records (see Chapter
whether patients fill them in at the time of 9). Computer linkage systems are a further type
receiving care, or whether they wait to fill them of secondary dataset, where available.
in immediately prior to the return date. These Health records include clinical records, and
design aspects will affect the validity of the patient notes. If detailed resource use informa-
method. A disadvantage of using diary cards tion is required, records have been suggested as
is that patients may fail to complete them, a better source of data than burdening the patient
although this can be overcome through reminder with detailed data collection.64 Even when
telephone calls.64 The costs of data collection by records exist, however, they are often incom-
diary card increases if researchers are required plete and difficult to access or retrieve.53 An
to motivate patients to complete diary cards and important practical issue is how easy existing
to monitor diary completion. records will be to access or retrieve. This may
Investigators often design a new data collec- also include issues regarding patient consenting
tion instrument for each study, as noted above for investigators to access their records. As with
for questionnaires determining travel and time data collection methods directly from patients,
costs. This results in wide variation in both the the type of resource use affects the validity of
information collected and its accuracy. These the method.74,75
instruments or adaptations are rarely tested for Administrative databases can be used to
validity and reliability. This increases flexibility derive resource use estimates.76 In the USA,
but results in wide variations in the information these administrative databases relate to clinical
collected and limits comparability with other and claims databases which are less common in
studies. In order to increase the comparability of the UK. The advantages of administrative data-
resource use estimates, some researchers have bases are that they are inexpensive to use com-
developed standard instruments for collecting pared to primary data collection.58 There are
resource use data. An American modular instru- several factors to be addressed when considering
ment, the Resource Utilisation Survey (RUS), using administrative databases to derive patient-
has been developed for use in clinical trials.68 In specific resource use. Firstly, missing data are
the UK, the Client Service Receipt Interview likely to be a problem since databases may not
(CSRI) has been developed to gather retrospec- have been designed to record all relevant infor-
tive information on health and social services mation for research purposes. Secondly, extrac-
resource use within a community/mental health tion of health record data is also subject to
context.69 If service provision has changed since problems relating to level of coder training and
the instrument was developed, however, stan- amount of interpretation necessary.77 Thirdly,
dardised instruments are unable to detect any measurement error occurs since there is a risk
changes in resource use associated with the that the data in such a database could be an
change in service delivery.53 artefact of the way the data were collected.78
The above data collection methods could be Accuracy of the data is also related to the
completed by those other than the patient, such intended purpose of the database.79 Differences
as staff at the hospital.70,71 If this approach is between patients may reflect measurement error
adopted, then guidelines on how to complete the rather than any real differences. Finally, data-
forms should be provided in order to maximise bases become dated and irrelevant as time goes
the quality of the data.64 on.
Collecting Resource Use Data for Costing in Clinical Trials 223
Non-patient-specific resource use methods involved in the trial and having clear procedures
for data entry.
Where cost-generating events are not considered
key, it is sufficient to measure them on a
non-patient-specific basis. Existing records and Data Analysis
databases are also used to derive non-patient-
specific resource use data, for example, length of Summarising cost data
stay may be derived from HRG data. The gen-
Once resource use and unit cost data have been
eral limitations of administrative databases are
collected, a cost per patient can be calculated.
discussed above, but their use as estimates of
For each resource item, the resource quantity
non-patient-specific resource use raises further
used by each patient is multiplied by the unit
issues. For example, sample selection bias may
cost of the resource item. These costs are then
occur, that is, the population coverage of the
summed for each intervention, or arm of the
database is not representative of the population
trial, to give the total cost per intervention, or
at large.80
per arm of the trial. At its simplest, the mean
Estimates of resource use can also be derived
total cost for each intervention or arm of the
using consensus gathering techniques from
trial, is calculated by dividing the total cost by
experts (see Chapter 24), although this method
the number of patients in each arm. Costs should
is unlikely to form the main mode of data
be measured in a specific base year and hence
collection. Evans81 reviewed consensus gather-
adjusted for the effects of inflation and dis-
ing methods and found that the criteria for
counted to convert the future costs into present
selecting experts and the baseline information
values.2,4 An indication of uncertainty, or vari-
provided to the panel are important aspects of
ability, should also be reported, such as standard
designing panel studies. Expert panels have
deviation around a mean total cost.
been used in several studies,82–84 the advantage
of this being that it is an inexpensive, quick and Missing and censored data Certain features,
easy method for estimating resource use, par- such as missing and censored data, complicate
ticularly in economic evaluations that are per- this summarising process and need to be
formed after the clinical trial has ended. The addressed.2,86 It is important to know why data
limitations of basing resource use on expert are missing and what implications missing data
panel estimates is that it may be an inaccurate may have.58 Data may be missing because
method if recollection is poor and if estimates patients did not return or complete the data
relate to the ideal service rather than what hap- collection instruments. Data are seldom missing
pens in practice.27 at random, and therefore the analytical method
used in the presence of missing data has impli-
Mixed modes of data collection cations for the results.87 Missing data may be
handled by imputing mean values, but this may
Often, mixed modes of data collection will be artificially reduce the variability in total cost.
used to collect data for different types of Missing data may also arise because of loss to
resource use. Patient self-report and records are follow-up, and these are termed censored data. If
often used together either sequentially, in a two- patients have not been followed for the entire
step process, or concurrently.64 An example of a duration of the trial, then their costs are
two-step process is where patients report the unknown.
number of events and investigators follow-up on Statistical approaches used to handle censored
the detail of those events through records. Since clinical data have been adapted for use with
the problem of defining a gold standard method economic data. Non-parametric methods, such
of data collection remains, multiple sources of as Kaplan–Meier survival analysis and life-table
data can be used for the same information in analysis, have been suggested.86 In using this
order to test for validity.53 approach, costs, rather than survival time, are
used as the dependent variable and are treated as
Organisation of data collection right censored. It has been argued, however, that
this approach is only valid if the additional cost
Piloting of data collection methods is an import- per period is constant over time and a preferred
ant aspect of data collection design and activ- method is therefore to partition the survival
ities, such as data monitoring, are important curve over a number of intervals.88
organisational aspects.57,64 Steps to assess qual-
ity assurance, such as visits to each centre, and Synthesising costs and effects Incremental
maintaining contact between site visits are also cost–effectiveness ratios are only calculated for
important.85 Further organisational aspects relate interventions that are more costly and produce
to quality control issues, such as training staff greater effects than the intervention being used
224 The Advanced Handbook of Methods in Evidence Based Healthcare
in comparison. If more than one strategy is events and total costs can be explored by using
being compared, strategies can be ranked in measures of variability such as a standard devi-
terms of increasing effectiveness and dominated ation or variance, and these should be presented
strategies omitted from further consideration. around the mean cost differences. Variability in
Strategies or interventions are dominated if an cost-generating events can also be examined by
intervention is less effective and more costly calculating the variance/mean ratio (or over-
than an alternative; if an intervention is less dispersion) of events.43 If the variance equals the
costly and more effective then the treatment mean, there is no over-dispersion. Variability in
dominates.89 costs can be examined by calculating the co-
efficient of variation which expresses the stan-
Handling uncertainty dard deviation of a cost-generating event as a
percentage of the mean.16 This measure allows
Uncertainty arises from a number of sources. the variability between different types of cost to
These include the data inputs, methods and be compared.
assumptions as well as from the extrapolation or Sensitivity analysis involves the systematic
generalisability of results.90 Uncertainty exists investigation of how changes in the selected as-
with respect to the parameters used as data sumptions affect the costs and cost–effectiveness.
inputs, such as the unit costs or discount rate. It can be used to examine the impact of changes
Uncertainty relates to the methods used, such as in, for example, length of stay, on total costs or
the data collection methods or model, as well as cost–effectiveness. It can also be used to gen-
to the assumptions made by the investigator.4 eralise results.92 The parameters to be varied and
The generalisability of results to other settings is the amount of variation (or ranges) to be imputed
a further source of uncertainty.90 are selected. The specification of ranges across
The two main methods for analysing un-
which to vary parameters is a key component
certainty are statistical analysis and sensitivity
of sensitivity analysis, and justification for the
analysis, and these two forms of analysis have
ranges imputed should be given.90 Ranges are
complementary roles.4 Uncertainty in resource
based on clinically meaningful ranges; 95% con-
use and cost differences can be examined using
fidence intervals for parameters; or highest and
confidence intervals or by performing statistical
tests of difference, depending on whether the lowest ranges possible.4 The amount of change
interest is in estimation or hypothesis testing. in base results that is acceptable or constitutes a
An estimation approach is often preferred since robust finding also has to be specified. All the
estimation of a relevant quantity is more inform- stages of sensitivity analysis thus involve sub-
ative than significance value. With either ap- jective elements.
proach, examining the distribution of cost data is
important, since a relatively small number of
high-cost events can skew the distribution of Presentation and Reporting of Results
total cost (see Chapter 18). The implications
of skewed data are that the mean cost and the The emphasis on reporting of studies should be
variance around the mean are disproportionately on transparency. The methods used should be
affected, and the ability to detect significant made transparent and therefore a description of
differences between patient groups is perhaps the costing methods used, including the type of
reduced.9 There are several options in such a costs included should be given and the reasons
situation: use non-parametric tests; transform for adopting methods reported.14 Sample size
the data and continue to use parametric tests, calculations and assumptions should also be
assuming the transformation produces a normal reported,93 along with the reasoning behind
distribution;91 or use bootstrapping to construct centre selection. The data collection instruments
a confidence interval empirically. The latter two used and evidence on their validity and reliabil-
approaches are preferred since the aim is usually ity should be reported. Reporting of cost infor-
to determine a mean difference and an asso- mation should include sources of resource use
ciated confidence interval since, unlike the and unit cost estimates, price bases and discount
median, the mean cost can be related back to rates. The methods used to handle missing and
total cost. Examining uncertainty in the cost– censored data should be reported. Costs should
effectiveness ratio requires more complex statis- be presented with some information on their
tical methods which are still in development58 variability. Mean, rather than median costs,
(see Chapter 18). convey more useful information. The cost–
Confidence intervals should be presented effectiveness results can be presented graph-
around point estimates to indicate significant ically, for example, using the cost–effectiveness
differences. The uncertainty in cost-generating plane.94
Collecting Resource Use Data for Costing in Clinical Trials 225
The methodological issues described above can There are several key methodological issues that
be classified into those where there is agreement remain open because they depend upon the
as to how to handle them, and those where there perspective and values adopted in the study:
is disagreement. Given, however, that methodo- choice of perspective, adoption of a welfare
logical approaches to the collection and analysis theoretical approach and use of hypothesis test-
of resource use data for costing are still develop- ing or estimation (as shown in Table 5). These
ing, recommendations may impose rigid stan- issues can only be resolved by agreement as to
dards and constrain further methodological what are the appropriate values and perspective
development and debate. Clearly, there are some in the context of a particular study or pro-
methodological issues where there is agreement. gramme of research. These could of course
There are other issues, however, where disagree- legitimately be determined by a funder or user
ment about how to handle them is the inevitable of such studies prescribing what values that
consequence of needing different methods to body wishes to see reflected in studies under-
answer different economic questions, or the taken for it, or with its funding. This indeed has
same question in different health service con- been the case with the context-specific guide-
texts. Furthermore, resolutions of many of the lines produced for example in Australia12 or
design decisions in costing, and in economic Ontario15 for economic evaluations required to
evaluation more generally, are fundamentally support submissions to have new drugs re-
dependent on the theoretical perspective of the imbursed within those publicly funded health-
analyst, and the perspective of the study. Such care systems. As discussed above, there is dis-
issues can only be resolved once there is a agreement about the scope of measurement of
decision about the appropriate value system, productivity costs and future costs. Resolution
viewpoint or theoretical standpoint for a particu- of these issues requires more considered argu-
lar study. ments about the basis upon which they should
Consequently, in discussing the status of cur- be included or excluded.
rent methodological issues, issues are separated The other category of issue of disagreement is
into those where there is general agreement and shown in Table 5 and includes study design,
those where there is disagreement. The latter are data collection, data analysis and presentation
further divided into those which reflect legitimate issues. Many of them concern the more detailed
differences in values and perspective and those, methodological issues, for example, the sources
often more practical issues, that are amenable to of information, levels of necessary detail, the
further elucidation by empirical research. frequency of collecting data, modes of data
226 The Advanced Handbook of Methods in Evidence Based Healthcare
collection, and so on. In principle, for such 3 Practitioners should also ensure that, in
issues it should be possible to make specific or, determining the design, full use is made of
at least context-specific, recommendations based existing evidence relating to costs of the
on past experience. Unfortunately, whilst there technologies or treatments under study in
are successful studies that can be drawn on as order to aid the design of well-focused data
examples of seemingly good practice, there are collection. In doing so, investigators may be
few experimental studies to test the relative able to draw on data archives, if such are
performance of different methods. Hence it is established.
probably premature to offer firm recommenda- 4 If, when designing studies, practitioners con-
tions on methods for handling even these meth- front methodological issues of disagreement
odological issues. These issues could, and that could be resolved by further empirical
should, be resolved by formal empirical testing research, they should consider whether these
of alternative methods and thus form part of a could be formally tested within the applied
future research agenda. study. This would contribute to the reduction
in uncertainty surrounding the relative merits
of alternative approaches.
5 In order to aid transparency of results, prac-
IMPLICATIONS FOR PRACTITIONERS OF HTA titioners should aim to ensure that details of
their data collection methods and instru-
There are several implications for practitioners ments are readily available and ensure that
arising from this review: they report results in such a way that analy-
sis is repeatable by other investigators.
1 Practitioners should adhere to the methodo- 6 Practitioners should also be willing to share
logical issues of agreement (Table 5). data from previous studies with other re-
2 In undertaking studies, practitioners should searchers designing costing studies in similar
consider, in a systematic way, the options areas. They should be prepared to deposit
for resource use data collection and ensure work in data archives if such are set up and,
that the rationale for the chosen design is in so doing so, contribute to the critical
made explicit and transparent. appraisal of the usefulness of such archives.
Collecting Resource Use Data for Costing in Clinical Trials 227
20. Rutten-Van Molken MPMH, Van Doorslaer of life? A response to the recommendation of the
EKA, Jansen MCC, Kerstijens HAM, Rutten Washington panel. Health Economics 1997; 6:
FFH. Costs and effects of inhaled corticosteroids 253–9.
and bronchodilators in asthma and chronic 34. Koopmanschap MA, Rutten FFH. Indirect costs
obstructive pulmonary disease. Am. J. Respir. in economic studies: confronting the confusion.
Crit. Care Med. 1995; 151: 975–82. PharmacoEconomics 1993; 4: 446–54.
21. Schulman KA, Buxton M, Glick H, et al. Results 35. van Roijen L, Essink-Bot ML, Koopmanschap
of the economic evaluation of the first study: a MA, Bonsel G, Rutten FFH. Labor and health
multinational prospective economic evaluation. status in economic evaluation of health care: the
Int. J. Technol. Assess. Health Care 1996; 12: health and labor questionnaire. Int. J. Technol.
698–713. Assess. Health Care 1996; 12: 405–15.
22. Sculpher MJ, Dwyer N, Byford S, Stirrat G. 36. Koopmanschap MA, Rutten FFH. A practical
Randomised trial comparing hysterectomy and guide for calculating indirect costs of disease.
transcervical endometrial resection: effect on PharmacoEconomics 1996; 10: 460–6.
health related quality of life and costs two years 37. Garber AM, Phelps CE. Economic foundations of
after surgery. Br. J. Obstet. Gynaecol. 1996; 103: cost-effectiveness analysis. J. Health Economics
142–9. 1997; 16: 1–31.
23. van Bergen PFMM. Costs and effects of long- 38. Weinstein MC, Manning WG. Theoretical issues
term oral anticoagulant treatment after myo- in cost-effectiveness analysis. J. Health Economics
cardial infarction. JAMA 1995; 273: 925–8. 1997; 16: 121–8.
24. Wonderling D, McDermott C, Buxton M, et al. 39. Birch S, Gafni A. Cost-effectiveness ratios: in
Costs and cost effectiveness of cardiovascular a league of their own. Health Policy 1994; 28:
screening and intervention: the British family 133–41.
heart study. Br. Med. J. 1996; 312: 1269–73. 40. Davidoff A, Powe NR. The role of perspective in
25. Rittenhouse B. Uses of models in economic evalu- defining measures for the economic evaluation of
ations of medicines and other health technologies. medical technology. Int. J. Technol. Assess.
London: Office of Health Economics, 1996.
Health Care 1996; 12: 9–21.
26. Langham S, Thorogood M, Normand C, Muir J,
41. Gerard K, Mooney G. QALY league tables:
Jones L, Fowler G. Costs and cost effectiveness
Handle with care. Health Economics 1993; 2:
of health checks conducted by nurses in primary
59–64.
care: the Oxcheck study. Br. Med. J. 1996; 312:
42. Johannesson M. Avoiding double-counting in
1265–8.
pharmacoeconomics studies. PharmacoEcon-
27. Drummond MF, Jefferson TO, on behalf of the
omics 1997; 11: 385–8.
BMJ economic evaluation working party. Guide-
43. Spiegelhalter DJ, Jones DR, Parmar MKB, Gore
lines for authors and peer reviewers of economic
submissions to the BMJ. Br. Med. J. 1996; 313: SM, Fitzpatrick R, Cox DR. Being economical
275–83. with the costs. Department of Epidemiology and
28. Meltzer D. Accounting for future costs in medical Public Health Technical Paper 96-06, University
cost-effectiveness analysis. J. Health Economics of Leicester, 1996.
1997; 16: 33–64. 44. Sculpher M, Drummond M, Buxton M. The
29. Smith K, Wright K. Informal care and economic iterative use of economic evaluation as part of
appraisal: a discussion of possible methodological the process of health technology assessment. J.
approaches. Health Economics 1994; 3: 137–48. Health Services Res. Policy 1997; 2: 26–30.
30. Bryan S, Buxton M, McKenna M, Ashton H, 45. Jonsson B, Weinstein MC. Economic evaluation
Scott A. Private costs associated with abdominal alongside multinational clinical trials: study con-
aortic aneurysm screening: the importance of siderations for GUSTO 11b. Int. J. Technol.
private travel and time costs. J. Med. Screening Assess. Health Care 1997; 13: 49–58.
1995; 2: 62–6. 46. Hurley SF, Bond LM, Carlin JB, Evans DB,
31. Sculpher MJ, Buxton MJ. The private costs incur- Kaldor JM. A method for estimating baseline
red when patients visit screening clinics: the cases health care costs. J. Epidemiol. Community Health
of screening for breast cancer and for diabetic 1995; 49: 525–31.
retinopathy. Health Economics Research Group 47. Dranove D. Measuring costs. In: Sloan FA (ed.)
Discussion Paper, Brunel University, Uxbridge, Valuing health care: cost, benefits and effective-
1993. ness of pharmaceuticals and other medical tech-
32. Weinstein MC, Siegel JE, Garber AM, et al. nologies. Cambridge; Cambridge University Press,
Productivity costs, time costs and health-related 1995.
quality of life: a response to the Erasmus group. 48. Buxton MJ, Drummond MF, Van Hout BA, et al.
Health Economics 1997; 6: 505–10. Modelling in economic evaluation: an unavoid-
33. Brouwer WBF, Koopmanschap MA, Rutten FFH. able fact of life. Health Economics 1997; 6:
Productivity costs measurement through quality 217–27.
Collecting Resource Use Data for Costing in Clinical Trials 229
49. Morris J, Goddard M. Economic evaluation and during phase II/III clinical trials. Pharmaco-
quality of life assessments in cancer clinical Economics 1996; 9: 264–77.
trials: the CHART trial. Eur. J. Cancer 1992; 65. Donaldson C, Hundley V, McIntosh E. Using
29A: 766–70. economics alongside clinical trials: why we can-
50. Knapp M, Beecham J. Reduced list costings: not choose the evaluation technique in advance.
examination of an informed short cut in mental Health Economics 1996; 5: 267–9.
health research. Health Economics 1993; 2: 66. Anie KA, Jones PW, Hilton SR, Anderson HR. A
313–22. computer-assisted telephone interview technique
51. Backhouse M, Mauskopf JA, Jones D, Wold DE. for assessment of asthma morbidity and drug use
Economics outcomes of colfosceril palmitate in adult asthma. J. Clin. Epidemiol. 1996; 49:
rescue therapy in infants weighing 1250g or more 653–6.
with respiratory distress syndrome. Pharmaco- 67. Streiner DL, Norman GR. Health measurement
Economics 1994; 6: 358–69. scales: A practical guide to their development and
52. Sheldon TA. Problems of using modelling in use. Oxford: Oxford University Press, 1989.
the economic evaluation of health care. Health 68. Copley-Merriman C, Egbuonu-Davis L, Kotsanos
Economics 1996; 5: 1–11. JG, Conforti P, Franson T, Gordon G. Clinical
53. Clark RE, Teague GB, Ricketts SK, et al. Meas- economics: a method for prospective health
uring resource use in economic evaluations: resource data collection. PharmacoEconomics
determining the social costs of mental illness. J. 1992; 1: 370–6.
Mental Health Admin. 1994; 21: 32–41. 69. Beecham J. Collecting information: The client
54. O’Brien BJ, Drummond MF, Labelle RJ, Willan service receipt interview. Mental Health Res. Rev.
A. In search of power and significance: issues 1994; 1: 6–8.
in the design and analysis of stochastic cost- 70. Hundley VA, Donaldson C, Lang GD, et al.
effectiveness studies in health care. Medical Care Costs of intrapartum care in a midwife- managed
1994; 32: 150–63. delivery unit and a consultant led labour ward.
55. Drummond M, O’Brien B. Clinical importance, Midwifery 1995; 11: 103–9.
statistical significance and the assessment of
71. Hughes J, Ryan M, Hinshaw K, Henshaw R,
economic and quality-of-life outcomes. Health
Rispin R, Templeton A. The costs of treating
Economics 1993; 2: 205–12.
miscarriage: a comparison of medical and sur-
56. Torgerson DJ, Ryan M, Ratcliffe J. Economics in
gical management. Br. J. Obstet. Gynaecol. 1996;
sample size determination for clinical trials. Q. J.
103: 1217–21.
Med. 1995; 88: 517–21.
72. Brown JB, Adams ME. Patients as reliable
57. Coyle D, Davies L, Drummond M. Trials and
reporters of medical care process: recall of ambu-
tribulations: Emerging issues in designing econ-
latory encounter events. Medical Care 1992; 30:
omic evaluations alongside clinical trials. Int. J.
Technol. Assess. Health Care 1998; 14: 135–44. 400–11.
58. Mullahy J, Manning W. Statistical issues in cost- 73. Revicki DA, Irwin D, Reblando J, Simon GE.
effectiveness analysis. In: Sloan FA (ed.) Valuing The accuracy of self-reported disability days.
health care: costs benefits and effectiveness of Medical Care 1994; 32: 401–4.
pharmaceutical and other medical technologies. 74. Boyer GS, Templin DW, Goring WP et al. Dis-
Cambridge; Cambridge University Press, 1995. crepancies between patient recall and the medical
59. Mullahy J. What you don’t know can’t hurt you? record. Arch. Intern. Med. 1995; 155: 1868–72.
Statistical issues and standards for medical tech- 75. Linet MS, Harlow SD, McLaughlin JK,
nology evaluation. Medical Care 1996; 34: McCaffrey LD. A comparison of interview data
DS124–DS135. and medical records for previous medical condi-
60. Wittes J, Brittain E. The role of internal pilot tions and surgery. J. Clin. Epidemiol. 1989; 42:
studies in increasing the efficiency of clinical 1207–13.
trials. Stat. Med. 1990; 9: 65–72. 76. Lave JR, Pashos CL, Anderson GF, et al. Costing
61. Jacobs P, Baladi JF. Biases in cost measurement medical care: Using medicare administrative data.
for economic evaluation studies in health care. Medical Care 1994; 32: JS77–JS89.
Health Economics 1996; 5: 525–9. 77. Aaronson LS, Burman ME. Focus on Psycho-
62. Rittenhouse B. Exorcising protocol-induced metrics. Use of health records in research: reli-
spirits: making the clinical trial relevant for econ- ability and validity issues. Research in Nursing
omics. Medical Decision Making 1997; 17: and Health 1994; 17: 67–73.
331–9. 78. Safran C. Using routinely collected data for clin-
63. CHKS. Acute Care 1996: Healthcare resource ical research. Stat. Med. 1991; 10: 559–64.
groups national statistics 1995/1996. Warwick- 79. Wolff N, Helminiak TW. Nonsampling measure-
shire: CHKS, 1996. ment error in administrative data: implications for
64. Mauskopf J, Schulman K, Bell L, Glick H. economic evaluations. Health Economics 1996; 5:
A strategy for collecting pharmacoeconomic data 501–12.
230 The Advanced Handbook of Methods in Evidence Based Healthcare
80. Kuykendall DH, Johnson ML. Administrative data in economic evaluation. Medical Care 1995;
databases, case-mix adjustments and hospital 33: 851–63.
resource use: the appropriateness of controlling 87. Crawford SL, Tennstedt SL, McKinlay JB. A
patient characteristics. J. Clin. Epidemiol. 1995; comparison of analytic methods for non-random
48: 423–30. missingness of outcome data. J. Clin. Epidemiol.
81. Evans C. The use of consensus methods and 1995; 48: 209–19.
expert panels in pharmacoeconomic studies. Prac- 88. Lin DY, Feuer EJ, Etzioni R, Wax Y. Estimating
tical applications and methodological shortcom- medical costs from incomplete follow-up data.
ings. PharmacoEconomics 1997; 12: 121–9. Biometrics 1997; 53: 419–34.
82. Phillips K, Lowe E, Kahn J. The cost- 89. Karlsson G, Johanesson M. The decision rules of
effectiveness of HIV testing of physicians and cost-effectiveness analysis. PharmacoEconomics
dentists in the United States. JAMA 1994; 271: 1996; 9: 113–20.
851–8. 90. Briggs A, Sculpher M, Buxton M. Uncertainty
83. Jonsson B, Bebbington PE. What price depres- in the economic evaluation of health care tech-
sion? The cost of depression and the cost- nologies: the role of sensitivity analysis. Health
effectiveness of pharmacological treatment. Br. J. Economics 1994; 3: 95–104.
Psychiatry 1994; 164: 665–73. 91. Coyle D. Statistical analysis in pharmacoeconomic
84. O’Brien B, Goeree R, Mohamed AH, Hunt R. studies. PharmacoEconomics 1996; 9: 506–16.
Cost-effectiveness of helicobacter pylori eradica- 92. Sculpher MJ, Michaels J, McKenna M, Minor
tion for the long-term management of duodenal J. A cost-utility analysis of laser assisted angio-
ulcer in Canada. Arch. Intern. Med. 1995; 155: plasty for peripheral arterial occlusions. Int. J.
1958–64. Technol. Assess. Health Care 1996; 12: 104–25.
85. Morris J, Goddard M, Coyle D, Drummond M. 93. Mason J, Drummond M. Reporting guidelines
Measuring cost and quality of life in radiotherapy for economic studies. Health Economics 1995; 4:
trial treatments. Centre for Health Economics, 85–94.
discussion paper 112 ,1993. 94. Black WC. The cost-effectiveness plane: a graphic
86. Fenn P, McGuire A, Phillips V, Backhouse M, representation of cost-effectiveness. Medical Deci-
Jones D. The analysis of censored treatment cost sion Making 1990; 10: 212–15.
14
Eliciting Time Preferences for Health
J O H N C A I R N S and M A R J O N V A N D E R P O L
future consumption can be enjoyed. Thus, the and stated preference. The distinction is that the
opportunity cost of current consumption is some former involves observing actual behaviour,
higher level of future consumption, and the specifically inter-temporal decisions, whereas
discount factor is a formal recognition of this the latter involves asking individuals what they
opportunity cost. At the individual level it is would do in particular hypothetical circum-
suggested that most individuals have time pref- stances. Despite a predisposition in favour of
erences; that is, they are not indifferent with revealed preference, economists have in recent
respect to the timing of future events. The years shown an increasing willingness to explore
reasons suggested for this include: that individ- the stated preference approach. There are still
uals cannot be sure that they will be alive at any concerns about the validity of the information
particular point in the future; that they anticipate generated, and the ideal corroborating evidence
being better off and will as a result attach less remains observed behaviour.
weight to further increments to their wealth or A wide range of behaviour has been studied
income; and what Pigou famously described as a including: the purchase of consumer dur-
defective telescopic faculty. ables;1–4 educational investment decisions;5 food
Discounting practices often play a central role consumption;6 and labour market wage-risk
in determining the relative cost–effectiveness choices.7–9 These studies generally are based on
of different interventions. If evaluations are larger sample sizes than those used in applica-
undertaken on an incorrect basis the quality of tions of the stated preference approach. Also, the
decision-making will suffer and health service estimation of discount rates is relatively indirect
efficiency will be reduced. Moreover, confusion and quite complicated. This results partly from
or lack of agreement over standard discounting the difficulty of using data collected primarily
practice potentially undermines the credibility for some other purpose and the many more
and value of economic evaluation. Different factors outwith the researchers control (as com-
rates are applied in different jurisdictions but pared to an experimental approach).
they are generally between 0.03 and 0.06. The The stated preference approach has also been
impact of discounting is especially marked when applied in a wide range of settings. These have
considering projects where some of the effects included: financial choices;10–13 purchases of
are fairly far in the future. For example, the consumer durables;14,15 saving lives; and non-
discount factor applied to costs arising in 20 fatal changes in health. The last two groups of
years time would be 0.31 at the currently recom- studies are of course of particular relevance.
mended UK rate (r 5 0.06). Underlying both of these approaches has been
Despite the potential importance of individ- a reliance on the Discounted Utility (DU) model
uals’ preferences for future health events, cur- which has dominated economic thought with
rent knowledge regarding the nature of these respect to inter-temporal choice for over 50
preferences is poor. Systematic investigation of, years. The most important assumption of the
the influence of type of choice on apparent time model is that individuals discount future events
preferences, the characteristics of different at a constant rate. Loewenstein and Prelec16
methods of eliciting preferences, and the under- identify four inter-temporal choice anomalies
lying models of time preferences for analysing that run counter to the predictions of the model.
responses, is required. They describe these as: (i) the common differ-
ence effect; (ii) the absolute magnitude effect;
(iii) the gain-loss asymmetry; and (iv) the
AIMS OF THE CHAPTER delay–speedup asymmetry. The last three are
self-explanatory. The DU model assumes that
This chapter has the following aims: the discount rate applied will not be related to
the magnitude of the event which is subject to
1 To review the stated preference approach to discounting, nor to whether the event represents
the elicitation of inter-temporal preferences. a gain or a loss, nor to whether it is being
2 To outline The Estimation of Marginal time brought forward or delayed. The common differ-
Preference in a UK-wide Sample (TEMPUS) ence effect refers to the impact on choice
project and explain the choice of methods between two delayed outcomes of a change in
used. the delay applied equally to both outcomes. The
3 To report some preliminary findings. DU model assumes that the choice depends
only on the absolute interval between the two
outcomes.
THE ESTIMATION OF TIME PREFERENCES These anomalies are explained by Loewen-
stein and Prelec in general terms with reference
Two broad approaches have been used to esti- to future consumption, and supported by evi-
mate time preference rates – revealed preference dence from monetary choices. However, there is
Eliciting Time Preferences for Health 233
no reason to suppose that they are any less in suggest that ‘. . . time preference is measured by
evidence when the outcomes are in terms of asking conventional time preference questions
health. . . . but cast in the health, as opposed to financial
domain.’ (p. 449). Also they claim that it is not
necessary to speculate on the nature of time
Is Health Different? preference ‘. . . since it is empirically determin-
able’ (p. 449). However, drawing on Loewen-
Before outlining the work done on the estima- stein and Prelec,24 which highlighted the import-
tion of time preferences for health it is worth ance of another class of effects – sequence
addressing the basic question – is health differ- effects, Gafni23 argues robustly that no measure-
ent? There are at least two ways in which this ment technique allows pure time preference
question can be addressed: as a practical issue to be distinguished, and the best that can be
arising in economic evaluation; and at a more achieved is a measure of time preference for a
methodological level. The practical question is given sequence of events. This may be true of
whether or not future health effects should be preferences over one’s own future health states;
discounted at the same rate as future costs. This however, it is less clear that the sequence of
question has generated considerable interest in events will be an important influence when con-
recent years, and much of the empirical work sidering preferences over life-saving profiles.
that has been undertaken has aimed to inform A further way in which some health benefits
this debate. may be different arises when the change in
Parsonage and Neuberger17 claim that ‘. . . in health is measured in terms of QALYs. Krahn
practice, for most purposes, it is appropriate to and Gafni25 argue that, because of the methods
use a zero discount rate for future health bene- sometimes used to measure QALYs, time pref-
fits’ (p. 72). They argue that the traditional sour- erences may already have been taken into
ces of time preference are unimportant in the account. As a result further discounting of the
context of future health benefits. Whether or not QALYs might represent a form of double
monetary costs and non-monetary benefits counting.
should be discounted at the same rate can be
viewed as largely an empirical question.18 Is the
sum of undiscounted health benefits a better or a Using Stated Preference to Estimate
poorer approximation to the true value of a Time Preferences for Health
stream of future benefits than the sum of the
discounted health benefits? The answer would As noted above, a distinction can be drawn
appear to depend on the social valuation of the between those studies with respect to life saving
health benefits and how it changes over time. and those with respect to non-fatal changes in
The greater the increase in social valuation over health. Table 1 provides a brief overview of this
time the closer will the undiscounted sum literature.
approximate the true value. The smaller the
increase in social valuation, the closer the
approximation provided by the discounted sum.
Saving lives
Currently the only jurisdiction where formal The first studies of time preferences with
guidance encourages the differential discounting respect to life saving26–29 adopted fairly similar
of monetary costs and non-monetary benefits is approaches. Horowitz and Carson26 offered
England and Wales, where 6% is recommended respondents dichotomous choices between a
for financial values and 1.5–2% for health programme saving lives for the next 15 years
effects quantified in physical units.19 Such advice and a programme saving lives for 10 years,
is at odds with that elsewhere, for example, starting in 5 years time. Cropper et al.28 asked
Canada20 and the USA21 (5% and 3% respect- two linked dichotomous questions offering a
ively applied to all costs and benefits). context-free choice between programmes saving
The methodological issue concerns whether lives now and various times in the future. Both
or not it is possible to identify individuals’ rates then used econometric methods to identify
of time preference with respect to future health median implied time preferences rates for the
effects and is rather more important in the sample interviewed. A broadly similar approach
context of this chapter. Gafni and Torrance22 was followed in a recent Swedish study30 of
hint at the potential difficulties in doing so, and intergenerational choices.
Gafni23 develops the argument more fully. Gafni Later studies by Olsen,31 Cairns32 and Cairns
and Torrance relate attitude towards risk in and van der Pol33,34 used variants of an open-
health to three distinct effects: a quantity effect; ended approach where the data were collected
a time preference effect; and a gambling effect. with a self-completed questionnaire. Through
It is interesting to note that Gafni and Torrance this approach the identification of rates of time
234
The Advanced Handbook of Methods in Evidence Based Healthcare
Table 1 Summary of empirical time preference literature in health
Study Median r Mean r Type Delay Sample
preference can be made on an individual basis health profiles which differ only in the timing of
through simple manipulation of the responses. the period of ill-health.
The price of this advantage may be increased
problems with respect to framing effects and
related biases. Research Issues Addressed by the
A third approach followed by Enemark et al.35 TEMPUS Project
differs in that it is not concerned with choices
between saving differing numbers of statistical Despite the growing number of empirical studies
lives at different times in the future. It concerns there are numerous opportunities for further
the choice between watchful waiting and sur- research to add to the understanding of time
gery for abdominal aortic aneurysms for a hypo- preferences, and ultimately lead to better
thetical cohort. In a classification between life- informed decision-making with respect to dis-
saving and non-fatal changes in health it is counting practice. This listing does not seek to
probably appropriate to put it with the life- be comprehensive but rather highlights three
saving studies, but it is rather different from the areas explored in the TEMPUS project.
rest of the literature and as such may offer a new
line of enquiry. Different methods of eliciting time
preferences
Non-fatal changes in health Little is known about the influence on estimates
The first significant study of time preferences for of implied discount rates of different methods of
non-fatal changes in health was by Lipscomb.36 eliciting preferences. A major issue in the field
His approach requires subjects to classify 96 of contingent valuation has been the use of
scenarios on a 0 to 10 category scale. Regres- closed-ended (are you willing to pay £100?)
sion analysis is then used to explain these scores versus open-ended questions (how much are you
using the characteristics of the health state. willing to pay?). With respect to the elicitation of
Lipscomb suggests that time preference ‘. . . be time preferences, open-ended methods31–34,37–39
regarded as operationally equivalent to the (mar- have proved more popular to date than closed-
ginal) influence of the delay-of-onset variable ended,26–30 particularly if those studies adopting
on preference scores’ (p. S243). Notwithstand- a more indirect approach (via the valuation of
ing the practical difficulties of applying the different health outcomes35,36,40–42) are classified
approach (such as the demands on individual as open-ended. No studies have been published
respondents), it represents an attractive means of which explicitly compare different methods of
determining the impact of delay on the valuation eliciting time preferences.
of an outcome. Further issues concern the use of self-
In the report by Cairns,37 implied discount completed questionnaires versus interviews
rates are calculated directly from the duration of (face-to-face and telephonic), and the presence
ill health which renders the respondent indif- and impact of framing effects. Both of these
have yet to be systematically investigated.
ferent between being ill in the further future and
Whilst there are undoubtedly lessons to be
in the near future. A broadly similar approach
learned from work in other fields, such as con-
has been followed by Chapman and Elstein38 tingent valuation, there is clearly considerable
and Chapman39 in which subjects indicated what scope for designing time preference studies to
period of relief from ill health at a specified address these issues directly.
point in the future would make them indifferent
between relief from ill health for a specified
period now. Nature of the inter-temporal choice
Redelmeier and Heller40 used the standard An important question on which little is known
gamble and categorical scaling to elicit prefer- is the extent and nature of differences between
ences for a number of different health states, one time preferences with respect to one’s own
aspect of which was the timing of onset of ill health and others’ health. The studies with
health. MacKeigan et al.41 also used a categor- respect to saving statistical lives clearly refer to
ical rating scale to elicit preferences for different the health of others.26–34 Generally, questions
health states with different delays of onset. They with respect to non-fatal changes in health have
examined the effect of duration of health gain been posed in terms of the respondents’ own
and health loss, and delay before health change, health,37–42 although MacKeigan et al.,41 whilst
on the valuation of health states, but did not inviting respondents to consider their own
estimate the implied discount rates. Finally, health, asked respondents to imagine that they
Dolan and Gudex42 estimate time preference were 50 years old and married with children
rates using time trade-off valuations for two (none of whom lived with them). The exceptions
236 The Advanced Handbook of Methods in Evidence Based Healthcare
are: Lipscomb,36 whose scenarios start ‘A per- models which allow for decreasing timing aver-
son in your community is now 25 years old’; sion. However, there has been little investigation
and part of the Olsen31 life-saving study in of alternative models.27,33,34
which the life-saving questions were repeated
substituting ‘. . . a programme which improves The TEMPUS Project
the health of people in a chronic state of dys-
function and distress’ for ‘. . . a programme The specific objectives of the TEMPUS study
which saves human lives’; and a study of are:
vascular surgeons time preferences for their
patients’ future health.35 The distinction between 1 To derive implied discount rates for future
own and others’ health is important because health benefits for a sample of the general
time preferences with respect to own health are public in the UK.
likely to be more relevant to explaining health- 2 To establish whether individual inter-
affecting behaviour, whereas preferences with temporal preferences with respect to their
respect to others’ health are potentially more own health differ from those with respect to
the health of others.
important if the focus is on the evaluation of
3 To investigate the effect of different ways of
publicly funded healthcare programmes. asking questions on apparent inter-temporal
Further issues concern: the nature and extent preferences (specifically closed-ended and
of differences between time preferences for sav- open-ended methods are compared).
ing lives and those for non-fatal changes in 4 To establish whether individuals value future
health; and whether time preferences are to health benefits in line with the traditional
some extent health state-specific. Loewenstein discounted utility model and to investigate in
claims ‘. . . the dependence of discounting on addition how well the hyperbolic and pro-
the characteristics of awaited consumption portional discounting models explain indi-
means that discount rates estimated in specific vidual responses.
contexts . . . cannot be generalised beyond the
domain of behaviour in which they were
derived’.43 Although this remark was made with DESIGN AND METHODS
respect to a comparison of the purchase of
consumer durables and savings behaviour, it
does raise the issue of whether implied discount Stated preference techniques comprising a series
of health-related choices are used to elicit the
rates are similar within the health domain. The
time preferences of a random sample of adults.
health states examined to date have included: In order to meet the specific objectives of the
an arthritic condition; a strangulated internal TEMPUS project, both open-ended and closed-
hernia; a depressive illness; and temporary ended questions were used. There are several
blindness. choices to be made when designing either open-
ended or closed-ended time preference ques-
Underlying model of time preferences tions. Most are relevant to both types of ques-
tionnaire but some are specific to one or other
Harvey44 has emphasised the distinction between question type.
constant and decreasing timing aversion. Most, The following are potentially relevant to both
but not all, studies36,41 have assumed constant types of question:
timing aversion, whereas the empirical evidence
(a) Number of points in time: respondents can
suggests that decreasing timing aversion might
be asked to consider two points in time or
represent individuals’ inter-temporal preferences they can be presented with a profile. The
more accurately. Preferences are defined as standard approach has been the former, with
constant timing averse where an individual is only two studies comparing profiles.26,35
indifferent between (x, s) and (y, t), and between (b) Base health state: the base health state can
(x, s 1 D) and (y, t 1 D), where x and y are be full-health and respondents make choices
benefits accruing at times s and t and 0 , x , y with respect to the consumption of ill-
and s , t. Preferences are defined as decreasing health,37,40,41 or the base health state is
timing averse where an individual is indifferent ill-health and respondents make choices
between (x, s) and (y, t) and prefers (y, t 1 D) to with respect to the consumption of full-
(x, s 1 D). That is, less importance is attached to health.38,39,41
a fixed difference t 2 s between two times s and (c) Number of ill-health states considered: one
t the farther into the future these times are ill-health state37–39,41 can be considered or
moved. Proportional discounting models44 and more than one ill-health state36,40,42 can be
hyperbolic discounting models45 are examples of considered.
Eliciting Time Preferences for Health 237
individual. Two different years are chosen for Closed-Ended Method: Discrete
the initial point at which ill health will be Choice Modelling
experienced if treatment is not received. Having
more than two starting points might make the The basic underlying idea is that individuals’
questions too difficult to answer, whereas hav- time preferences are revealed by how much
ing a common starting point for all six questions longer they are willing to be ill for in exchange
might result in respondents getting bored. for a delay in the onset of that ill-health. In order
The periods of delay chosen range from 2 to to establish the relative importance of duration
13 years from the starting point. The range of of a selected ill-health state and the relative
delays were selected on the pragmatic grounds importance of the year in which the spell of ill-
that if the delay was too short, factors other than health occurs, these two attributes are included.
time preference would have a major influence on The Euroqol health state 11221 used in the
the responses, and if it was too long, some open-ended study was also used in the discrete
respondents might have difficulties imagining it choice modelling.
(particularly older respondents). Each subject is The choice of attribute levels (for duration of
asked six questions: three with a starting point 2 ill-health and year in which the ill-health occurs)
years in the future; and three with a starting is to some extent arbitrary. It is important that
point 3 years in the future. Each questionnaire the options are plausible and that they are
contains three different delays for each starting informative, in the sense that it is not always
point: a short-term delay (2–5 years); a medium- obvious which choice will be made. The levels
term delay (6–9 years); and a long-term delay chosen for the year in which the spell of ill-
(10–13 years). Also the difference between the health occurs are: 2, 7 and 15 years in the future.
short- and medium-term delay and between the The levels chosen for the duration of the health
medium- and long-term delay is not the same in state are: 20, 24, 26, 33 and 38 days.
any questionnaire (for each starting point). These attributes and levels give rise to 15
Table 2 shows four different versions of the possible scenarios (31 3 51). A discrete choice
questionnaire (A, B, C, and D) which have these method is used where individuals are asked to
properties. The purpose of this somewhat com- choose between pairs of scenarios. Each pairwise
plex design is to collect data for a wide range comparison implies a particular discount rate.
of delays, and to ensure that the questions do Two sets of eight pairwise choices are selected
not appear to conform to any pattern in case based on the following criteria: (i) inclusion of
respondents think that they are expected to all 15 scenarios; (ii) a wide range of implied
respond in a certain way. discount rates; and (iii) one pairwise comparison
Derivation of implied discount rates is straight- representing a negative discount rate. Table 3
forward with such data. The individual chooses shows the selected pairwise comparisons.
a duration of ill-health x such that if it were The design of the questionnaire has to place
experienced y years in the future it would be the pairwise choices in a realistic context, and
equivalent to 20 days experienced 2 years in the has to minimise the number and the effect of
future. Letting b 5 1/(1 1 r) the respondent’s possible biases. Generally, individuals them-
choice implies that 20b2 5 xby22 and selves cannot delay their ill-health. In order to
make it as realistic as possible the respondent is
SD
1
x asked to imagine that s/he will be ill and that
y22
r5 21 (1) two alternative treatments are available. The
20 effects of the alternative treatments vary with
regard to when the illness occurs and how long
it lasts. They are then asked to indicate which
Table 2 The four versions of the treatment they prefer. (See Appendix 3 for an
questionnaire (A–D) example.)
Years before delayed ill health
A relationship must be specified between the
attributes and utility. The difference in utility
Starting point A B C D between option B and A can be expressed as:
Version 1
1 7 33 2 26 0.05
2 2 33 15 38 0.01
3 2 20 7 38 0.14
4 15 38 7 20 0.08
5 2 38 7 33 20.03
6 7 20 15 26 0.03
7 7 33 2 20 0.11
8 7 20 15 24 0.02
Version 2
1 15 38 7 24 0.06
2 2 33 15 38 0.01
3 7 38 2 20 0.14
4 7 20 15 38 0.08
5 15 20 7 24 20.02
6 2 26 15 33 0.03
7 2 20 7 33 0.11
8 7 26 2 24 0.02
The ratio of the coefficients of Days and Year The Four Questionnaires
represent the marginal rate of substitution
between duration and year, in other words how The different type of questions that are required
much longer individuals are willing to be ill for with respect to both own and others’ health are:
a delay in the onset of that ill-health. The ratios six open-ended questions (four versions); and
b2/b1 and b4/b3 can be used to identify (year, eight discrete choices (two versions). In order to
duration) pairs between which the individual avoid overburdening respondents there are four
would be indifferent. From this the implied different types of questionnaires:
discount rate can be estimated.
1 Six open-ended questions with respect to
b2 b4 own health;
r5 or r 5 (3) 2 Six open-ended questions with respect to
b1Daysn b3Daysn
others’ health;
The b coefficients are estimated using a random 3 Eight discrete choices with respect to own
effects ordered probit model.47 This technique is health;
preferred to ordinary least squares because of 4 Eight discrete choices with respect to others’
the ordinal nature of the dependent variable health.
(prefer A, no preference, prefer B) and because The comparisons that are made using these
multiple observations are obtained from each four distinct questionnaires are summarised in
individual. Figure 1.
Table 4 Response rate by questionnaire and a university degree. The average age of the
area sample is 49.9 years, with a minimum of 17 and
a maximum of 95. The ill-health state was
I II III IV valued at 54.6 on the visual–analogue scale.
Table 6 shows how the sample compares with
Caithness 0.152 0.158 0.268 0.219
Edinburgh 0.220 0.210 0.286 0.331
the population in terms of age, gender and educa-
Norfolk 0.146 0.169 0.275 0.273 tion. The age–gender structure of the sample
Manchester 0.157 0.131 0.231 0.190 seems fairly representative for the population.
Pembrokeshire 0.176 0.144 0.189 0.188 However, the younger age group (,30 years)
Cardiff 0.189 0.169 0.269 0.268 tends to be under-represented. A higher per-
Overall 0.173 0.163 0.252 0.244 centage of respondents have a university degree
Overall (I–IV) 0.218 compared to the population.
Age (years)
,30 13.6 23.6 9.6 20.8 11.8 20.9
30–44 28.4 25.9 31.3 25.0 24.1 27.8
45–64 42.6 30.6 32.2 31.0 43.6 31.6
.64 14.8 19.9 26.5 23.2 20.0 19.8
Gender
Male 44.3 48.6 49.6 46.2 41.0 49.4
Female 55.7 51.4 50.0 53.8 58.5 50.6
Education
University 17.6 5.8 27.4 10.4 21.5 7.2
Age (years)
,30 28.7 28.9 14.6 17.8 13.2 20.3
30–44 26.8 27.4 22.1 24.4 25.2 25.3
45–64 20.4 24.0 30.7 30.5 37.7 31.4
.64 23.6 19.7 32.2 27.3 23.2 23.0
Gender
Male 45.2 47.3 47.2 46.1 52.3 48.6
Female 54.8 52.7 52.3 53.9 47.7 51.4
Education
University 49.0 16.9 34.7 16.6 25.8 7.4
investigate functional forms for alternative dis- TEMPUS study and the methods used were
counting models, but this is not reported in this described, after which some preliminary results
chapter.48 were reported.
The implied discount rates elicited in this
study should not be over-emphasised because of
CONCLUSIONS the unrepresentativeness of the study sample. In
particular, the respondents were markedly better
This chapter has reviewed the stated preference educated than the population as a whole. This is
approach to the elicitation of inter-temporal not surprising given the deliberate policy of
preferences for health events. Several important sampling in more socio-economically advant-
research issues which remain to be addressed aged areas. However, it is notable how close the
were identified. Three issues, in particular, were estimated median rates are to the rates advocated
addressed by the TEMPUS project. Firstly, for use in economic evaluation in a range of
whether individual inter-temporal preferences countries (for example, 3% in the US, 5% in
with respect to own health differ from those Australia and Canada). Also the degree of con-
with respect to the health of others. Secondly, vergence between rates for own and others’
the impact of different methods of eliciting health and between the open-ended and closed-
preferences on apparent inter-temporal prefer- ended approach is striking.
ences. Thirdly, whether or not individuals value The over-representation of more highly edu-
future health events in line with the traditional cated groups is of less concern when consider-
discounted utility model. The design of the ing the three main research issues addressed by
Eliciting Time Preferences for Health 243
TEMPUS. Implied discount rates with respect to R&D Health Technology Assessment Pro-
own health and others’ health were shown to be gramme (project 94/35/1).
similar. An open-ended method elicited higher
discount rates than a closed-ended method.
Finally, some preliminary evidence for decreas- REFERENCES
ing timing aversion was presented, specifically
that implied discount rates are a function of the
period of delay. 1. Hausman JA. Individual discount rates and the
There are limitations inherent in the stated purchase and utilisation of energy-using durables.
preference approach – specifically there are con- Bell J. Economics 1979; 10: 33–54.
cerns over the validity of responses compared to 2. Gately D. Individual discount rates and the pur-
values derived from observing behaviour. How- chase and utilisation of energy-using durables:
ever, it is difficult to observe inter-temporal comment. Bell J. Economics 1980; 11: 373–4.
choices in a healthcare context and then to 3. Dreyfus MK, Viscusi WK. Rates of time prefer-
ence and consumer valuations of automobile
control for the large number of potential influ-
safety and fuel efficiency. J. Law Economics
ences. The use of postal questionnaires was
1995; 38: 79–105.
driven largely by the desire to obtain reasonably
4. Kooreman P. Individual discounting, energy con-
large sample sizes. Such an approach offers less servation, and household demand for lighting.
opportunity to establish whether or not the Resource and Energy Economics 1996; 18:
respondent understands the question; moreover, 103–14.
the questions are pre-determined and do not 5. Lang K, Ruud PA. Returns to schooling, implicit
give an opportunity for the questions to be discount rates and black-white wage differentials.
responsive to the previous answers. Rev. Economics Statist. 1986; 69: 41–7.
Regarding future research, the topics identi- 6. Lawrance EC. Poverty and the rate of time prefer-
fied above remain relevant. A single, albeit ence: evidence from panel data. J. Political Econ-
multifaceted, project such as TEMPUS can add omy 1991; 99: 54–77.
significantly to our understanding, but cannot by 7. Viscusi WK, Moore MJ. Rates of time preference
itself resolve the outstanding research issues. and valuations of the duration of life. J. Public
Thus the three themes addressed in this paper: Economics 1989; 38: 297–317.
the influence on estimates of implied discount 8. Moore MJ, Viscusi WK. The quantity-adjusted
rates of different methods of eliciting prefer- value of life. Economic Inquiry 1988; 26: 369–88.
ences; the nature of the inter-temporal choice; 9. Moore MJ, Viscusi WK. Discounting environ-
and the underlying model of time preferences, mental health risks: new evidence and policy
all require further investigation. Further issues implications. J. Environ. Economics Management
concern the use of self-completed questionnaires 1990; 18: S51–S62.
versus interviews (face-to-face and telephonic), 10. Thaler R. Some empirical evidence on dynamic
and the presence and impact of framing effects. inconsistencies. Economics Lett. 1981; 8: 201–7.
Also, there is considerable scope for the invest- 11. Fuchs VR. Time preference and health: an
igation of the nature and extent of differences exploratory study. In: Fuchs VR (ed.). Economic
between time preferences for saving lives and Aspects of Health. Chicago: The University of
those for non-fatal changes in health; and Chicago Press, 1982.
whether time preferences are to some extent 12. Benzion U, Rapoport A, Yagil J. Discount rates
health state-specific. The empirical investigation inferred from decisions: an experimental study.
of alternative models of time preference, in Management Science 1989; 35: 270–84.
particular, models which allow for decreasing 13. Myerson J, Green L. Discounting of delayed
rewards: models of individual choice. J. Exp.
timing aversion, such as proportional and hyper-
Anal. Behav. 1995; 64: 263–76.
bolic discounting models, is at an early stage in
14. Houston DA. Implicit discount rates and the pur-
the economics literature.
chase of untried, energy-saving durable goods. J.
Consumer Res. 1983; 10: 236–46.
15. Sultan F, Winer RS. Time preferences for products
ACKNOWLEDGEMENTS and attributes and the adoption of technology-
driven consumer durable innovations. J. Economic
Psychol. 1993; 14: 587–613.
HERU is funded by the Chief Scientist Office 16. Loewenstein GF, Prelec D. Anomalies in inter-
of the Scottish Executive Health Department temporal choice: evidence and an interpretation.
(SEHD). The views expressed in this paper Q. J. Economics 1992; 107: 573–97.
are those of the authors and not SEHD. This 17. Parsonage M, Neuberger H. Discounting and
study received financial support from the NHS health benefits. Health Economics 1992; 1: 71–6.
244 The Advanced Handbook of Methods in Evidence Based Healthcare
18. Cairns JA. Discounting and health benefits: 34. Cairns JA, van der Pol M. Constant and decreas-
another perspective. Health Economics 1992; 1: ing timing aversion for saving lives. Soc. Sci.
76–9. Med. 1997; 45: 1653–9.
19. Department of Health. Policy Appraisal and 35. Enemark U, Lyttkens CH, Tröeng T, Weibull
Health. London: HMSO, 1995. H, Ranstam J. Time preferences of vascular
20. Torrance GW, Blaker D, Detsky A, Kennedy W, surgeons. Medical Decision Making 1998; 18:
Schubert F, Menon D, Tugwell P, Konchak R, 168–77.
Hubbard E, Firestone T. Canadian guidelines 36. Lipscomb J. The preference for health in cost-
for economic evaluation of pharmaceuticals. effectiveness analysis. Medical Care, 1989; 27:
PharmacoEconomics 1996; 9: 535–59. S233–S253.
21. Gold MR, Siegel JE, Russell LB, Weinstein MC. 37. Cairns JA. Health, wealth and time preference.
Cost-effectiveness in health and medicine. Oxford: Project Appraisal, 1992; 7: 31–40.
Oxford University Press, 1996. 38. Chapman GB, Elstein AS. Valuing the future:
22. Gafni A, Torrance GW. Risk attitude and time temporal discounting of health and money. Medi-
preference in health. Management Science 1984; cal Decision Making 1995; 15: 373–86.
30: 440–51. 39. Chapman GB. Temporal discounting and utility
23. Gafni A. Time in health: can we measure individ- for health and money. J. Exp. Psychol. [Learn.,
uals’ pure time preferences? Medical Decision Mem. Cogn.] 1996; 22: 771–91.
Making 1995; 15: 31–7. 40. Redelmeier DA, Heller DM. Time preference in
24. Loewenstein GF, Prelec D. Preferences for medical decision making and cost effectiveness
sequences of outcomes. Psychol. Rev. 1993; 100: analysis. Medical Decision Making 1993; 13:
91–108. 212–17.
25. Krahn M, Gafni A. Discounting in the economic 41. MacKeigan LD, Larson LN, Draugalis JR, Boot-
evaluation of health care interventions. Medical man JL, Burns LR. Preferences for health gains
Care 1993; 31: 403–18. versus health losses. PharmacoEconomics 1993;
26. Horowitz JK, Carson RT. Discounting statistical 3: 374–86.
lives. J. Risk Uncertainty 1990; 3: 403–13. 42. Dolan P, Gudex C. Time preference, duration and
27. Cropper ML, Aydede SK, Portney PR. Discount- health state valuations. Health Economics 1995;
ing human lives. Am. J. Agric. Economics 1991; 4: 289–99.
73: 1410–15. 43. Loewenstein G. Anticipation and the valuation
28. Cropper ML, Aydede SK, Portney PR. Rates of of delayed consumption. Economic J. 1987; 97:
time preference for saving lives. Am. Economic 666–84.
Rev. 1993; 82: 469–72. 44. Harvey, CM. The reasonableness of non-constant
29. Cropper ML, Aydede SK, Portney PR. Prefer- discounting. J. Public Economics 1994; 53:
ences for life saving programs: how the public 31–51.
discount time and age. J. Risk Uncertainty 1994; 45. Ainslie, G. Derivation of rational economic
8: 243–65. behaviour from hyperbolic discount rates. Am.
30. Johannesson M, Johansson P-O. The discounting Economic Rev. 1991; 81: 334–40.
of lives saved in future generations – some 46. Dolan P. Modelling valuations for health states:
empirical results. Health Economics 1996; 5: the effect of duration. Health Policy 1996; 38:
329–32. 189–203.
31. Olsen JA. Time preferences for health gains: an 47. McKelvey R., Zaviona W. Statistical Model for
empirical investigation. Health Economics 1993; the Analysis of Ordinal Level Dependent Vari-
2: 257–65. ables. J. Math. Sociol. 1975; 4: 103–20.
32. Cairns JA. Valuing future benefits. Health Econ- 48. van der Pol M, Cairns JA. The discounted utility
omics 1994; 3: 221–9. model versus a hyperbolic discounting model.
33. Cairns JA, van der Pol M. Saving future lives: a Paper presented at the Health Economists’ Study
comparison of three discounting models. Health Group Meeting, University of Birmingham 6–8th
Economics, 1997; 6: 341–50. January, 1999.
Appendix 1 Options and criteria
Other factors than time preference
Degree of difficulty to answer Degree of realism Computation of discount factor influence response
Number of points in time Having to consider only two A profile is more realistic Discount factors are more easily
Two points points in time is easier. whereby ill-health is spread computed when only two points
Profile over a couple of years instead in time are considered. Discount
of being concentrated in one factors can, in some case, not
particular year. be computed from profiles.
Timing versus quantity It is easier to choose when Choosing how ill or how The choice of timing versus With giving a choice of timing it
Timing something takes place than to healthy is less realistic. Timing quantity does not have an is more likely that other factors
Quantity choose how much health or ill- is more realistic (treatments can influence on how difficult or will influence the response.
245
more difficult. limited period is less complex. consider a long remaining life.
246 The Advanced Handbook of Methods in Evidence Based Healthcare
Appendix 2 Example of open-ended question If the ill-health would then start 4 years from now,
what is the maximum number of days of ill-health that
Imagine the following ill-health: You have some would still make the treatment worthwhile?
problems with performing your usual activities (e.g.
work, study, housework, family or leisure activities) Appendix 3 Example of closed-ended question
and you have moderate pain or discomfort. You
have no problems in walking about, nor with washing Imagine the following ill-health:
and dressing yourself, and you are not anxious or
You have some problems with performing your
depressed.
usual activities (e.g. work, study, housework,
Imagine that you will be ill, as described above, family or leisure activities) and you have moder-
starting 2 years from now for 20 days. There is a ate pain or discomfort. You have no problems in
minor one-off treatment available that will postpone walking about, nor with washing and dressing
this spell of ill-health to a point further in the future. yourself and you are not anxious or depressed.
For instance, the treatment could have the following
effects: your period of ill-health would start 9 years
Imagine that you will be ill (as described in the box
from now instead of 2 years from now; and you would
above). There are two alternative treatments (A and
then be ill for 30 days instead of 20 days. B) available. The effects of the alternative treatments
You might think this treatment is a good idea: the vary with regard to when the illness will occur and
advantage of postponing the ill-health outweighs the how long you will be ill for (you cannot be cured
disadvantage of being ill for a longer period. Or you completely). Below is an example of how they can
might think the treatment is not worthwhile: you do vary. With treatment A you will be ill starting 2 years
value the postponement but the advantage of this is from now for 20 days, and with treatment B you will
outweighed by the disadvantage of being ill for a be ill starting 6 years from now for 48 days. Assuming
longer period; or you might simply prefer to be ill 2 everything else about the treatments is the same (i.e.
years from now instead of 9 years from now. Imagine severity of the treatment, side-effects, costs), which
again that you will be ill starting 2 years from now for treatment would you prefer?
20 days, and that treatment is again available that will
postpone this spell of ill-health. ● Example Which
We will now ask you to state the maximum number treatment
When you How long would you
of days of ill-health that would still make the treat-
are ill for prefer?
ment worthwhile for you. For example, say that the
treatment can postpone the period of ill-health to 6 in 2 years for
years in the future. If the number of days of ill-health Treatment A
time 20 days
in that year would be zero everyone would probably
in 6 years for
choose the treatment. As the number of days of ill-
health in that year increases individuals would at
Treatment B
time 48 days ✓
some point no longer prefer to be treated. What we are
interested in is the maximum number of days of ill- No preference
health at which you would still choose to be treated.
15
The Conduct and Design
of Questionnaire Surveys in
Healthcare Research
E L A I N E MC C O L L , A N N J A C O B Y , L O I S T H O M A S ,
JENNIFER SOUTTER, CLAIRE BAMFORD,
NICK STEEN, ROGER THOMAS, EMMA HARVEY,
A N D R E W G A R R A T T and J O H N B O N D
WHY THIS REVIEW WAS NEEDED The Readership for this Review
Questionnaire surveys are frequently the method This review of best practice in question-
of choice for gathering primary quantitative data naire design and survey administration is aimed
248 The Advanced Handbook of Methods in Evidence Based Healthcare
primarily at researchers and practitioners using Defining the Scope of the Review
questionnaires in the evaluation of health tech-
nologies, that is ‘. . . all the methods used by An operational definition of
health professionals to promote health, to pre-
vent and treat disease, and to improve rehabilita-
‘questionnaire’
tion and long-term care’.6 We recognise that We defined ‘questionnaire’ as ‘a structured
questionnaires are also widely used in other schedule used to elicit predominantly quantita-
fields of healthcare research and in research on tive information, by means of direct questions,
topics unrelated to health. Our findings and from respondents, either by self-completion or
recommendations will be of general interest to via interview’. We chose this deliberately broad
all researchers for whom questionnaire surveys definition because we recognised that inter-
are the chosen method of data collection. None- viewer administration may be the most appro-
theless, in synthesising the findings and making priate mode of data collection in certain cir-
recommendations, we distinguished between cumstances. However, we also accepted that,
evidence from studies on health-related topics because of resource constraints, self-completion
and those on other subjects. We did this because questionnaires are likely to be the method of
there is empirical evidence that patterns of choice in many health surveys. Our review
response to surveys on topics perceived to be therefore commenced with an examination of
important and relevant to respondents (such as the relative strengths and weaknesses of inter-
health) are significantly different from those for viewer administration and self-report. Through-
out the review, in synthesising and interpreting
surveys of general populations, or on topics
evidence from primary studies and from pre-
perceived as less personally relevant.7,8 There vious reviews of the literature, we distinguished
are also likely to be ethical concerns particular between different modes of administration and
to health surveys. sought to highlight the extent to which findings
from one mode can be extrapolated to another.
stimulate response rates to self-completion ques- the public and professionals, reflected in a
tionnaires were also effective in interviewer- downwards trend in response rates for health
administered surveys. professionals.21 Without doubt, recommenda-
In designing some experiments, scant atten- tions made on the basis of which methods of
tion was paid to the complexity of the survey questionnaire design and administration are
process, in particular the way in which a number effective today will need to be reviewed in the
of factors interact to influence respondent future to take account of these changes.
behaviour.18 Ignoring likely interactions led, in
some cases, to inappropriate and unrealistic
experiments (such as a highly personalised letter
with an assurance of anonymity!). KEY ISSUES IN SURVEY DESIGN AND
Personal contact Time and effort required Actual difficulty encountered Provision of SAE
personalisation of letter length of questionnaire clarity of question
personalisation of size of pages wording
envelope supply of addressed clarity of instructions
class of mail return envelope complexity of questions
supply of stamped
return envelope
Questionnaire appearance Cursory evaluation of Sensitivity of requests Reminders to return
cover illustration difficulty number and nature of
colour of cover number of questions sensitive questions
layout and format complexity of questions
quality/clarity of type
Topic Actual time required
questionnaire title
cover illustration
content of cover letter
timeliness
relevance/salience
Source credibility/trust
image of sponsor
credentials of individual
investigator
message in cover letter
Reward for participation
tangible rewards;
monetary and other
incentives
intangible rewards;
appeals to altruism,
self-interest, etc.
Persistence of source
follow-up procedures
Finally, ensuring that the target respondent, collect more data than is strictly necessary,
and not a relative or carer, actually completes which wastes resources and may be unethical.
the questionnaire may also be a problem, par- Data of greater complexity can also be gath-
ticularly in postal surveys. ered in face-to-face interviews, because visual
aids can be used and the interviewer can more
readily provide explanations and probe and
Volume, Complexity and Nature of prompt for additional information. However,
Data Collected self-completion questionnaires are generally held
to be superior to telephone surveys for collecting
The desired volume, complexity and nature of complex data. For example, questions involving
data to be collected will vary from survey to the ranking of several items, or those using mul-
survey. It has been asserted11 that the largest tiple response categories (for instance, a seven-
volume of data can be collected in face-to- point scale of satisfaction) are difficult to use in
face interviews (since the burden of recording telephone surveys because there is no visual
responses falls on the interviewer rather than the stimulus; instead, the respondent must mentally
respondent). The downside is the temptation to retain and process the items or categories.
Conduct and Design of Healthcare Surveys 253
Response rates
General samples Good Good Poor to Good
Specialised samples Good Good Satisfactory to Good
Representative samples
Avoidance of refusal bias Good Good Poor
Control over who completes the questionnaire Good Satisfactory Satisfactory to Good
Gaining access to the selected person Satisfactory Good Good
Locating the selected person Satisfactory Good Good
Ability to handle:
Long questionnaires Good Satisfactory Satisfactory
Complex questions Good Poor Satisfactory
Boring questions Good Satisfactory Poor
Item non-response Good Good Satisfactory
Filter questions Good Good Satisfactory
Question sequence control Good Good Poor
Open-ended questions Good Good Poor
Quality of answers
Minimise social desirability responses Poor Satisfactory Good
Ability to avoid distortion due to:
Interviewer characteristics Poor Satisfactory Good
Interviewer’s opinions Satisfactory Satisfactory Good
Influence of other people Satisfactory Good Poor
Allows opportunities to consult Satisfactory Poor Good
Avoids subversion Poor Satisfactory Good
Implementing the survey
Ease of finding suitable staff Poor Good Good
Speed Poor Good Satisfactory
Cost Poor Satisfactory Good
Interviews facilitate the use of open-ended or attitudes. Response bias can arise in a number
questions or open-ended probes, where the inter- of ways. ‘Social desirability’ bias is particularly
viewer can record verbatim the answers given likely in respect of questions on sensitive topics,
by respondents. This may generate richer and and occurs when respondents conceal their true
more spontaneous information than would be behaviour or attitudes and instead give an
possible using self-completion questionnaires. answer that shows them in a good light, or is
Although open-ended questions can be used in perceived to be socially acceptable. Related to
self-completion questionnaires, responses are this is ‘sponsorship’ bias, by which respondents’
typically less detailed since respondents tend to answers differ according to who is conducting
be less expansive in writing than in speaking. the survey. ‘Memory’ bias occurs in questions
Collecting data on ‘boring’ topics is easier in involving recall of past events or behaviour and
interviewer-administered surveys; interviewers can include omission and telescoping (misplac-
can engage subjects’ attention and interest and ing an event in time). Biases can also arise from
can use their powers of persuasion to maximise the respondent mishearing or misunderstanding
participation.31 By contrast, response rates to the question or accompanying instructions.
postal surveys on topics perceived to be un- The relative threat of different types of
important or lacking in relevance are generally response bias depends in part on the mode of
low.32 administration. In surveys using mixed modes
of administration, care is needed in interpreting
Response Bias findings; observed differences between respond-
ents may be attributable to the methods of data
Response bias occurs when the answers pro- collection, rather than to true underlying differ-
vided by respondents are invalid – in other ences in experiences and attitudes. The greater
words, they do not reflect their true experiences anonymity afforded by a postal questionnaire,
254 The Advanced Handbook of Methods in Evidence Based Healthcare
and to a lesser extent by a telephone interview, from restrictions caused by the hospital regime
may mean that patients are more likely to report rather than by impaired health.33
truthfully on potentially embarrassing behaviour
or experiences (for example, mental health prob-
lems). Social desirability bias is more likely in Practicalities
interviewer-administered surveys, since respond-
ents may wish to ‘save face’. By contrast, other Interviewer-administered surveys are generally
forms of deception can be minimised by inter- held to be more expensive than postal sur-
views, because some responses can be verified veys.11,12 The additional expense derives mainly
by observation. from the costs of training and paying inter-
Questionnaire wording and design can also viewers and of their travel costs. Face-to-face
induce response bias, for example through ques- interviews also generally require an initial con-
tion sequencing effects (where the response tact (by letter, telephone or in person) to set up
given to a particular question differs according an appointment for the interview; multiple con-
to the placement of that question relative to tacts may be needed to agree a convenient time.
others), and the labelling and ordering of Although multiple contacts may also be required
response categories. to achieve a telephone interview, the elimination
of travel costs and time can mean that they are a
low cost and speedy method of data collec-
Other Sources of Bias tion.11,34 However, costs increase with the num-
ber of attempts required to contact subjects who
A significant disadvantage of interviewer- are not available at the first call, and with the
administered surveys is that the interviewers number of long-distance calls required. Whilst
themselves can introduce errors both in a ran- the costs of all interview surveys increase sig-
dom and a systematic way.11 Random errors are nificantly with geographical dispersal of sample
more likely to be due to interviewer inaccuracy, members, postal surveys can cover scattered
for example recording answers incorrectly or populations without increasing survey costs.
altering the wording of a question by mistake. Similarly, in ‘captive audience’ self-completion
Examples of systematic effects are: selective surveys (for example, administered to a group of
recording of subjects’ responses; differences in patients in a waiting room), data can be col-
the extent to which interviewers probe for a lected simultaneously from a large number of
substantive response or accept ‘don’t know’ respondents. Telephone interviews31 and postal
answers; and consistent re-wording of questions. surveys allow survey researchers to access areas
Systematic bias can occur even when there is where safety might be compromised, such as
only one interviewer, if the interviewer does not inner cities.
accurately record the respondent’s answers, but Interviewer-administered surveys are gener-
instead is consistently selective in what he or ally more appropriate when a large number of
she records. It is an even greater problem where open-ended questions are required. However,
a survey requires multiple interviewers; coding responses to such open-ended questions
observed differences across respondents may be can be both time-consuming and costly. As a
an artefact of the way in which different inter- result, interviewer-administered surveys may
viewers have posed questions or recorded take longer to produce results than postal sur-
answers, rather than an indication of true under- veys. Balanced against this, however, there is no
lying differences between respondents. The need to wait for questionnaires to be returned
advent of computer-assisted personal interview- through the post; if two rounds of reminders are
ing (CAPI) may prevent routine errors and used, the interval between dispatch and return of
omissions, but at present requires prohibitively questionnaires can be as much as 12–16 weeks.
high levels of investment for many survey In interviewer-administered surveys, the speed
researchers. with which data are available for analysis can be
Personal characteristics of the interviewer, increased by the use of computer-assisted tech-
such as age, gender, social class, race or level of niques, which minimise the need for subsequent
experience and training may also affect both data entry and checking.
response rates and the nature of the responses Face-to-face interviews generally require
given. Similarly, the setting in which data is highly trained and motivated interviewers, since
collected may affect responses. For example, it they must work autonomously. Less experienced
has been shown that responses to physical func- interviewers may be used in telephone surveys,
tion items on a health status measure may be since stricter control and closer supervision is
confounded by place of administration, with possible.31 This greater control can also reduce
higher scores for hospital in-patients resulting inter-interviewer variability. Postal surveys
Conduct and Design of Healthcare Surveys 255
require the fewest personnel and need minimal Computer-assisted versus self-completion
equipment. questionnaires
Four comparative studies were found;48–51 gen-
eralisability is limited by the fact that all were
BEST PRACTICE IN SURVEY RESEARCH: on non-health topics. Findings with respect to
EXPERT OPINION AND EMPIRICAL EVIDENCE response rates, response quality and answers to
sensitive questions were equivocal. However,
Mode of Administration evidence from a single study48 suggested that
respondents to computerised questionnaires may
use a wider range on rating scales. Quicker
Telephone interviews versus postal surveys responses may also be obtained using computer-
In three identified studies, all on health-related assisted questionnaires.48
topics,35–37 higher response rates were obtained
for telephone surveys; in a fourth38 (also health- Computer-assisted telephone interviewing
related), overall response rates to the postal (CATI) versus conventional telephone
survey were significantly higher. In this latter interviewing
study, however, rates of item non-response
(question omission) were higher for postal We identified one health-related study on this
administration. These studies demonstrated no topic.52 Response rates for the non-CATI group
consistent advantage of one approach over the were significantly higher, and the time taken to
other with respect to non-response bias, quality complete a CATI was longer. On most of the
of response, anonymity or cost. It is, however, criteria measured, only small differences were
important to note that all the identified studies found between the two approaches, but there
were carried out in the USA, where telephone was some evidence of less interviewer variabil-
ownership rates are higher. In the UK, telephone ity in CATI.
surveys may suffer from non-response bias
against the young, poor, ethnic minorities and Question Wording and Sequencing
the domestically mobile, since these groups are
most likely to be among the 4% of the popula- Some caution is required in interpreting the
tion who do not have a telephone. available evidence on the effects of question-
naire wording and sequencing, and on the
Face-to-face interviews versus presentation of response categories, in terms of
the validity and quality of the data obtained.
self-completion questionnaires Both theory and limited empirical evidence sug-
We identified six relevant studies,35,39–43 five of gest that these effects may vary with mode of
which were on health-related topics. Evidence administration, but the majority of empirical
from two studies35,42 supported expert views that studies involve interview approaches. Further-
face-to-face interviews tend to yield higher more, few identified studies were in the context
response rates. Evidence from a single study39 of health surveys; those that were health-related
suggested that respondents may be more likely are highlighted below.
to give no answer or to say ‘don’t know’ in an
interview. However, there was no unequivocal Question wording
evidence to support the view that postal surveys
Schuman and colleagues53 found that open-ended
yield more truthful responses to sensitive ques-
questions produce more non-common responses
tions or elicit more critical or less socially than closed questions, but most additional
acceptable responses. response categories are small and miscellaneous.
Therefore, use of either question form will ordi-
Telephone versus face-to-face interviews narily lead to similar conclusions. However,
open-ended questions are preferable in develop-
Four comparative studies were identified,44–47 ment stages and pilot studies, to generate appro-
three of which were on health topics. Although priate response categories for closed questions.
one non-health study45 found telephone inter- Bishop and co-workers54 compared the effects
views to be quicker, there was no consistent of presenting one versus two sides of an issue in
evidence of the relative superiority of either survey questions and demonstrated that giving a
mode in terms of overall response rates, item second substantive choice on attitude questions
non-response rates or the elicitation of sensitive increased the likelihood of respondents express-
information. ing an opinion.
256 The Advanced Handbook of Methods in Evidence Based Healthcare
Researchers have shown that survey respond- It has been argued that question order effects
ents employ a wide range of cognitive pro- are less common in self-completion question-
cesses in formulating responses to behavioural naires than in interviewer-administered sur-
frequency questions, including both episodic veys.63 This may be the result of reduced serial-
enumeration (recalling and counting specific order constraints (for example, in interviewer-
instances on which the behaviour occurred) and administered surveys, respondents may wish to
rate processing (aggregating from the ‘normal’ present a consistent image to the interviewer64).
rate at which the behaviour takes place in a unit Moreover, in self-completion questionnaires,
of time, such as a week). Task conditions, such respondents have the opportunity to read all the
as time-framing, have been shown to influence questions before responding, and thus have a
the processes employed, both in health-related55 better feel for the relationship between ques-
and other surveys.56 tions; they are not forced by time constraints to
One form of response bias is the acquiescence give ‘top-of-the-head’ responses.65 However,
effect (the tendency to agree with questions research64 which tested whether specific ques-
presented in the form of statements); this is most tion order effects, previously observed in inter-
common amongst less-educated respondents.54 view surveys, occurred also in postal surveys
Conventional wisdom suggests that presenting a was inconclusive.
mixture of positively and negatively phrased The content of opening items may influence
statements will control or offset acquiescence subjects’ motivation to complete a question-
tendencies; however, the actual effect of mixing naire; the more interesting and relevant these
statements in this way may be to reduce items are, the greater the likelihood of
response validity.57 response.66 Findings by Jones and Lang67 also
The wording of filter questions, asking whether suggested that topic ordering within a ques-
the respondent has knowledge of or has thought tionnaire may differentially affect response rates
about an issue, can significantly affect the per- amongst groups holding different attitudes, thus
centages of ‘don’t know’ responses elicited at increasing non-response bias.
a subsequent substantive question, particularly It is usually recommended4,68 that general
for topics less familiar to the respondent. Con- questions should precede more specific ones,
versely, the content of the question can have an since specific questions may create a ‘saliency’
important independent effect on ‘don’t know’ effect which influences responses to more gen-
responses, regardless of the filter wording. The eral questions.68 Empirical evidence supports
use of filter questions can thus alter the conclu- this view. Prior items in a questionnaire have
sions drawn.58,59 been shown to exert a ‘carry-over’ effect by
Interpretation of questions that include pres- priming respondents about their beliefs/attitudes
tige names is complicated by the fact that sub- towards a particular topic.69,70 Similarly, the
jects respond not only on the basis of the content meaning of a potentially ambiguous question
of issues but also on the basis of the names.60 may be clarified by the wording of a preced-
ing question, through a ‘contrast’ effect,62 as
demonstrated by Colasanto and colleagues71 in a
Question sequencing study on beliefs about transmission of AIDS and
Question order effects may influence overall HIV.
response rates; they may also be a significant Expert opinion on the relative placement of
source of response bias. The direction and questions on related topics varies. Moser and
strength of these effects can vary with survey Kalton1 equivocate and suggest that question-
topic, the context in which the questions are naires may be designed so that earlier questions
presented and the study population. Context (or groups thereof) set the context for later ones,
effects appear to be especially likely when or conversely so that responses to later questions
researchers attempt to summarise complex are not influenced by those preceding them. By
issues in a single general item.61 It seems that contrast, Sudman and Bradburn4 recommend
question order effects are consistent across gen- that all questions on one topic should be com-
der and education levels, and so are as much of pleted before embarking on a new topic. The
a concern in surveys of restricted populations as view that questions should be grouped by topic,
in those of general populations.62 and topics ordered so that related ones are
Ordering effects may bias estimates of the adjacent to one another, ignores the issue of
prevalence of attitudes and behaviour. If other- context effects; yet empirical evidence suggests
wise identical questions are posed in a different that such effects are common. However, evi-
order or a different context across surveys, dence from studies on both health-related72 and
apparent differences in response patterns may other topics69,70 indicates that ‘buffering’ (inter-
reflect these context effects rather than true posing unrelated items between questions
differences between settings or over time.62 known or thought likely to influence one
Conduct and Design of Healthcare Surveys 257
another) is not guaranteed to completely elim- first response category presented) and ‘recency’
inate context effects. Researchers therefore need effects (the tendency to choose the last response
to balance the risk of context effects with the option). Schuman and Presser62 assert that
desirability of coherence and continuity. primacy effects are more common in self-
completion questionnaires (a view supported by
Response categories empirical evidence64) and that recency effects
are more usual in interviews.
Sudman and Bradburn4 suggest that the distinc- In pre-coded (closed) questions, the ‘codes’ or
tion between question wording and response ‘labels’ attached to each response category may
format is to some extent artificial, since the form also be seen as part of the question–response
of the question often dictates the most appro- stimulus. One identified study77 compared num-
priate response format. erical and verbal labelling of response categories
Nonetheless, the question of whether to and showed no significant differences. However,
include or exclude ‘no opinion’, ‘don’t know’ or findings by the same authors suggested that fully
middle response categories is one which has defined scales (labelling all scale points) may
been considerably debated. It has been argued act as a check on leniency errors (the tendency
that inclusion of such options, especially in self- to bias ratings in an upward direction). In
completion questionnaires, may lead respondents another study,78 however, there was no sig-
to take the easy way out rather than spending nificant differences in response patterns to scales
time formulating an opinion or recording factual in which only the end-points were explicitly
details, thereby yielding invalid data. The labelled and those in which all scale points were
counter-argument is that their exclusion encour- labelled to indicate equal spacing. These con-
ages respondents to make wild guesses or to flicting findings suggest that the wording of
omit the question altogether, presenting different response alternatives may be as critical as ques-
threats to data completeness and quality. Em- tion wording.
pirical evidence,59,64,73,74 including findings from
one health-related study,75 suggests that response
distributions vary according to whether or not Response formats
middle alternatives and explicit ‘don’t know’
Most experts3,4,19,79 recommend a vertical answer
response categories are included. It appears that
format (Figure 1) for multiple choice questions,
offering such response options is generally
except in respect of rating scales. In the latter
preferable for attitudinal questions. Providing
case, they favour a horizontal format (Figure 2),
respondents with an explicit opportunity to have
both to save space and to avoid the erroneous
no opinion may also avoid spurious representa-
impression that each category is independent of
tiveness.59,73 Middle alternatives and ‘don’t
each other.80
know’ response categories may be less important
for factual questions.
The order of response alternatives may affect
both distribution of responses to individual
items and associations between these and other
items. Israel and Taylor76 have shown this effect
to be particularly marked in respect of questions
prone to social desirability bias; the percentage
endorsing the ‘socially acceptable’ response was
significantly higher when this was presented
first.
Two specific types of response order effects Figure 1 Example of vertical format for
are ‘primacy’ effects (the tendency to select the factual questions.
not been shown to have a significant impact ups enhances response rates.99–101 Including a
on response rates.91 However, guidelines on duplicate questionnaire with the first reminder
enhancing the readability of documents, pro- does not appear to have a significant impact, but
duced by the United Kingdom based Basic inclusion of a replacement questionnaire with a
Skills Agency (whose remit is assisting those second reminder appears to be effective.102 Con-
with literacy problems) and the Royal National trary to the recommendations of Dillman3 and
Association for the Blind, indicate that dark the conclusions from an earlier review,32 we
backgrounds (particularly shades of blue and found no conclusive evidence that special mail-
purple) should be avoided; black type on a white ing techniques for final reminders are superior to
or yellow background provides the best contrast. standard mailing.103,104 One health-related study
These organisations recommend the use of a suggested that postcard reminders appear to
clear and distinct type face, with a minimum be as effective as letters and are generally
font size of 12 points (larger if readers are likely cheaper.105
to have visual impairment). They advise against Findings both from primary studies94,95,104,106–113
excessive use of upper-case letters, suggesting and from previous reviews7,8,27,32,114,115 show no
that emphasis is more appropriately provided by consistent advantage of class of mail, or of
underlining or emboldening; such formatting stamped envelopes over franked or reply-paid
may also be used to distinguish questions, envelopes, either for outgoing or return mail.
answers and instructions from each other.
General motivational factors
Enhancing Response Rates
Anonymity and confidentiality are frequently
Factors influencing survey response rates may confused by survey researchers, but they are not
be conveniently divided into mechanical and equivalent.116 Under conditions of anonymity,
perceptual factors (timing of survey; number, no identification appears on the questionnaire,
timing and method of contacts; postage rates and and it is not possible to link individual responses
types); general motivational factors (anonym- to a specific named person. As a result, remin-
ity and confidentiality; personalisation; nature ders cannot be targeted specifically to initial
of appeal; other aspects of covering letters; non-respondents. Under conditions of confiden-
sponsorship; saliency); and financial and other tiality, an identification code (usually a number)
incentives. appears on the questionnaire; researchers can
link this code to a named individual, but individ-
ual responses cannot be attributed to a specific
Mechanical and perceptual factors person by anyone who does not have access to
Evidence from a single, health-related, study of the link between code and name. Ethical prin-
day of posting92 found no significant effect on ciples also require that respondents’ answers are
response rates. Expert advice3 and evidence not revealed to a third party without their
from a series of non-health surveys18 suggest explicit permission, and that results are presented
that month of posting may affect response rates in such a way that individual respondents cannot
(in particular, December should be avoided); be identified. Evidence from the five studies (one
however, it is unclear whether this effect is on a health-related topic) we identified117–121
universal or is population- or topic-specific. failed to provide consistent evidence of the
One of the most potent ways of enhancing effects of anonymity and confidentiality on
response rates is through multiple contacts.32,67,93 response rates or data quality.
Both pre-notification and follow-up contacts (re- Despite Dillman’s3 recommendation regarding
minders) have been found to be effective.7,94,95 the personalisation of covering letters ‘to show
We found one non-health study which sug- regard for respondents’, only two (both non-
gested that a letter may be more effective than a health)104,122 out of 11 studies94,95,104,108,119,122–128
telephone call in pre-notifying subjects.96 How- showed a significant positive effect of personal-
ever, the findings from three studies (one on a isation on response rates; in neither health-
health-related topic) with respect to ‘high related study123,128 did personalisation sig-
involvement’ methods of pre-notification (such nificantly enhance response rates. However,
as ‘foot-in-door’ approaches, whereby subjects it appears that personalisation may interact
are first asked to comply with a small task, in with such factors as the nature of the appeal
the hope that they will continue to comply with made in the covering letter and assurances of
the larger task of survey participation) were confidentiality.129
equivocal.87,97,98 We found some evidence130 that traditional-
Nor did we find any conclusive evidence from style letters are more effective than novel ap-
three identified studies (one health-related) that proaches. However, only one108 out of three
a ‘threat’ in the reminder letter of further follow- identified studies108,131,132 suggested that the
260 The Advanced Handbook of Methods in Evidence Based Healthcare
characteristics of the signatory affect response our own extensive experience of conducting
rates. Nor does response rate appear to be surveys. Where recommendations are not evi-
positively related to hand-written signatures131 dence based, we highlight this fact.
or colour of ink.132 Although a ‘social utility’
appeal in the covering letter (emphasising the
value of the response to society in general) has
Mode of Administration
been advocated,3 empirical evidence from ten
Empirical evidence35–52 does not demonstrate
studies (four of which were health-related) indi-
any consistent advantage of any single method
cates that no single type of appeal offers a con-
of questionnaire administration over all others.
sistent advantage;67,107,113,123,124,131,133–136 it seems
The choice of mode of administration should
that the appeal should be matched to the antici-
therefore be made on a survey-by-survey basis,
pated motivations of the recipients.
taking into account: study population; survey
Specifying that the questionnaire will only
topic; sampling frame and method; volume and
take a short time to complete has been shown to
complexity of data to be collected; and resources
be effective in stimulating responses.20 How-
available.
ever, although specification of a deadline for
responding appears to increase the speed of
response (and thereby reduce the number of Question Wording and Sequencing
reminders needed), it may have no effect on
overall response rates.99,123 Theories of respondent behaviour4,5,62 and find-
A ‘salient’ topic is generally effective in ings from empirical studies53–56 suggest that
enhancing response rates.32,137,138 The impact researchers aiming to increase the validity and
of sponsorship appears to be situation- and reliability of respondents’ answers should con-
location-specific.67,86,107,117,139 sider the range of cognitive processes involved
in response formulation and the potential impact
Incentives of task variables such as: the likely salience and
temporal regularity of events; method of survey
In general surveys, incentives are usually a administration; and question design issues such
highly effective means of increasing response as time-frame.
rates.109,135,140–145 However, the benefit of incen- In the absence of any strong evidence to the
tives in health surveys (where response rates are contrary, general principles of questionnaire
typically higher in any case) is less clear- wording (Figure 4) should be maintained. More-
cut.133,146–152 Financial incentives may be more over, Sudman and Bradburn’s4 guidelines for
effective than non-monetary incentives of similar the wording of specific types of questions, the
value.140 Regardless of setting and topic, enclosed presentation of response categories and question
incentives generally generate higher response ordering (Figures 5–10) have stood the test of
rates than promised incentives.141,142,146,151,153–155 time and, in many instances, are supported
Findings on the effects of different sizes of by empirical evidence.53–78 Avoid the use of
incentive were equivocal.141,152 questions involving prestige names60 wherever
Although a number of experts3,156,157 have possible.
recommended offering survey participants a Caution should be exercised in the use of
copy of the results, we found little evidence that negatively phrased attitudinal items, specifically
such an offer has a significant effect on response in using a mix of positive and negative items in
rates, at least in general surveys.88,122,137,158 the hope of reducing acquiescence bias.57
Finally, it appears that personal drop-off of Consider carefully the implications of includ-
questionnaires for self-completion may offer ing or excluding filter questions on response
some advantages, but the cost–effectiveness of distributions. Filter questions may be appro-
this approach may be situation-specific.159 priate to identify which respondents should
answer subsequent questions (for example, to
RECOMMENDATIONS FOR USERS OF avoid asking questions about smoking behaviour
of lifetime non-smokers). Filters can also be
QUESTIONNAIRE SURVEYS used to avoid ‘spurious representativeness’58,59
resulting from trying to elicit opinions from
In so far as possible, the recommendations we those who have no views on the subject. Bear in
make below are based on empirical evidence mind also that filter questions and skip instruc-
from the primary research studies identified in tions are less well-handled by respondents to
our review. However, for some aspects of the self-completion questionnaires than they are by
design and conduct of surveys, there was little trained interviewers
such evidence. In these cases, we have drawn on Although the potential for question order
the recommendations of survey experts and on effects may be reduced in postal surveys,63 it is
Conduct and Design of Healthcare Surveys 261
Figure 4 General principles of question wording (After Moser and Kalton1 and
Oppenheim11).
unlikely to be eliminated entirely; general prin- on results, the order should be randomised.
ciples of question ordering should therefore be Because they may be potentially sensitive,
adhered to regardless of mode of administration. demographic questions should be placed at
As experts4 suggest, the questionnaire should the end of the questionnaire, as Sudman and
open with interesting but non-threatening ques- Bradburn recommend.4
tions. If respondents are expected to have Follow expert advice4 and use open-ended
stronger opinions on some survey topics than questions sparingly, particularly in self-
others, the questionnaire should be assembled to completion questionnaires. However, in piloting
reflect these priorities.66 General questions and pre-testing, open-ended questions should be
should precede specific ones.68–70 Since evidence used to ensure that the response categories sub-
from primary research studies69,70,72 suggests sequently used for closed questions adequately
that ‘buffering’ questions are unlikely to elim- represent the likely range of responses. As rec-
inate context effects, the conventional wisdom ommended in key texts,10,12 response categories
of grouping questions by topic should be for closed questions should be mutually exclus-
adhered to. In situations where researchers are ive (unambiguous, not overlapping) and col-
uncertain about the impact of question order lectively exhaustive (all contingencies catered
262 The Advanced Handbook of Methods in Evidence Based Healthcare
for, if necessary by the inclusion of an option of subsequent data coding and editing time and
‘Other, please specify’). costs, a ‘circle the number’ format is preferable
Since the middle response category on to a ‘tick the box’ format, since evidence indi-
attitude/opinion questions does not necessarily cates no difference in response rates or data
represent a position of neutrality, it should be quality.23
included.59,64,73–75 However, for factual ques-
tions, the ‘don’t know’ response can be omitted. Questionnaire Appearance
Since the evidence is inconsistent,77,78 it is prefer-
able to label all response categories rather than In the absence of empirical evidence, a good
only the endpoints. deal of expert opinion makes sound sense and is
Experts3,4,19,79 recommend a vertical response supported by theories of cognition, perception
format (Figure 1) for multiple-choice questions, and pattern recognition.19 The principles shown
except for rating scales. Empirical evidence and in Figure 3 provide appropriate guidelines.
expert opinion19,80 suggests that a horizontal Although we did not identify any literature on
response format (Figure 2) is more appropri- the topic, computer-readable questionnaires
ate for item sets involving the same response (optical character recognition) are likely to
categories throughout, and in rating scales. Con- assume greater importance in the future. Their
sider natural reading style (left to right, and design should take into account the hardware
horizontally oriented) in placing headings and and software to be used. In health surveys, it is
codes for responses. If a ‘remote scale’ format is likely that a significant proportion of the sample
used, the scale should be repeated after every will have some degree of visual impairment.
three or four questions to minimise the rate of Their needs should be borne in mind in design-
inappropriate neutral responses.81 To reduce ing self-completion questionnaires.
264 The Advanced Handbook of Methods in Evidence Based Healthcare
The ‘optimal’ length of questionnaire will only a short response (a number or a few words)
depend upon data requirements, mode of admin- is required.
istration, survey topic and study population. Follow sound design principles19 and use ele-
Although we did not find any evidence of lower ments of brightness, colour, shape and location,
response rates in longer questionnaires on as well as verbal indicators, to ‘steer’ respond-
health-related topics,85,86 it would be rash to ents through the questionnaire. Graphical means,
conclude that there is no upper limit to the for example arrows and boxes, help to indicate
length of questionnaire feasible in health sur- skip patterns. Place instructions and directions at
veys. Principles of design19 and empirical evi- the point where they are required; if a series of
dence160 indicate that it is desirable to avoid questions involves turning a page, it may be
crowding questions or reducing ‘white space’ in necessary to repeat instructions on the new
an effort to reduce apparent length. page.
As Dillman3 suggests, a booklet format with
double-sided printing presents a more pro-
fessional image. To facilitate use of standard Enhancing Response Rates
stationery and to avoid excessive production
costs, use standard-sized paper (A4 folded to A5 The choice of techniques for enhancing response
booklet or A3 folded to A4 booklet, as dictated rates should be informed by consideration of the
by length of questionnaire). likely barriers and motivational factors for each
White paper or a light tint enhance legibility. particular survey topic and study population.
However, coloured covers can facilitate ques- The frameworks presented by Dillman3 and
tionnaire identification. Dillman3 recommends Brown and colleagues18 (Tables 1 and 2) form a
that the front cover should contain the title of useful basis for deliberation. In assessing poten-
the survey, the identity of the organisation carry- tial methods, consider the potential for sample
ing it out and, for self-completion question- composition bias, response bias, item non-
naires, a neutral graphic illustration. The back response effects, and resource implications, as
cover should provide some blank space for well as the likely impact on response rates. The
respondents’ open comments, and should specify
marginal benefits of intensive approaches to
the address of the organisation conducting the
enhancing response rates may be outweighed by
survey (if not on the front cover) and say ‘thank
the marginal costs. Manipulation of a single
you’ to the respondent.
factor is unlikely to prove fruitful.18 Instead,
A font size of at least 12 points should be
used; a larger font size (up to 14 or 16 points, consideration should be given to the total
depending on type face) is desirable if respond- ‘package’ of: questionnaire wording; question-
ents may have some visual impairment. Choose a naire appearance; general motivational factors;
distinct type face. Emboldening, italicisation mechanical and perceptual factors; and financial
and underlining should be used to distinguish and other incentives. The possibility of inter-
questions, response categories and instructions. actions between factors should be taken into
Maintain a consistent format throughout the count and care should be exercised to avoid
questionnaire. As Dillman3 advocates, respond- apparent mismatches.
ents should not be asked to do two things (for Expert opinion3 and empirical evidence18 sug-
example, rating and ranking) in the one ques- gest that, if possible, the month of December
tion. Avoid splitting a question, its associated should be avoided in conducting postal surveys.
response categories and instructions for answer- Depending on the survey topic and the study
ing over two pages. In questions where the list population, avoidance of the peak holiday months
of response categories is too long to fit on a (July and August) may also be advisable.
single page, continue the response categories on Evidence from primary studies and previous
a facing page if possible; otherwise repeat the reviews32,67,93 suggest that multiple contacts are a
question on the subsequent page. Where one potent means of enhancing response. If resources
question is logically dependent upon another, allow, consider pre-notification,7,94,95 preferably
every effort should be made to place both on the by letter,96 to alert target respondents to the
same page. Avoid placing a short question at the arrival of the questionnaire. However, if
foot of a page, especially if preceded by a long resource constraints limit the number of contacts
question with a number of subparts. which can be made, put resources into follow-
When open-ended questions are used, suffi- ups rather than pre-notification. Regardless of
cient space should be left to record responses; whether subjects have been pre-notified, at
respondents take the amount of space provided least one reminder should be sent to non-
as indicative of the level of detail required. Do respondents.7,94,95 No one style of reminder is
not use lines for open-ended questions, unless universally appropriate;99–101 match the appeal in
Conduct and Design of Healthcare Surveys 265
the reminder letter to the perceived motivations third party without the explicit permission of the
of the study population. A consensual approach respondent concerned).
(stressing how many people have already In surveys of general populations, the balance
responded) may be appropriate to some groups, of evidence94,95,104,108,119,122–128 suggests that per-
while an indication of the likelihood of further sonalisation may offer no significant advantage.
follow-up in the absence of a response may be However, personalisation is probably appropri-
more effective with others. Our own experience ate if the message in the letter suggests personal
leads us to recommend the inclusion of a dupli- knowledge of the circumstances of the recipient
cate questionnaire with at least one reminder, if or uses a self-interest appeal. We also favour
initial non-response is believed to be related to personalisation when the target respondents are
non-delivery or mislaying of the original. If in fact personally or professionally known to the
multiple reminders are used, it may be appro- surveyor.
priate to wait until the second or subsequent A traditional letter format, using headed note-
reminder to enclose the duplicate.102 Choose a paper, is most appropriate.130 Keep the covering
mode of contact appropriate to the survey topic letter short and use language appropriate to the
and study population. Intensive techniques, such target recipients. If extensive or detailed infor-
as certified or recorded-delivery mailing, have mation needs to be given, consider including a
not been conclusively demonstrated103,104 to separate information sheet. In most circum-
have an advantage over standard mailing and stances, a facsimile signature is likely to be
may be considered by target respondents, or by adequate.131 However, care should be taken to
ethical committees, to be overly intrusive or match the style of signature to the degree of
unduly coercive. Whilst postcard reminders may personalisation of the body of the letter. No
be cost-effective,105 concerns regarding confi- single type of cover letter appeal is universally
dentiality may preclude their use in surveys on appropriate.67,107,113,123,124,131,133–136 Rather, the
health-related topics. nature of the appeal should be based on the
Respondents should not have to bear the cost perceived motivations of the study population
of returning the questionnaire; a pre-paid and and should be ethically sound. Consider includ-
ing a realistic indication of the time required for
addressed return envelope should always be
completion of the questionnaire.20 A deadline
provided. Since evidence from empirical stud-
for questionnaire return may also be specified,
ies94,95,104,106–115 and previous reviews7,8,27,32,114,115
especially if a timely response is of the
shows no consistent effect of type of postage,
essence.99,123
for convenience use franking rather than postage If ethical and practical constraints permit,
stamps for out-going mail, and use business choose a study ‘sponsor’ appropriate to the
reply envelopes for return of questionnaires. The survey topic and study population;67,86,117,139
choice between first-class and second-class mail manipulate the covering letter and return address
should involve consideration of the relative appropriately. In surveys on health-related topics,
costs, the speed with which results are required, response rates may be enhanced if the covering
and whether it is anticipated that respondents letter purports to come from the recipients’
will be aware of or influenced by the class of healthcare provider.86,139 However, considera-
mail. tion should be given to whether this approach
In general, total anonymity is not appropriate may induce response bias, and to whether it is
since it precludes targeting reminders and link- practicable.
ing questionnaire responses with other data In as far as possible, ensure the relevance and
sources (such as medical records). Moreover, interest of the survey topic to the study popula-
empirical evidence does not provide any clear tion.8,32,137,138 Fortunately, surveys on health-
support for a totally anonymous approach.117–121 related topics are generally perceived to be
We therefore recommend the use of coded highly salient!
(numbered and therefore identifiable) question- If ethical and budgetary constraints allow,
naires. It is appropriate to be explicit in a consider the use of enclosed, financial incen-
covering letter or information sheet about how tives,109,135,140–146,151,153–155 though their impact
the code number will be used (to keep a check may be less in health surveys.133,146–152 In mak-
on who has responded and thereby to allow non- ing the choice, the most relevant cost to consider
respondents to be followed-up). Provide appro- is the projected cost per returned questionnaire;
priate assurances of confidentiality, clarifying will the likely additional yield in responses
what this means (generally that only the outweigh the additional cost of providing the
research team will be able to link the numbered incentive? Note also that incentives are often
questionnaire to a named individual and that regarded as unethical in health research, and
individual responses will not be revealed to a grant-awarding bodies tend to disapprove of the
266 The Advanced Handbook of Methods in Evidence Based Healthcare
practice.9 Providing feedback to study respond- mode of administration). Such data would facili-
ents is probably unnecessary in surveys of the tate further reviews and meta-analyses.
general public or of patients,88,122,137,158 but may
be appropriate in surveys of health professionals.
REFERENCES
PRIORITIES FOR FUTURE RESEARCH 1. Moser CA, Kalton G. Survey methods in social
investigation. 2nd edn. Aldershot: Gower, 1971.
2. Bennett AE, Ritchie R. Questionnaires in medi-
As we have already noted, many of the identi-
cine: a guide to their design and use. London:
fied studies were of surveys on non-health top-
Oxford University Press, 1975.
ics, thus potentially limiting their generalisabil- 3. Dillman DA. Mail and telephone surveys: The
ity to design and conduct of health surveys. We total design method. New York: John Wiley and
recommend further research to test whether the Sons, Inc., 1978.
findings from other sectors are replicated in 4. Sudman S, Bradburn N. Asking questions: a
health-related surveys. In particular, we recom- practical guide to questionnaire design. San
mend further investigation of: the effect of mode Francisco: Jossey-Bass, 1982.
of administration in collecting data of a sensitive 5. Bradburn N, Sudman S and Associates. Improv-
and personal nature; the effect on acquiescence ing interview method and questionnaire design.
bias of mixing positive and negative statements San Francisco: Jossey Bass, 1979.
in scales to measure attitudes; the impact of 6. Advisory Group on Health Technology Assess-
various types of appeal in covering letters (in ment. Assessing the effects of health technologies.
particular the comparison of self-interest and London: Department of Health, 1993.
social utility appeals) on response rates and data 7. Linsky AS. Stimulating responses to mailed ques-
quality; the effectiveness and cost–effectiveness tionnaires: a review. Public Opinion Quarterly
of the number and mode of contacts (pre- 1975; 39: 82–101.
notification and reminders) in postal surveys. 8. Kanuk L, Berenson C. Mail surveys and
Particularly in relation to question wording response rates: A literature review. J. Marketing
and sequencing, much of the current evidence Res. 1975; 12: 440–53.
has been derived from interview surveys. It 9. Bowling A. Research methods in health: Invest-
has been postulated63 that these effects are igating health and health services. Buckingham:
minimised in postal surveys, but there is little Open University Press, 1997.
empirical evidence to back this assertion. We 10. Abramson JH. Survey methods in community
therefore recommend further investigation of medicine. Edinburgh: Churchill Livingstone,
question ordering effects in self-completion 1990.
questionnaires. Of particular interest in the con- 11. Oppenheim AN. Questionnaire design, inter-
text of health surveys is whether the relative viewing and attitude measurement. 2nd edn.
placement of entire scales (such as generic ver- London: Pinter Publishers, 1992.
sus condition-specific health status measures) 12. Fink A. The survey kit. Thousand Oaks: Sage
Publications, 1995.
affects response rates and scale scores. We iden-
13. Crombie IK, Davies HTO. Research in health
tified one study on this topic,162 which indicated
care: design, conduct and interpretation of health
no ordering effect. However, we concur with the
services research. Chichester: John Wiley and
authors of that study, who emphasise the need for
Sons, 1996.
larger studies in other disease groups to determine 14. Øvretveit J. Evaluating health interventions.
whether their results are generalisable. Buckingham: Open University Press, 1998.
Rather than designing and implementing 15. Czaja R, Blair J. Designing surveys: a guide to
studies specifically to investigate methodological decisions and procedures. Thousand Oaks: Pine
aspects of survey design and administration, we Forge Press, 1995.
recommend incorporating such experiments into 16. Sackett DL. Bias in analytic research. J. Chron.
surveys which are being carried out anyway. For Dis. 1979; 32: 51–63.
example, a split-half design, examining the 17. Murphy EA. The logic of medicine. Baltimore:
impact of different styles of reminder letter on Johns Hopkins University Press, 1976.
response rates and response quality, could be 18. Brown TL, Decker DJ, Connelly NA. Response
easily included in many studies. to mail surveys on resource-based recreation
Finally, we recommend that all reports of topics: A behavioral model and an empirical
survey research should include methodological analysis. Leisure Sciences 1989; 11: 99–110.
details (for example, length of questionnaire; 19. Jenkins CR, Dillman DA. Towards a theory
mode of administration; number, timing and of self-administered questionnaire design. In:
Conduct and Design of Healthcare Surveys 267
Lyberg L, Biemer P, Collins M, et al (eds). 37. Pederson LL, Baskerville JC, Ashley MJ,
Survey measurement and process quality. New Lefcoe NM. Comparison of mail questionnaire
York: John Wiley & Sons Inc, 1997; pp. 165– and telephone interview as data gathering strat-
96. egies in a survey of attitudes toward restrictions
20. Hornik J. Time cue and time perception effect on cigarette smoking. Can. J. Public Health
on response to mail surveys. J. Marketing Res. 1994; 76: 179–82.
1981; 18: 243–8. 38. McHorney CA, Kosinski M, Ware JE, Jr. Com-
21. McAvoy BR, Kaner EFS. General practice parisons of the costs and quality of norms for the
postal surveys: a questionnaire too far? Br. Med. SF–36 health survey collected by mail versus
J. 1996; 313: 732–3. telephone interview: results from a national sur-
22. Goyder J. The silent minority: Nonrespondents vey. Medical Care 1994; 32: 551–67.
on sample surveys. Cambridge: Polity Press, 39. Newton RR, Prensky D, Schuessler K. Form
1987. effect in the measurement of feeling states. Soc.
23. Cartwright A. Some experiments with factors Sci. Res. 1982; 11: 301–17.
that might affect the response of mothers 40. Nederhof AJ. Visibility of response as a mediat-
to a postal questionnaire. Stat. Med. 1986; 5: ing factor in equity research. J. Soc. Psychol.
607–17. 1984; 122: 211–15.
24. Cartwright A. Professionals as responders: 41. Oei TI, Zwart FM. The assessment of life
variations in and effects of response rates to events: self-administered questionnaire versus
questionnaires, 1961–77. Br. Med. J. 1978; 2: interview. J. Affect. Disord. 1986; 10: 185–90.
1419–21. 42. Cartwright A. Interviews or postal question-
25. Fowler FJ. Survey research methods. Beverly naires? Comparisons of data about women’s
Hills, CA: Sage, 1984. experiences with maternity services. Milbank Q.
26. Borg WR, Gall MD. Educational research: an 1988; 66: 172–89.
introduction. New York: Longman, 1983. 43. Boekeloo BO, Schiavo L, Rabin DL, Conlon
27. Asch DA, Jedrziewski K, Christiakis NA. RT, Jordan CS, Mundt DJ. Self-reports of HIV
Response rates to mail surveys published in risk factors by patients at a sexually transmitted
medical journals. J. Clin. Epidemiol. 1997; 50: disease clinic: audio vs written questionnaires.
1129–36. Am. J. Public Health 1994; 84: 754–60.
28. Thibaut JW, Kelley HH. The social psychology 44. Jordon LA, Marcus AC, Reeder LG. Response
of groups. New York: Wiley, 1959. styles in telephone and household interviewing:
29. Homans GC. Social behavior: Its elementary A field experiment. Public Opinion Q. 1980; 44:
forms. New York: Harcourt, Brace and World, 210–22.
1961. 45. Quinn RP, Gutek BA, Walsh JT. Telephone inter-
30. Blau PM. Exchange and power in social life. viewing: A reappraisal and a field experiment.
New York: Wiley, 1964. Basic Appl. Soc. Psychol. 1980; 1: 127–53.
31. Morton Williams J. Interviewer approaches. 46. Aneshensel CS, Frerichs RR, Clark VA,
Aldershot: Dartmouth Publishing Company Yokopenic PA. Telephone versus in person sur-
Limited, 1993. veys of community health status. Am. J. Public
32. Heberlein TA, Baumgartner R. Factors affecting Health 1982; 72: 1017–21.
response rates to mailed questionnaires: A quan- 47. Fenig S, Levav I, Kohn R, Yelin N. Telephone
titative analysis of the published literature. Am. vs face-to-face interviewing in a community
Sociol. Rev. 1978; 43: 447–62. psychiatric survey. Am. J. Public Health 1993;
33. Jenkinson C, Ziebland S, Fitzpatrick R, Mowat 83: 896–8.
A, Mowat A. Hospitalisation and its influence 48. Allen DF. Computers versus scanners: An
upon results from health status questionnaires. experiment in nontraditional forms of survey
Int. J. Health Sci. 1993; 4: 13–19. administration. J. Coll. Student Personnel 1987;
34. Frey JH; Oishi SM. How to conduct interviews 28: 266–73.
by telephone and in person. Thousand Oaks: 49. Higgins CA, Dimnik TP, Greenwood HP. The
Sage Publications, 1995. DISKQ survey method. Special Issue: Political
35. Hinkle AL, King GD. A comparison of three opinion polling. J. Market Research Soc. 1987;
survey methods to obtain data for community 29: 437–45.
mental health program planning. Am. J. Com- 50. Liefeld JP. Response effects in computer-
munity Psychol. 1978; 6: 389–97. administered questioning. J. Marketing Res.
36. Talley JE, Barrow JC, Fulkerson KF, Moore 1988; 25: 405–9.
CA. Conducting a needs assessment of uni- 51. Helgeson JG, Ursic ML. The decision process
versity psychological services: A campaign of equivalency of electronic versus pencil-and-
telephone and mail strategies. J. Am. Coll. paper data collection methods. Soc. Sci. Com-
Health 1983; 32: 101–3. puter Rev. 1989; 7: 296–310.
268 The Advanced Handbook of Methods in Evidence Based Healthcare
52. Groves RM, Mathiowetz NA. Computer assisted 68. Bradburn NM, Mason WM. The effect of ques-
telephone interviewing: Effects on interviewers tion order on response. J. Marketing Res. 1964;
and respondents. Public Opin. Q. 1984; 48: 1: 61
356–69. 69. Tourangeau R, Rasinski KA, Bradburn N,
53. Schuman H, Ludwig J, Krosnick JA. The per- D’Andrade R. Belief accessibility and context
ceived threat of nuclear war, salience, and open effects in attitude measurement. J. Exp. Soc.
questions. Public Opin. Q. 1986; 50: 519–36. Psychol. 1989; 25: 401–21.
54. Bishop GF, Oldendick RW, Tuchfarber AJ. 70. Tourangeau R, Rasinski KA, Bradburn N,
Effects of presenting one versus two sides of an D’Andrade R. Carryover effects in attitude sur-
issue in survey questions. Public Opin. Q. 1982; veys. Public Opin. Q. 1989; 53: 495–524.
46: 69–85. 71. Colasanto D, Singer E, Rogers TF. Context
55. Larsen JD, Mascharka C, Toronski C. Does the effects on responses to questions about AIDS.
wording of the question change the number of Public Opin. Q. 1992; 56: 515–18.
headaches people report on a health question- 72. Schuman H, Kalton G, Ludwig J. Context and
naire? Psychol. Record 1987; 37: 423–7. contiguity in survey questionnaires. Public
56. Blair E, Burton S. Cognitive processes used by Opin. Q. 1983; 47: 112–15.
survey respondents to answer behavioral fre- 73. Hawkins DI, Coney KA. Uninformed response
quency questions. J. Consumer Research 1987; error in survey research. J. Marketing Res. 1981;
14: 280–8. 18: 370–4.
57. Schriesheim CA, Hill KD. Controlling acquies- 74. Wandzilak T, Ansorge CJ, Potter G. Utilizing
cence response bias by item reversals: The effect ‘undecided’ option with Likert items: Associ-
on questionnaire validity. Educational and Psy- ated measurement problems. Int. J. Sport Psy-
chological Measurement 1981; 41: 1101–14. chol. 1987; 18: 51–8.
58. Bishop GF, Oldendick RW, Tuchfarber AJ. 75. Poe GS, Seeman I, McLaughlin J, Mehl E.
Effects of filter questions in public opinion sur- ‘Don’t know’ boxes in factual questions in a
veys. Public Opin. Q. 1983; 47: 528–46.
mail questionnaire: Effects on level and quality
59. Bishop GF, Tuchfarber AJ, Oldendick RW.
of response. Public Opin. Q. 1988; 52: 212–22.
Opinions on fictitious issues: the pressure to
76. Israel GD, Taylor CL. Can response order bias
answer survey questions. Public Opin. Q. 1986;
evaluations? Evaluation and Program Planning
50: 240–50.
1990; 13: 365–71.
60. Smith ER, Squire P. The effects of prestige
77. Frisbie DA, Brandenburg DC. Equivalence of
names in question wording. Public Opin. Q.
questionnaire items with varying response for-
1990; 54: 97–116.
mats. J. Educational Measurement 1979; 16:
61. Schuman H, Presser S, Ludwig J. Context
43–8.
effects on survey responses to questions about
78. Lam TC, Klockars AJ. Anchor point effects on
abortion. Public Opin. Q. 1981; 45: 216–23.
62. Schuman H, Presser S. Questions and answers the equivalence of questionnaire items. J. Edu-
in attitude surveys: experiments on question cational Measurement 1982; 19: 317–22.
form, wording and content. New York: Aca- 79. Bourque LB, Fielder EP. How to conduct self-
demic Press, 1981. administered and mail surveys. Thousand Oaks:
63. Smith TW. Condition order effects: GSS tech- Sage, 1995.
nical report No. 33. Chicago: National Opinion 80. Gaskell GD, O’Muircheartaigh CA, Wright DB.
Research Centre, 1982. Survey questions about vaguely defined events:
64. Ayidiya SA, McClendon MJ. Response effects the effects of response alternatives. Public Opin.
in mail surveys. Public Opin. Q. 1990; 54: Q. 1994; 58: 241–54.
229–47. 81. Stem DE, Lamb CW, MacLachlan DL. Remote
65. Hippler H-J, Schwarz N. Response effects in versus adjacent scale questionnaire designs. J.
surveys. In: Hippler H-J, Schwarz N, Sudman S Market Res. Soc. 1978; 20: 3–13.
(eds). Social information processing and survey 82. Market Research Society R&DC. Report of the
methodology. New York: Springer-Verlag, Second Working Party on Respondent Co-
1987. operation: 1977–80. J. Market Res. Soc. 1981;
66. Roberson MT, Sundstrom E. Questionnaire 23: 3–25.
design, return rates, and response favorableness 83. Cannell CF, Kahn RL. Interviewing. In: Lindzey
in an employee attitude questionnaire. J. Appl. G, Aronson E. (eds). The handbook of social
Psychol. 1990; 75: 354–7. psychology – volume II. Reading: Addison-
67. Jones WH, Lang JR. Sample composition bias Wesley, 1968;
and response bias in a mail survey: A com- 84. US Department of Health, Education and Wel-
parison of inducement methods. J. Marketing fare and National Centre for Health Services
Res. 1980; 17: 69–76. Research. Health survey methods – respondent
Conduct and Design of Healthcare Surveys 269
burden. 1977; Proceedings from the 2nd Bi- 101. Dommeyer CJ. The effects of negative cover
ennial Conference. letter appeals on mail survey response. Special
85. Cartwright A. Who responds to postal ques- Issue: political opinion polling. J. Market Res.
tionnaires? J. Epidemiol. Community Health Soc. 1987; 29: 445–51.
1986; 40: 267–73. 102. Swan JE, Epley DE, Burns WL. Can follow-up
86. Jacoby A. Possible factors affecting response to response rates to a mail survey be increased by
postal questionnaires: findings from a study of including another copy of the questionnaire?
general practitioner services. J. Public Health Psychological Reports 1980; 47: 103–6.
Med. 1990; 12: 131–5. 103. Gitelson RJ, Drogin EB. An experiment on the
87. Hansen RA, Robinson LM. Testing the effec- efficacy of a certified final mailing. J. Leisure
tiveness of alternative foot-in-the-door manipu- Res. 1992; 24: 72–8.
lations. J. Marketing Res. 1980; 17: 359–64. 104. Kahle LR, Sales BD. Personalization of the
88. Powers DE, Alderman DL. Feedback as an outside envelope in mail surveys. Public Opin.
incentive for responding to a mail questionnaire. Q. 1978; 42: 547–50.
Res. Higher Education 1982; 17: 207–11. 105. Roberts H, Pearson JCG, Dengler R. Impact of
89. Roszkowski MJ, Bean AG. Believe it or not – a postcard versus a questionnaire as a first
Longer questionnaires have lower response reminder in a postal lifestyle survey. J. Epi-
rates. J. Business Psychol. 1990; 4: 495–509. demiol. Community Health 1993; 47: 334–5.
90. Herzog AR, Bachman JG. Effects of question- 106. Harris JR, Guffey HJ. Questionnaire returns:
naire length on response quality. Public Opin. Q. Stamps versus business reply envelopes re-
1981; 45: 549–59. visited. J. Marketing Res. 1978; 15: 290–3.
91. Jobber D, Sanderson S. The effects of a prior 107. Jones WH, Linda G. Multiple criteria effects in a
letter and coloured questionnaire paper on mail mail survey experiment. J. Marketing Res. 1978;
survey response rates. J. Market Res. Soc. 1983; 15: 280–4.
25: 339–49. 108. Labrecque DP. A response rate experiment
92. Olivarius NdF, Andreasen AH. Day-of-the-week using mail questionnaires. J. Marketing 1978;
effect on doctors’ response to a postal ques- 42: 82–3.
tionnaire. Scand. J. Primary Health Care 1995; 109. Hopkins KD, Podolak J. Class-of-mail and the
13: 65–7. effects of monetary gratuity on the response
93. Peterson RA, Albaum G, Kerin RA. A note on rates of mailed questionnaires. J. Exp. Education
alternative contact strategies in mail surveys. J. 1983; 51: 169–70.
Market Res. Soc. 1989; 31: 409–18. 110. Corcoran KJ. Enhancing the response rate in
94. Martin WS, Duncan WJ, Powers TL, Sawyer survey research. Social Work Research and
JC. Costs and benefits of selected response Abstracts 1985; 21: 2.
inducement techniques in mail survey research. 111. Elkind M, Tryon GS, de Vito AJ. Effects of type
J. Business Res. 1989; 19: 67–79. of postage and covering envelope on response
95. Martin WS, Duncan WJ, Sawyer JC. The inter- rates in a mail survey. Psychological Rep. 1986;
active effects of four response rate inducements 59: 279–83.
in mail questionnaires. College Student J. 1984; 112. Harvey L. A research note on the impact of
18: 143–9. class-of-mail on response rates to mailed ques-
96. Faria AJ, Dickinson JR, Filipic TV. The effect tionnaires. J. Market Res. Soc. 1986; 28:
of telephone versus letter prenotification on mail 299–300.
survey response rate, speed, quality and cost. J. 113. Cartwright A, Windsor J. Some further experi-
Market Res. Soc. 1990; 32: 551–68. ments with factors that might affect the response
97. Allen CT, Schewe CD, Wijk G. More on self- to postal questionnaires. Survey Methodology
perception theory’s foot technique in the pre- Bulletin 1989; 25: 11–15.
call/mail survey setting. J. Marketing Res. 1980; 114. Duncan W. Mail questionnaires in survey
17: 498–502. research: A review of respondent inducement
98. Kamins MA. The enhancement of response rates techniques. J. Management 1979; 5: 39–55.
to a mail survey through a labelled probe foot- 115. Armstrong JS, Lusk EJ. Return postage in mail
in-the-door approach. J. Market Res. Soc. 1989; surveys: A meta-analysis. Public Opin. Q. 1987;
31: 273–83. 51: 233–48.
99. Nevin JR, Ford NM. Effects of a deadline and a 116. Zeinio RN. Data collection techniques: mail
veiled threat on mail survey responses. J. Appl. questionnaires. Am. J. Hosp. Pharmacy 1980;
Psychol. 1976; 61: 116–18. 37: 1113–19.
100. Blass T, Leichtman SR, Brown RA. The effect 117. Jones WH. Generalizing mail survey induce-
of perceived consensus and implied threat upon ment methods: Population interactions with
responses to mail surveys. J. Soc. Psychol. 1981; anonymity and sponsorship. Public Opin. Q.
113: 213–16. 1979; 43: 102–11.
270 The Advanced Handbook of Methods in Evidence Based Healthcare
118. McDaniel SW, Rao CP. An investigation of mailed questionnaires. Rehabilitation Counsel-
respondent anonymity’s effect on mailed ques- ing Bulletin 1978; 22: 138–41.
tionnaire response rate and quality. J. Market 134. McKillip J, Lockhart DC. The effectiveness of
Res. Soc. 1981; 23: 150–60. cover-letter appeals. J. Soc. Psychol. 1984; 122:
119. Childers TL, Skinner SJ. Theoretical and empiri- 85–91.
cal issues in the identification of survey respond- 135. Biner PM. Effects of cover letter appeal and
ents. J. Market Res. Soc. 1985; 27: 39–53. monetary incentives on survey response: A
120. Campbell MJ, Waters WE. Does anonymity reactance theory application. Basic Appl. Soc.
increase response rate in postal questionnaire Psychol. 1988; 9: 99–106.
surveys about sensitive subjects? A randomised 136. Biner PM, Barton DL. Justifying the enclosure
trial. J. Epidemiol. Community Health 1990; 44: of monetary incentives in mail survey cover
75–6. letters. Psychol. Marketing 1990; 7: 153–62.
121. McKee DO. The effect of using a questionnaire 137. Dommeyer CJ. Does response to an offer of
identification code and message about non- mail survey results interact with questionnaire
response follow-up plans on mail survey response interest? J. Market Res. Soc. 1985; 27: 27–38.
characteristics. J. Market Res. Soc. 1992; 34: 138. Hovland EJ, Romberg E, Moreland EF. Non-
179–91. response bias to mail survey questionnaires
122. Green KE, Kvidahl RF. Personalization and within a professional population. J. Dental
offers of results: Effects on response rates. J. Education 1980; 44: 270–4.
Exp. Education 1989; 57: 263–70. 139. Smith WC, Crombie IK, Campion PD, Knox JD.
123. Roberts RE, McCrory OF, Forthofer RN. Fur- Comparison of response rates to a postal ques-
ther evidence on using a deadline to stimulate tionnaire from a general practice and a research
responses to a mail survey. Public Opin. Q. unit. Br. Med. J. 1994; 1985: 1483–5.
1978; 42: 407–10. 140. Hansen RA. A self-perception interpretation of
124. Childers TL, Pride WM, Ferrell OC. A reassess- the effect of monetary and nonmonetary incen-
ment of the effects of appeals on response to tives on mail survey respondent behavior. J.
mail surveys. J. Marketing Res. 1980; 17: Marketing Res. 1980; 17: 77–83.
365–70. 141. Furse DH, Stewart DW. Monetary incentives
125. Woodward JM, McKelvie SJ. Effects of topical versus promised contribution to charity: new
interest and mode of address on response to mail evidence on mail survey response. J. Marketing
survey. Psychological Rep. 1985; 57: 929–30. Res. 1982; 19: 375–80.
126. Worthen BR, Valcarce RW. Relative effective- 142. Paolillo JG, Lorenzi P. Monetary incentives and
ness of personalized and form covering letters in mail questionnaire response rates. J. Advertising
initial and follow-up mail surveys. Psycho- 1984; 13: 46–8.
logical Rep. 1985; 57: 735–44. 143. Blythe BJ. Increasing mailed survey responses
127. Green KE, Stager SF. The effects of personal- with a lottery. Social Work Research and
ization, sex, locale, and level taught on educators’ Abstracts 1986; 22: 18–19.
responses to a mail survey. J. Exp. Education 144. Trice AD. Maximizing participation in surveys:
1986; 54: 203–6. hotel ratings VII. J. Soc. Behav. Personality
128. Wunder GC, Wynn GW. The effects of ad- 1986; 1: 137–41.
dress personalization on mailed questionnaires 145. Brennan M. The effect of a monetary incentive
response rate, time and quality. J. Market Res. on mail survey response rates: new data. J.
Soc. 1988; 30: 95–101. Market Res. Soc. 1992; 34: 173–7.
129. Wiseman F. A reassessment of the effects of 146. Little RE, Davis AK. Effectiveness of various
personalization on response patterns in mail sur- methods of contact and reimbursement on
veys. J. Marketing Res. 1976; 13: 110–11. response rates of pregnant women to a mail
130. Wagner WG, O’Toole WM. The effects of cover questionnaire. Am. J. Epidemiol. 1984; 120:
letter format on faculty response rate in mail 161–3.
survey research. Educational Psychol. Res. 147. Cook JR, Schoeps N, Kim S. Program responses
1985; 5: 29–37. to mail surveys as a function of monetary incen-
131. Dodd DK, Boswell DL, Litwin WJ. Survey tives. Psychol. Rep. 1985; 57: 366.
response rate as a function of number of sig- 148. Mortagy AK, Howell JB, Waters WE. A useless
natures, signature ink color, and postscript on raffle. J. Epidemiol. Community Health 1985;
covering letter. Psychological Rep. 1988; 63: 39: 183–4.
538. 149. Woodward A, Douglas B, Miles H. Chance of
132. Dodd DK, Markwiese BJ. Survey response rate free dinner increases response to mail ques-
as a function of personalized signature on cover tionnaire [letter]. Int. J. Epidemiol. 1985; 14:
letter. J. Soc. Psychol. 1987; 127: 97–8. 641–2.
133. Salomone PR, Miller GC. Increasing the 150. Weltzien RT, McIntyre TJ, Ernst JA, Walsh JA.
response rates of rehabilitation counselors to Crossvalidation of some psychometric properties
Conduct and Design of Healthcare Surveys 271
of the CSQ and its differential return rate as a 156. Erdos PL, Morgan AJ. Professional mail sur-
function of token financial incentive. Community veys. New York: McGraw-Hill, 1970.
Mental Health J. 1986; 22: 49–55. 157. Sudman S. Mail surveys of reluctant profession-
151. Berry SH, Kanouse DE. Physician response to a als. Evaluation Rev. 1985; 9: 349–60.
mailed survey: An experiment in timing of pay- 158. Dommeyer CJ. Offering mail survey results in a
ment. Public Opin. Q. 1987; 51: 102–14. lift letter. J. Market Res. Soc. 1989; 31:
152. Fiset L, Milgrom P, Tarnai J. Dentists’ response 399–408.
to financial incentives in a mail survey of mal- 159. Lovelock CH, Stiff R, Cullwick D, Kaufman
practice liability experience. J. Public Health IM. An evaluation of the effectiveness of drop-
Dentistry 1994; 54: 68–72. off questionnaire delivery. J. Marketing Res.
153. Hubbard R, Little EL. Promised contributions to 1976; 13: 358–64.
charity and mail survey responses: replication 160. Layne BH, Thompson DN. Questionnaire page
with extension. Public Opin. Q. 1988; 52: length and return rate. J. Soc. Psychol. 1981;
223–30. 113: 291–2.
154. Gajraj AM, Faria AJ, Dickinson JR. A com- 161. de Vaus DA. Surveys in social research. 3rd edn.
parison of the effect of promised and provided London: UCL Press Ltd, 1991.
lotteries, monetary and gift incentives on mail 162. Barry MJ, Walker-Corkery E, Chang Y, Tyll
survey response rate, speed and cost. J. Market LT, Cherkin DC, Fowler FJ. Measurement of
Res. Soc. 1990; 32: 141–62. overall and disease specific health status: does
155. Dommeyer CJ. How form of the monetary order of the questionnaires make a difference. J.
incentive affects mail survey response. J. Market Health Services Research and Policy 1996; 1:
Res. Soc. 1988; 30: 379–85. 20–7.
Part IV
ANALYTICAL METHODS
I N T R O D U C T I O N by K E I T H R . A B R A M S
This section of the book considers developments Hierarchical modelling enables proper allow-
in analytical methodology. Over the past 5–10 ance to be made for the often natural nesting
years, these have predominantly been con- or clustering that occurs in much healthcare
centrated in four separate, but linked areas: research. For example, repeated observations are
Bayesian methods; hierarchical modelling; qual- nested within patients, who themselves are
ity of life research; and economic evaluation. nested within healthcare units. Ignoring such
Bayesian methods, first advocated by the clustering can lead to biased estimation, whilst
Reverend Thomas Bayes in the 1700s, and for modelling it often specifically addresses ques-
much of the 20th century thought controversial tions of interest, for example identifying where
by many statisticians, have developed rapidly in the largest source of variation is.
recent years. This development has been fuelled Quality of life research has over the past
by computational advances and a realisation that decade become an integral part of almost every
such methods, now that they can be applied evaluation of a health technology. In terms of
relatively easily, can make a major contribution methodological development to date, much
in a number of key areas in health services, work has been undertaken on the development,
technology and policy research,1,2 for example validation and choice of specific instruments, as
monitoring and interpreting clinical trials and in has been discussed in Chapter 11 by Fitzpatrick
meta-analysis. The key distinction between the et al. and by Bowling (1996).4 By comparison,
Bayesian approach to statistical inference and relatively little work has been done on either
the ‘classical’ approach, is that evidence external specific design issues for studies that contain a
to the current study, either in the form of evi- quality of life element or in the development of
dence from other studies or as subjective judge- appropriate methods of analysis.5 The wide-
ments, can be – and is – formally incorporated spread use of quality of life measures in both
into the analysis. randomised and non-randomised studies and
The other area of rapid development over across a diverse range of clinical and service/
approximately the same time period has been care delivery settings makes consideration of
that of hierarchical modelling.3 As in the devel- methods for the design and analysis of such
opment of Bayesian methods, the foundations of studies particularly timely.
hierarchical modelling had been laid many years Economic evaluation concerns the systematic
previously – in the early part of the 20th century evaluation of alternative courses of action in
by Fisher and colleagues working in agricultural terms of both costs and health outcomes.6 Whilst
research. Again, it has only been with computa- economic evaluation within a health technology
tional advances that such methods can now be context has previously not been quite as prolific
applied to problems in healthcare research – an as the use of quality of life measures, it is
area which does not benefit from the same level quickly reaching a stage where it is almost
of experimental control as agricultural research. mandatory that any evaluative study comprises
274 The Advanced Handbook of Methods in Evidence Based Healthcare
both a quality of life and an economic element. primary outcome measures should be analysed,
During the 1980s much of health economics and the methods currently available for doing
literature addressed the question of the relevance so. Gulliford and colleagues consider the design
and justification of economic evaluation within a and analysis of studies which evaluate area-wide
health context, but by the 1990s the literature interventions. The very nature of such evalua-
had started to consider what the most appro- tive studies means that explicit account must be
priate approaches and methods are for such taken of the hierarchical or clustered nature of
evaluations – for example whether it was neces- the data both in the design, because subjects
sary for them to be conducted within a random- within the same interventional unit will be more
ised controlled trial. In terms of methodological similar than those in different units, and in the
development, one of the key – if not the over- analysis. Gray and Briggs consider the issue of
riding – questions is how can uncertainty, in uncertainty in economic evaluation; both the
terms of both methodology and costs/outcomes, extent to which, and the means by which, it is
be taken into account in any analysis. In ad- taken into account in the healthcare research
dressing this question a dialogue is developing literature. In addition they consider issues sur-
between methodologists regarding the merits or rounding the distribution of cost data in economic
otherwise of the various possible approaches, evaluations.
and in particular as regards the use of sensitivity
analyses compared with stochastic or prob-
abilistic modelling.7 REFERENCES
Whilst a methodological health technology
literature is beginning to be established in all
1. Berry DA, Stangl DK. Bayesian Biostatistics. New
four fields, they will all undoubtedly be active
York: Marcel Dekker, 1996.
areas of methodological research over the next 2. Gilks WR, Richardson S, Spiegelhalter DJ. Markov
5–10 years. This section of the book is organ- Chain Monte Carlo in Practice. London: Chapman
ised into five chapters which specifically con- & Hall, 1996.
sider the use of Bayesian methods, design and 3. Goldstein H. Multilevel Statistical Models. 2nd
analysis issues in quality of life research, the edn. London: Edward Arnold, 1995.
evaluation of area-wide technologies, and un- 4. Bowling A. Measuring disease: a review of
certainty in economic evaluation. Spiegelhalter disease-specific quality of life measurement scales.
and colleagues consider the use of Bayesian Buckingham: Open University Press, 1996.
methods in Health Technology Assessment 5. Cox DR, Fitzpatrick R, Fletcher AE, Gore SM,
(HTA). They explain what Bayesian methods Spiegelhalter DJ, Jones DR. Quality-of-life assess-
are, and consider the advantages and disadvant- ment: can we keep it simple? Journal of the Royal
ages of using such methods. Campbell and col- Statistical Society 1992; 155: 353–93.
leagues and Billingham and colleagues consider 6. Drummond MF, O’Brien B, Stoddart GL, Torrence
different aspects of designing and analysing G. Methods for the economic evaluation of health
studies in which quality of life is a primary care programmes. 2nd edn. Oxford: Oxford Uni-
outcome measure. Whilst Campbell looks at versity Press, 1997.
issues concerning the design, and in particular 7. Briggs A, Sculpher M, Buxton M. Uncertainty in
the issue of how large such studies should be, the economic evaluation of health care technologies:
Billingham addresses the problem of how studies the role of sensitivity analysis. Health Analysis
in which both quality and quantity of life are 1994; 3: 95–104.
16
Bayesian Methods
The final combination is done using a rule in We also emphasise that our definition of
probability theory known as Bayes’ theorem, ‘Bayesian’ may be more restrictive than that of
thus giving rise to the ‘Bayesian’ label. This other researchers, in that we are only concerned
chapter attempts to provide a guide to the with revision of beliefs after evidence, and do
diverse literature on the ramifications of this not particularly focus on decision-making or
apparently straightforward idea. decision analysis. We are also not concerned
What, then, is a Bayesian approach to HTA? with using Bayes’ theorem for prognostic or
We have defined it1 as ‘the explicit quantitative diagnostic statements.
use of external evidence in the design, monitor- Our initial electronic search yielded over
ing, analysis, interpretation and reporting of a 4000 papers on ‘Bayesian’ methods between
health technology assessment study’. 1990 and 1998, and all of the abstracts were
After describing the methods used in search- hand-searched for relevant material. Forward
ing and summarising the literature, we provide and backward searches, personal collections and
a brief critique of conventional statistical ap- hand-searching of recent journals provided addi-
proaches, and then outline the main features of a tional references, and a subjective judgement as
general Bayesian philosophy as applied in the to their relative importance led to the selection
context of HTA. Sections then deal in more in this chapter. The references provided are by
detail with the source of the prior distribution, no means exhaustive, but should form a con-
randomised trials, observational studies, evi- venient starting point for the literature.
dence synthesis and policy making, and finally
we provide some suggestions for future research.
Mathematical and computational methods will
be barely mentioned, and details should be THE ‘CLASSICAL’ STATISTICAL APPROACH
sought in the references provided. The chapter IN HTA
is therefore structured by context rather than
methods, although some methodological themes It would be misleading to dichotomise statistical
inevitably run throughout; for example, what methods as either ‘classical’ or ‘Bayesian’, since
form of prior distribution is appropriate, how both terms cover a bewildering range of tech-
can the posterior distributions be presented, and niques. It is a little more fair to divide conven-
what is the contrast to traditional methods? tional statistics into two broad schools: Fisherian
It will be clear that Bayesian methods are a and Neyman–Pearson.
controversial topic in that they may involve the
explicit use of subjective judgements in what is 1 The Fisherian approach to inference on an
supposed to be a rigorous scientific exercise in unknown intervention effect u is based on
HTA. We should therefore be quite explicit as to the likelihood function mentioned earlier,
our own subjective biases, which will doubtless which gives rise to an estimate comprising
be apparent from the organisation and text of the ‘most-likely’ value for u, intervals based
this chapter. We favour the Bayesian philos- on the range of values of u supported by the
ophy, would like to see its use extended in HTA, data, and the evidence against specified null
but feel that this should be carried out cautiously hypotheses summarised by P-values (the
and in parallel with the currently accepted chance of getting a result as extreme as that
methods. observed were the null hypothesis true).
2 The Neyman–Pearson approach is focused
on the chances of making various types of
error, so that clinical trials are designed to
METHODS USED IN THIS REVIEW
have a fixed Type I error (the chance of
incorrectly rejecting the null hypothesis),
In common with all such methodological usually 5%, and fixed power (one minus the
reviews, it is essential to define what we mean Type II error, the chance of not detecting the
by ‘systematic’. In our full report2 we have only alternative hypothesis), often 80% or 90%.
attempted a comprehensive review (i.e. all refer- The situation is made more complex if a
ences) of ‘proper’ Bayesian HTA studies, which sequential design is used, in which the data
we define as those in which an informative prior are periodically analysed and the trial
distribution has been given to the quantity of stopped if sufficiently convincing results
interest. In this chapter we have necessarily obtained. Repeated analysis of the data has a
been more selective, and so use ‘systematic’ in strong effect on the Type I error, since there
the sense of identifying the major issues and are many opportunities to obtain a false
references. positive result, and thus the P-value and
Bayesian Methods 277
Figure 1 Prior, likelihood and posterior distributions arising from GREAT trial
of home thrombolysis.
trial), the posterior probability that the reduction This is known as the frequency interpretation of
is at least 50% is only 5%, and a 95% interval probability, and so both Fisherian and Neyman–
runs from a 43% to 0% risk reduction. Pearson schools are often referred to as ‘fre-
quentist’. We have allowed probability to refer
to generic uncertainty about any unknown quan-
The Subjective Interpretation of tity, and this is an application of the subjectivist
Probability interpretation of probability.
The Bayesian perspective thus extends the
The normal use of probability describes long run remit of standard statistical analysis, in that
frequency properties of repeated random events. there is explicit concern for what it is reasonable
Bayesian Methods 279
for an observer to believe in the light of data. priors to represent archetypal positions of ig-
Thus the perspective of the consumer of the norance, scepticism and enthusiasm and, when
analysis is explicitly taken into account: for multiple related studies are being simultan-
example, in a trial on a new drug being carried eously analysed, the assumption of a common
out by a pharmaceutical company, the view- prior that may be ‘estimated’.
points of the company, the regulatory authorities Two important points should be emphasised
and the medical profession may be substantially immediately:
different. The subjective nature of the analysis is 1 Despite the name ‘prior’ suggesting a tem-
therefore unapologetically emphasised, although poral relationship, it is quite feasible that a
it has been criticised elsewhere.10 The prior prior distribution is decided after seeing the
distribution shown in Figure 1(a) was based on results of a study, since it is simply intended
the subjective judgement of a senior cardiologist, to summarise reasonable uncertainty given
informed by empirical evidence derived from evidence external to the study in question.
one unpublished and two published trials. 2 There is no such the thing as the ‘correct’
Of course, conclusions strongly based on prior. Instead, researchers have suggested
beliefs that cannot be supported by concrete using a ‘community’ of prior distributions
evidence are unlikely to be widely regarded as expressing a range of reasonable opinions.
convincing, and so it is important to attempt to
find consensus on reasonable sources of external
evidence. The assessment and use of prior Predictions
beliefs is discussed further in pp. 280–3.
Suppose we wish to predict some future obser-
The Relation to the Use of Bayes’ vations z, which depend on the unknown quan-
tity u through a distribution p(z ) u): for example,
Theorem in Diagnostic Tests z may be the difference in outcomes to be
observed on some future patients. Since our
If u is something that is potentially observable, current uncertainty concerning u is expressed by
and p(u) is available from known data, then the posterior distribution p(u ) y), then we can
the use of Bayes’ theorem is uncontroversial. integrate out the unknown u from the product of
For example, it has long been established that p(z ) u) and p(u ) y) to obtain the predictive dis-
sensitivity and specificity are insufficient char- tribution p(z ) y) for the future observations z.
acteristics to judge the result of a diagnostic test Such predictive distributions are useful in
for an individual – the disease prevalence is also many contexts: Berry and Stangl15 describe their
needed. use in design and power calculations, model
From the HTA perspective, the more inter- checking, and in deciding whether to conduct a
esting and controversial context is when the future trial. See pp. 284–6 and 288 for applica-
unknown u is a quantity that is not potentially tions of predictions.
directly observable, such as the mean benefit of
a new therapy in a defined group of patients.
There have been many arguments11–14 for the Decision-Making
connection between the use of Bayes’ theorem
in diagnostic testing and in general clinical There is a long history of attempts to apply the
research, pointing out that just as the prevalence theory of optimal decision-making to medicine,
is required for the assessment of a diagnostic in which (posterior) probabilities are combined
test, so the prior distribution on u is required to with utilities, although the impact has been
supplement the usual information (P-values and rather limited: Healy16 gives a good tutorial intro-
confidence intervals) which summarises the duction. Relevant applications within the HTA
likelihood. We need only think of the huge context include the explicit use of utility func-
number of clinical trials that are carried out, tions in the design and monitoring of clinical
with few clear successes, to realise that the trials (see pp. 285–6 and 288 for further discus-
‘prevalence’ of truly effective treatments is low. sion and other examples).
Thus we should be cautious about accepting
extreme results, such as observed in the GREAT Multiplicity
trial, at face value.
The context of clinical trials may present issues
The Prior Distribution of ‘multiple analyses of accumulating data, ana-
lyses of multiple endpoints, multiple subsets
The next section provides a full discussion of of patients, multiple treatment group contrasts
the source and use of prior distributions, includ- and interpreting the results of multiple clinical
ing elicitation from experts, the use of ‘default’ trials’.17 Observational data may feature multiple
280 The Advanced Handbook of Methods in Evidence Based Healthcare
institutions, and meta-analysis involve synthesis 4 The decision–theoretic or ‘full’ Bayes ap-
of multiple studies. The general Bayesian ap- proach, in which explicit utility functions are
proach to multiplicity involves specifying a used to make decisions based on maximising
common prior distribution for the substudies that expected utility.
expresses a belief in the expected ‘similarity’ of
Our focus in this chapter is on the third, proper,
all the individual unknown quantities being esti-
school of Bayesianism.
mated. This produces a degree of pooling,
in which an individual study’s results tend to
be ‘shrunk’ towards the average result by an Further Reading
amount depending on the variability between
studies and the precision of the individual A wide-ranging introductory chapter by Berry
study: relevant contexts include subset analysis, and Stangl15 in their textbook19 covers a whole
between-centre variability, and random-effects range of modelling issues, including elicitation,
meta-analysis. This is discussed on pp. 283 and model choice, computation, prediction and
286. decision-making. Non-technical tutorial articles
include those by Lewis and Wears20 and Lilford
Computational Issues and Braunholtz.21 Other authors emphasise dif-
ferent merits of Bayesian approaches in HTA:
The Bayesian approach applies probability theory Eddy et al.22 concentrate on the ability to deal
to a model derived from substantive knowledge with varieties of outcomes, designs and sources
and can, in theory, deal with realistically com- of bias, Breslow23 stresses the flexibility with
plex situations. It has to be acknowledged, how- which multiple similar studies can be handled,
ever, that the computations may be difficult, Etzioni and Kadane24 discuss general applica-
with the specific problem being to carry out the tions in the health sciences with an emphasis on
integrations necessary to obtain the posterior decision-making, while Freedman25 and Lilford
distributions of interest in situations where and Braunholtz21 concentrate on the ability to
non-standard prior distributions are used, or combine ‘objective’ evidence with clinical
where there are additional ‘nuisance parameters’ judgement.
in the model. These problems in integration for There is a huge methodological statistical
many years restricted Bayesian applications to literature on general Bayesian methods, much
rather simple examples. However, there has of it quite mathematical. Cornfield26 provides
recently been enormous progress in methods a theoretical justification of the Bayesian ap-
for Bayesian computation, generally exploiting proaches, in terms of ideas such as coherence. A
modern computer power to carry out simulations rather old article27 is still one of the best intro-
known as Markov Chain Monte Carlo (MCMC) ductions to the Bayesian philosophy. Good
methods.18 tutorial introductions are provided by Lindley28
and Barnett,29 while more recent books, in order
of increasing technical difficulty, include those
Schools of Bayesians by Lee,30 Gelman et al.,31 Carlin and Louis,32
and Bernardo and Smith.33
It is important to emphasise that there is no such
thing as the Bayesian approach, and that many
ideological differences exist between the re-
searchers whose work is discussed in this chap- WHERE DOES THE PRIOR DISTRIBUTION
ter. Four levels of Bayesian approach, of COME FROM?
increasing ‘purity’, may be identified:
1 The empirical Bayes approach, in which a There is no denying that prior beliefs exist in
prior distribution is estimated from multiple medicine. For example, in the context of clinical
experiments. Analyses and reporting are in trials, Peto and Baigent34 state that ‘it is gen-
frequentist terms. erally unrealistic to hope for large treatment
2 The reference Bayes approach, in which a effects’ but that ‘it might be reasonable to hope
Bayesian interpretation is given to the con- that a new treatment for acute stroke or acute
clusions expressed as posterior distributions, myocardial infarction could reduce recurrent
but an attempt is made to use ‘objective’ stroke or death in hospital from 10% to 9% or
prior distributions. 8%, but not to hope that it could halve in-
3 The proper Bayes approach, in which infor- hospital mortality’. However, turning informally
mative prior distributions are used based on expressed opinions into a mathematically
available evidence, but conclusions are sum- expressible prior distribution is perhaps the most
marised by posterior distributions without controversial aspect of Bayesian analysis. Four
explicit incorporation of utility functions. broad approaches are outlined below, ranging
Bayesian Methods 281
from the elicitation of subjective opinion to the Kadane and Wolfson40 distinguish between
use of statistical models to estimate the prior experiments to learn for which only the prior of
from data. Sensitivity analysis to alternative the experimenter needs to be considered, and
assumptions is considered vital, whatever the experiments to prove, in which the priors (and
method used to construct the prior distribution. utilities) of observers to be persuaded need to be
From a mathematical and computational per- considered. They discuss general methods of
spective, it is extremely convenient if the prior elicitation, and give an example applied to an
distribution is a member of a family of distribu- observational study in healthcare.
tions that is conjugate to the form of the like- The methods used in case studies can be
lihood, in the sense that they ‘fit together’ to divided into four main categories using increas-
produce a posterior distribution that is in the ingly formal methods:
same family as the prior distribution. For ex-
ample, many likelihoods for treatment effects 1 Informal discussion: examples include
can be assumed to have an approximately nor- Rosner and Berry,41 who consider a trial of
mal shape,35 and thus in these circumstances it Taxol in metastatic breast cancer, in which a
will be convenient to use a normally shaped beta prior for the overall success rate was
prior (the conjugate family), provided it approx- assessed using a mean 0.25 and 50% belief
imately summarises the appropriate external evi- that the true success rate lay between 15% and
dence. Similarly, when the observed data are to 35%. Similarly, Lilford and Braunholtz21
be proportions (implying a binomial likelihood), describe how priors were obtained from two
the conjugate prior is a member of the ‘beta’ doctors for the relative risk of venous throm-
family of distributions which provide a flexible bosis from oral contraceptives.
way of expressing beliefs about the magnitude 2 Structured interviewing and formal pooling
of a true unknown frequency. of opinion: Freedman and Spiegelhalter42
Modern computing power is, however, reduc- describe an interviewing technique in which
ing the need for conjugacy, and in this section a set of experts were individually inter-
we shall concentrate on the source of the prior viewed and hand-drawn plots of their prior
rather than its precise mathematical form. distributions elicited. Deliberate efforts were
made to prevent the opinions being over-
confident (too ‘tight’). The distributions
Elicitation of Opinion were converted to histograms and averaged
to produce a composite prior. The approach
A true subjectivist Bayesian approach only was subsequently used in a bladder cancer
requires a prior distribution that expresses the trial and these priors used later in discussing
personal opinions of an individual but, if the initial and interim power of the study43,44 –
health technology assessment is to be generally
see p. 284.
accepted by a wider community, it would appear
3 Structured questionnaires: a technique in
to be essential that the prior distributions have
some evidential or at least consensus support. In which a histogram is directly elicited via a
some circumstances there may, however, be questionnaire was used by Abrams et al.45
little ‘objective’ evidence available and sum- for a trial of neutron therapy, and by Parmar
maries of expert opinion may be indispensable. et al.46 to elicit prior distributions for the
There is a very large literature on eliciting effect of a new radiotherapy regime
subjective probability distributions from experts, (CHART). In this latter study the possible
with some good early references on statistical36 treatment effect was discretised into 5%
and psychological aspects,37 as well as on meth- bands, a distribution over these bands ob-
ods for pooling distributions obtained from mul- tained from each of nine clinicians and an
tiple experts.38 Chaloner39 provides a thorough arithmetic mean taken.
review of prior elicitation for clinical trials, 4 Computer-based elicitation: Chaloner et al.47
including interviews with clinicians, postal provide a detailed case study of the use of
questionnaires, and the use of an interactive a rather complex computer program that
computer program to draw a prior distribution. interactively elicited distributions from five
She concludes that fairly simple methods are clinicians for a trial of prophylactic therapy
adequate, using interactive feedback with a in AIDS.
scripted interview, providing experts with a sys-
tematic literature review, using 2.5% and 97.5%
percentiles, and using as many experts as pos- Summary of Evidence
sible. Berry and Stangl15 discuss methods for
eliciting conjugate priors and checking whether If the results of previous similar studies are
the assessments are consistent with each other. available it is clear that they may be used as the
282 The Advanced Handbook of Methods in Evidence Based Healthcare
basis for a prior distribution. Three main ap- 2 ‘Sceptical’ priors: Informative priors that
proaches have been used: express scepticism about large treatment
effects have been put forward both as a
1 Use of a single previous trial result as a prior
reasonable expression of doubt, and as a way
distribution: Brown et al.48 and Spiegelhalter
of controlling early stopping of trials on
et al.35 both provide examples of prior dis-
the basis of fortuitously positive results.
tributions proportional to the likelihoods
Mathematically speaking a sceptical prior
arising from a single pilot trial.
2 Direct use of a meta-analysis of many about a treatment effect will have a mean of
previous studies: Lau et al.49 point out that zero and some spread which determines the
cumulative meta-analysis can be given a degree of scepticism. Fletcher et al.53 use
Bayesian interpretation in which the prior such a prior, while Spiegelhalter et al.35
for each trial is obtained from the meta- argue that a reasonable degree of scepticism
analysis of preceding studies, while Smith may be equivalent to feeling that the altern-
et al.50 and Higgins and Whitehead51 both ative hypothesis is optimistic, formalised by
use a meta-analysis of past meta-analyses as a prior with only a small probability that the
a basis for a prior distribution for the degree treatment effect is as large as the alternative
of heterogeneity in a current meta-analysis. hypothesis.
3 Use of previous data in some discounted This approach has been used in a number
form: Previous studies may not be directly of case studies,46,54 and has been suggested
related to the one in question, and we may as a basis for monitoring trials55 (see p. 285).
wish to discount past evidence in some way. A senior FDA biostatistician56 has stated that
This approach has been most widely used in he ‘would like to see [sceptical priors]
studies with non-concurrent controls (see applied in more routine fashion to provide
p. 287, observational studies). insight into our decision making’.
3 ‘Enthusiastic’ priors: As a counter-balance
to the pessimism expressed by the sceptical
Default Priors prior, Spiegelhalter et al.35 suggest a prior
centred on the alternative hypothesis and
It would clearly be attractive to have prior with a low chance (say 5%) that the true
distributions that could be taken ‘off-the-shelf’, treatment benefit is negative. Use of such a
rather than having to consider all available evi- prior has been reported in case studies54,57
dence external to the study in their construction. and as a basis for conservatism in the face of
Four main suggestions can be identified: early negative results55 (see p. 285).
1 ‘Non-informative’ priors: There has been 4 Priors with a point mass at zero (‘lump-and-
substantial research in the Bayesian litera- smear’ priors): The traditional statistical
ture into so-called non-informative or refer- approach expresses a qualitative distinction
ence priors, which usually take the form of between the role of a null hypothesis, gen-
uniform distributions over the range of inter- erally of no treatment effect, and alternative
est. Formally, a uniform distribution means hypotheses. A prior distribution that retains
the posterior distribution has the same shape this distinction would place a ‘lump’ of
as the likelihood function, which in turn probability on the null hypothesis, and
means that the resulting Bayesian intervals ‘smear’ the remaining probability over
and estimates will essentially match the tra- the whole range of alternatives. Cornfield
ditional results, so that posterior tail areas repeatedly argued for this approach, which
will match one-sided P-values (ignoring any he termed the ‘relative betting odds’ on
adjustment for sequential analysis). the null hypothesis,58 both on a practical
Results with such reference priors are basis and as a rebuttal to the criticism of
generally quoted as one part of a Bayesian ‘sampling to foregone conclusion’ (see
analyses. In particular, Burton52 suggests p. 285). It has been used in a cancer trial59
that most doctors interpret frequentist con- and for sensitivity analysis in trial report-
fidence intervals as posterior distributions, ing60 (see p. 286). Although such an analysis
and also that information prior to a study provides an explicit probability that the null
tends to be vague, and that therefore results hypothesis is true, and so appears to answer
from a study should be presented by per- the question of interest, the prior might be
forming a Bayesian analysis with a non- somewhat more realistic were the lump to be
informative prior and quoting posterior placed on a small range of values represent-
probabilities for the parameter of interest ing the more plausible null hypothesis of ‘no
being in various regions. effective difference’.
Bayesian Methods 283
clinician’s priors and only assigning treatments is generally set to 80% or 90%, and the altern-
that at least one clinician considers optimal. ative hypothesis may best be defined as being
Papineau74 refutes Urbach’s position and claims both ‘realistic and important’. The requirement
that despite it not being essential for statistical for the trial being designed to detect a plausible
inference, experimental randomisation forms a difference has naturally led to the use of prior
vital role in drawing causal conclusions. The distributions: either the prior mean could be the
relationship between randomisation and causal alternative hypothesis or the power for each
inferences is beyond the scope of this chapter, value of u could be averaged with respect to the
but in general the need for sound experimental prior distribution to obtain an ‘expected’ or
design appears to dominate philosophical statis- ‘average’ power, which should be a more real-
tical issues.75 In fact, Berry and Kadane76 sug- istic assessment of the chance that the trial will
gest that if there are several parties who make yield a positive conclusion. It is natural to
different decisions and observe different data, express a cautionary note issued on projecting
randomisation may be a strictly optimal pro- from previous studies,82 and possible techniques
cedure since it enables each observer to draw for discounting past studies is very relevant (see
their own appropriate conclusions. pp. 287–8).
The use of power calculations is so rooted
When is it ethical to randomise? into trial design that alternatives are rarely men-
tioned. However, it is natural within a Bayesian
If we agree that randomisation is in theory context to be less concerned with hypothesis
useful, then the issue arises of when it is ethical testing and more focused on the eventual preci-
to randomise. This is closely associated with the sion of the posterior distribution of the treatment
process of deciding when to stop a trial (dis- effect. There is an extensive literature on non-
cussed further on p. 285) and is often represented power-based Bayesian sample size calculations
as a balance between individual and collective which may be very relevant for trial design.83,84
ethics.77 It has been argued that a Bayesian
model naturally formalises the individual ethical
position,78,79 in that it explicitly confronts the Monitoring and Interim Analyses
personal belief in the clinical superiority of one and Predictions
treatment.
Introduction
Specification of Null Hypotheses Whether or not to stop a trial early is a complex
ethical, financial, organisational and scientific
Attention in a trial usually focuses on the null issue, in which statistical analysis plays a con-
hypothesis of treatment equivalence expressed siderable role.77 Furthermore, the monitoring of
by u 5 0, but realistically this is often not the sequential clinical trials can be considered the
only hypothesis of interest. Increased costs, ‘front line’ between Bayesian and frequentist
toxicity and so on may mean that a certain approaches, and the reasons for their diverg-
improvement would be necessary before con- ence ‘reach to the very foundations of the two
sidering the new treatment clinically superior, paradigms’.24
and we shall denote this value uSUP. Similarly, Four main statistical approaches can be
the new treatment might not actually be con- identified:
sidered clinically inferior unless the true benefit 1 Formal frequentist: attempts to retain a fixed
were less than some threshold denoted uINF. Type I error through tightly controlled stop-
The interval between uINF and uSUP has been ping boundaries. Whitehead3,72 is a major
termed the ‘range of equivalence’:80 often uINF is proponent, and Jennison and Turnbull85 pro-
taken to be 0. vide a detailed review.
This is not a specifically Bayesian idea,65 2 Informal non-Bayesian: best exemplified in
and there are several published examples of trials influenced by Richard Peto’s group, in
the elicitation and use of such ranges of which protocols state86 that the data monitor-
equivalence.42,81 ing committee should only alert the steering
committee if there is ‘both (a) ‘proof
beyond reasonable doubt’ that for all, or for
Alternative Hypotheses, Power and some, types of patient one particular treat-
Sample Size of Trials ment is clearly indicated . . ., and (b) evi-
dence that might reasonably be expected to
Trials have traditionally been designed to have influence the patient management of many
reasonable power, defined as the chance of cor- clinicians who are already aware of the
rectly detecting an alternative hypothesis. Power results of other main studies’.
Bayesian Methods 285
3 Informal Bayesian: uses probabilities derived For example, Berry et al.97 consider a trial of
from a posterior distribution for monitoring, influenza vaccine for Navajo children, in which
without formally pre-specifying a stopping they attempt to model the consequences of con-
criterion. tinuing a trial for each successive month.
4 Decision-theoretic Bayesian: explicitly as- The level of detail required for such an analy-
sesses the losses associated with conse- sis has been criticised as being unrealistic,23,35
quences of stopping or continuing the study. but it has been argued that trade-offs between
Here we briefly describe some of the huge benefits for patients within and outside the trial
literature on this subject. should be explicitly confronted.24
confirmatory trial about whether the new treat- care organization conducting a randomised trial
ment provides a clinically important benefit. instead of continuing to treat patients in the
absence of more information.’116 Detsky117 con-
Sensitivity Analysis, Robustness ducted an early attempt to model the impact of a
trial in terms of future lives saved. In other
and Reporting examples,116,118 sample-sizes are explicitly deter-
mined by a trade-off between the cost of the trial
An integral part of any good statistical report is
a sensitivity analysis of assumptions concerning and the expected future benefit.
the form of the model (the likelihood). Bayesian More generally, the role of decision analysis
approaches have the additional concern of in health services research has been emphasised
sensitivity to the prior distribution, both in view by Lilford and Royston.119
of its controversial nature and because it is by
definition a subjective assumption that is open to
valid disagreement. There is therefore a need Data-Dependent Allocation
to consider the impact of a ‘community of
priors’.107 So far we have only covered standard random-
A ‘robust’ Bayesian approach, in which an isation designs, but a full decision-theoretic
attempt is made to characterise the class of approach to trial design would consider data-
priors leading to a given decision, is illustrated dependent allocation so that, for example, in
by Hughes108 and others.109,110 There is limited order to minimise the number of patients getting
experience of reporting such analyses in the the inferior treatment, the proportion random-
medical literature, and it has been suggested that ised to the apparently superior treatment could
a separate ‘interpretation’ section is required to be increased as the trial proceeded. An extreme
display how the data in a study would add to a example is Zelen’s ‘play-the-winner’ rule in
range of currently held opinions.35 which the next patient is given the currently
superior treatment, and randomisation is dis-
Subset and Multi-Centre Analysis pensed with entirely.120 Tamura et al.121 report
one of the few adaptive clinical trials to take
The discussion on multiplicity on pp. 277–80 has place, in patients with depressive disorder. The
already described how multiple simultaneous review by Armitage122 describes the practical
inferences may be made by assuming a common problems in carrying out such trials.
prior distribution with unknown parameters,
provided an assumption of exchangeability (the
prior does not depend on units’ identities) is Alternative Trial Designs
appropriate. Within the context of clinical
trials this has immediate relevance to the issue Equivalence Trials
of estimating treatment effects in subgroups of
patients. There is a large statistical literature on trials
Published applications are generally based on designed to establish equivalence between ther-
assigning a reference (uniform) prior on the apies. From a Bayesian perspective the solution
overall treatment effect, and then assuming the is straightforward: define a region of equiva-
subgroup-specific deviations from that overall lence (see p. 284) and calculate the posterior
effect have a common prior distribution with probability that the treatment difference lies in
zero mean. This prior expresses scepticism about this range – a threshold of 95% or 90% might be
widely differing subgroup effects. Donner111 chosen to represent strong belief in equivalence.
sets out the basic ideas, and Simon and co- Several examples of this remarkably intuitive
authors elaborate the techniques in several approach have been reported.23,123–126
examples.112,113 Explicit estimation of individual
institutional effects114,115 relates strongly to the
methods used for institutional comparisons of Crossover and repeated N-of-1 trials
patient outcomes (see p. 288). The Bayesian approach to crossover designs, in
which each patient is given two or more treat-
Cost–Effectiveness of Carrying Out ments in an order selected at random, is fully
Trials: ‘Payback Models’ reviewed by Grieve.127 More recent references
concentrate on Gibbs sampling approaches.128,129
‘By explicitly taking into consideration the costs N-of-1 studies can be thought of as repeated
and benefits of a trial, Bayesian statistical meth- crossover trials in which interest focusses on the
ods permit estimation of the value to a health response of an individual patient, and Zucker
Bayesian Methods 287
et al.130 pool multiple N-of-1 studies using a A GUIDE TO THE BAYESIAN HTA
hierarchical model. LITERATURE: OBSERVATIONAL STUDIES
Factorial designs
Since the probability models used in Bayesian
Factorial trials, in which multiple treatments are analysis do not need to be based on actual
given simultaneously to patients in a structured randomisation, non-randomised studies can be
design, can be seen as another example of mul- analysed in exactly the same manner as random-
tiplicity and hence a candidate for hierarchical ised comparisons, perhaps with extra attention
models. Simon and Freedman131 and Miller and to adjusting for co-variates in an attempt to
Seaman132 suggest suitable prior assumptions control for possible baseline differences in the
that avoid the need to decide whether inter- treatment groups. The dangers of this approach
actions do or do not exist. have, of course, been well described in the
medical literature,142 but nevertheless there are
circumstances where randomisation is either
Other Aspects of Drug Development impossible or where there is substantial valuable
information available in historical data. There is,
Other aspects of drug development include: of course, then a degree of subjective judgement
1 Pharmacokinetics: The ‘population’ ap- about the comparability of groups, which fits
proach to pharmacokinetics is well estab- well into the acknowledged judgement under-
lished, and is essentially an empirical Bayes lying all Bayesian reasoning.
procedure. Proper Bayes analysis has been The fullest Bayesian analysis of non-
extensively reported by Wakefield and randomised data for HTA is probably the Con-
colleagues.133,134 fidence Profile Method of Eddy and colleagues,143
2 Phase 1 trials: Phase 1 trials are conducted but discussion of this technique is placed within
to determine that dosage of a new treatment evidence synthesis (p. 288). In this section we
which produces a risk of a toxic response consider three aspects of non-randomised com-
which is deemed to be acceptable. The pri- parisons: direct analysis of epidemiological data
mary Bayesian contribution to the develop- including case-control studies; the use of his-
ment of methodology for Phase 1 trials has torical controls; and comparisons of institutions
been the Continual Reassessment Method on the basis of their outcomes.
developed by O’Quigley and colleagues,135
and extended by Goodman et al.136 amongst
others. A prior for a parameter underlying a Epidemiological Data
dose–toxicity curve is updated sequentially
and used to find the current ‘best’ estimate The results from an epidemiological study,
of the dosage. whether of a cohort or case-control design, pro-
3 Phase 2 trials: Phase 2 clinical trials are vide a likelihood which can be combined with
carried out in order to discover whether a prior information using standard Bayesian meth-
new treatment is promising enough to be ods. Case-control designs have been considered
submitted to a controlled Phase 3 trial, and in detail,144,145 while Lilford and Braunholz’s
often a number of doses might be compared. tutorial article is based on likelihoods arising
Bayesian work has focussed on monitoring from case-control studies.21
and sample size determination. Monitoring There is also a substantial literature on
on the basis of posterior probability of ex- Bayesian methods for complex epidemiological
ceeding a desired threshold response rate has modelling, particularly concerning problems with
been recommended by Mehta and Cain,137 measurement error and spatial correlation.146–148
while Heitjan57 adapts the proposed use of
sceptical and enthusiastic priors (see p. 285).
Thall and co-workers have also developed Using Historical Controls
stopping boundaries for Phase 2 studies
based on posterior probabilities of clinically A Bayesian basis for the use of historical con-
important events,138 although Stallard139 has trols in clinical trials, generally in addition to
criticised this approach as being demon- some contemporaneous controls, was first sug-
strably suboptimal when evaluated using a gested by Pocock,149 and has since been particu-
full decision-theoretic model with a monetary larly developed within the field of carcino-
loss function. Finally, Whitehead and col- genicity studies.150 The crucial issue is the extent
leagues140,141 have taken a full decision- to which the historical information can be con-
theoretic approach to allocating subjects sidered equivalent to contemporaneous data.
between Phase 2 and Phase 3 studies. Three broad approaches have been taken.
288 The Advanced Handbook of Methods in Evidence Based Healthcare
1 Assume the historical control groups are will be less than adequate due to economic,
exchangeable with the current control group, organisational or ethical considerations.164 Con-
and hence build or assume a hierarchical sidering all the available evidence, including
model for the response within each that from non-randomised studies may then be
group.151,152 This leads to a degree of pooling necessary or advantageous. Cross-design syn-
between the control groups, depending on thesis, an approach for synthesising evidence
their observed or assumed heterogeneity. from different sources, has been outlined165 and
2 If there is only one historical group, then subsequently applied to a data on breast cancer
assume a parameter representing the prob- screening.166,167 A Bayesian hierarchical model
ability that any past individual is exchange- approach specifically allows for the quantitative
able with current individuals, so discounting within- and between-sources heterogeneity, and
the contribution of past data to the likeli- for a priori beliefs regarding qualitative differ-
hood, as used by Berry15,153 in reanalysis of ences between the various sources of evidence.
the ECMO study. It has yet to be established when such analyses
3 Assume that the parameter being estimated are appropriate, as there is concern that includ-
in historical data is some function of the ing studies with poorer designs will weaken the
parameter that is of interest, thus explicitly analysis, though this issue is partially addressed
modelling potential biases, as in the Con- by conducting sensitivity analyses under various
fidence Profile Method.143 assumptions.
Institutional Comparisons
Confidence Profile Method
A classic ‘multiplicity’ problem arises in the use
of performance indicators to compare institu- This approach was developed by Eddy and col-
tions with regard to their health outcomes or use leagues and promulgated in a book with numer-
of particular procedures. Analogously to subset ous worked examples and accompanying soft-
estimation (p. 286), hierarchical models can be ware,143 as well as tutorial articles.22,168 They use
used to make inferences based on estimating a directed conditional independence graphs to rep-
common prior or ‘population’ distribution.154,155 resent the qualitative way in which multiple
An additional benefit of using Markov Chain contributing sources of evidence relate to the
Monte Carlo methods is the ability to derive quantity of interest, explicitly allowing the user
uncertainty intervals around the rank order of to discount studies due to their potential internal
each institution.156 bias or their limited generalisability. Their ana-
lysis is essentially Bayesian, although it is pos-
sible to avoid prior specification and use only
A GUIDE TO THE BAYESIAN HTA the likelihoods. The necessity to make explicit
LITERATURE: EVIDENCE SYNTHESIS AND subjective judgements, and the limited capacity
and friendliness of the software, has perhaps
POLICY-MAKING
limited the application of this technique.
Meta-analysis
Policy-Making
Hedges157 reviews Bayesian approaches to meta-
analysis, which is similar to that assumed for The use of Bayesian ideas in decision-making is
random effects models, but requires that prior a huge area of research and application, in which
distributions are specified for the mean effect attention is more focussed on the utility of
size and the between- and within-study vari- consequences rather than the use of Bayesian
ances. This leads to wider interval estimates. methods to revise beliefs.169 This activity blends
Several authors158–160 have implemented a hier- naturally into cost–effectiveness analysis which
archical model. Empirical Bayes approaches is outside the domain of this chapter, but never-
have received most attention in the literature theless the subjective interpretation of probabil-
until recently, largely because of computa-
ity is essential, since the expressions of uncer-
tional difficulties in the use of fully Bayesian
tainty required for a decision analysis can rarely
modelling.161–163
be based purely on empirical data.
The journal Medical Decision Making con-
Cross-Design Synthesis tains an extensive collection of policy analyses
based on maximising expected utility, some
As noted when discussing observational studies, of which particularly stress the importance of
in some circumstances randomised evidence Bayesian considerations.170
Bayesian Methods 289
IMPLICATIONS FOR FUTURE RESEARCH health technology assessment. Br. Med. J. 1999;
319: 508–12.
2. Spiegelhalter DJ, Myles JM, Jones DR, Abrams
The HTA field has been slow to adopt Bayesian KR. Bayesian methods in health technology
methods, perhaps due to a reluctance to use assessment. Health Technology Assessment
prior opinions, unfamiliarity with the tech- 2000, to appear.
niques, mathematical complexity, lack of soft-
3. Whitehead J. The Design and Analysis of
ware, or the innate conservatism of the medical
Sequential Clinical Trials, 2nd edn. Chichester:
establishment and regulatory authorities.
There are strong philosophical reasons for J Wiley & Sons, 1997.
using Bayesian approaches in HTA, but the 4. Senn S. Statistical Issues in Drug Development.
current literature emphasises the practical claims Chichester: J. Wiley & Sons, 1997.
of being able to handle complex inter-related 5. Rothman KJ. No adjustments are needed for
problems, as well as making explicit and multiple comparisons. Epidemiology 1990; 1:
accountable what is usually implicit and hidden, 43–6.
and so to clarify discussions and disagreements. 6. Perneger TV. What’s wrong with Bonferroni
The main advantage is that Bayesian analyses adjustments. Br. Med. J. 1998; 316: 1236–8.
can tell us directly what we want to know: how 7. GREAT group. Feasibility, safety and efficacy
should this piece of evidence change what we of domiciliary thrombolysis by general practi-
currently believe? tioners: Grampian region early anisteplase trial.
There are clearly a number of perceived prob- Br. Med. J. 1992; 305: 548–53.
lems with the Bayesian approach, largely con- 8. Bayes T. An essay towards solving a problem in
cerning the source of the prior and the inter- the doctrine of chances. Phil. Trans. Roy. Soc.
pretations of the conclusions. There are also 1763; 53: 418.
practical difficulties in implementation and soft- 9. Pocock SJ, Spiegelhalter DJ. Domiciliary throm-
ware. It seems appropriate, therefore, that any bosis by general practitioners. Br. Med. J. 1992;
serious introduction of Bayesian principles into 305: 1015.
HTA should be gradual and in parallel with 10. Cooper MW. Should physicians be Bayesian
traditional approaches, and in such a way as to agents? Theoret. Med. 1992; 13: 349–61.
facilitate examination of the sensitivity of results 11. Staquet MJ, Rozencweig M, Von Hoff DD,
to prior distributions. This is emphasised in the Muggia FM. The delta and epsilon errors in the
international statistical guidelines for submis- assessment of cancer clinical trials. Cancer
sions to regulatory agencies, in which it is stated Treatment Rep. 1979; 63: 1917–21.
that ‘the use of Bayesian and other approaches 12. Pocock S. Clinical Trials: A Practical Approach.
may be considered when the reasons for their Chichester: Wiley, 1983.
use are clear and when the resulting conclusions
13. Diamond GA, Forrester JS. Clinical trials and
are sufficiently robust’.171
statistical verdicts: probable grounds for appeal.
Practical methodological developments to
Ann. Intern. Med. 1983; 98: 385–94.
promote the appropriate introduction and exten-
sion of the use of Bayesian methods in HTA 14. Browner WS, Newman TB. Are all significant p
include: values created equal? The analogy between
diagnostic tests and clinical research. JAMA
(a) preparation of an extended set of case 1987; 257: 2459–63.
studies showing practical aspects of the 15. Berry DA, Stangl DK. Bayesian methods in
Bayesian approach, in particular for predic- health-related research. In: Berry DA, Stangl
tion and handling multiple substudies, in DK (eds). Bayesian Biostatistics. New York:
which mathematical details are minimised, Marcel Dekker, 1996, pp. 3–66.
and 16. Healy MJR. Probability and decisions. Arch.
(b) development of standards for the perform- Dis. Child. 1994; 71: 90–4.
ance and reporting of Bayesian analyses, 17. Simon R. Problems of multiplicity in clinical
the first steps towards which – the Bayes- trials. J. Statist. Planning Inference 1994; 42:
Watch check-list – are reported in the full
209–21.
report of the project on which this chapter
18. Gelman A, Rubin DB. Markov chain Monte
is based.2
Carlo methods in biostatistics. Statist. Methods
Med. Res. 1996; 5: 339–55.
19. Berry DA, Stangl DK (eds). Bayesian Bio-
REFERENCES statistics. New York: Marcel Dekker, 1996.
20. Lewis RJ, Wears RL. An introduction to
1. Spiegelhalter DJ, Myles JM, Jones DR, Abrams the Bayesian analysis of clinical trials. Ann.
KR. An introduction to Bayesian methods in Emergency Med. 1993; 22: 1328–36.
290 The Advanced Handbook of Methods in Evidence Based Healthcare
21. Lilford RJ, Braunholtz D. For debate – the 41. Rosner GL, Berry DA. A Bayesian group
statistical basis of public policy – a paradigm sequential design for a multiple arm randomized
shift is overdue. Br. Med. J. 1996; 313: 603–7. clinical trial. Statist. Med. 1995; 14: 381–94.
22. Eddy DM, Hasselblad V, Shachter R. A 42. Freedman LS, Spiegelhalter DJ. The assessment
Bayesian method for synthesizing evidence: The of subjective opinion and its use in relation to
confidence profile method. Int. J. Technol. stopping rules for clinical trials. The Statistician
Assess. Health Care 1990; 6: 31–55. 1983; 32: 153–60.
23. Breslow N. Biostatistics and Bayes. Statist. Sci. 43. Spiegelhalter DJ, Freedman LS. A predictive
1990; 5: 269–84. approach to selecting the size of a clinical trial,
24. Etzioni RD, Kadane JB. Bayesian statistical based on subjective clinical opinion. Stat. Med.
methods in public health and medicine. Annu. 1986; 5: 1–13.
Rev. Public Health 1995; 16: 23–41. 44. Spiegelhalter DJ. Probabilistic prediction in
25. Freedman L. Bayesian statistical methods – a patient management and clinical trials. Stat.
natural way to assess clinical evidence. Br. Med. Med. 1986; 5: 421–33.
J. 1996; 313: 569–70. 45. Abrams K, Ashby D, Errington D. Simple
26. Cornfield J. The Bayesian outlook and its appli- Bayesian analysis in clinical trials – a tutorial.
cations. Biometrics 1969; 25: 617–57. Controlled Clinical Trials 1994; 15: 349–59.
27. Edwards WL, Lindman H, Savage LJ. Bayesian 46. Parmar MKB, Spiegelhalter DJ, Freedman
statistical inference for psychological research. LS. The CHART trials: Bayesian design and
Psychol. Rev. 1963; 70: 193–242. monitoring in practice. Stat. Med. 1994; 13:
28. Lindley DV. Making Decisions, 2nd edn. 1297–312.
Chichester: J. Wiley & Sons, 1985. 47. Chaloner K, Church T, Louis TA, Matts JP.
29. Barnett V. Comparative Statistical Inference Graphical elicitation of a prior distribution for a
(Second Edition). Chichester: J. Wiley & Sons, clinical trial. Statistician 1993; 42: 341–53.
1982. 48. Brown BW, Herson J, Atkinson EN, Rozell ME.
30. Lee PM. Bayesian Statistics: an Introduction.
Projection from previous studies – a Bayesian
London: Edward Arnold, 1989.
and frequentist compromise. Controlled Clinical
31. Gelman A, Carlin J, Stern H, Rubin DB.
Trials 1987; 8: 29–44.
Bayesian Data Analysis. New York: Chapman &
49. Lau J, Schmid CH, Chalmers TC. Cumulative
Hall, 1995.
meta-analysis of clinical trials builds evidence
32. Carlin BP, Louis TA. Bayes and Empirical
for exemplary medical care. J. Clin. Epidemiol.
Bayes Methods for Data Analysis. London:
1995; 48: 45–57.
Chapman & Hall, 1996.
50. Smith TC, Spiegelhalter DJ, Parmar MKB.
33. Bernardo JM, Smith AFM. Bayesian Theory.
Bayesian meta-analysis of randomized trials
Chichester: J. Wiley & Sons, 1994.
using graphical models and BUGS. In: Berry
34. Peto R, Baigent C. Trials: the next 50 years. Br.
Med. J. 1998; 317: 1170–1. DA, Stangl DK (eds). Bayesian Biostatistics.
35. Spiegelhalter DJ, Freedman LS, Parmar MKB. New York: Marcel Dekker, 1996, pp. 411–27.
Bayesian approaches to randomized trials. J. 51. Higgins JP, Whitehead A. Borrowing strength
Roy. Statist. Soc. Series A – Statistics Society from external trials in a meta-analysis. Stat.
1994; 157: 357–87. Med. 1996; 15: 2733–49.
36. Savage LJ. Elicitation of personal probabilities 52. Burton PR. Helping doctors to draw appropriate
and expectations. J. Am. Statist. Assoc. 1971; 66: inferences from the analysis of medical studies.
783–801. Stat. Med. 1994; 13: 1699–713.
37. Tversky A. Assessing uncertainty (with dis- 53. Fletcher A, Spiegelhalter D, Staessen J, Thijs L,
cussion). J. Roy. Statist. Soc. B 1974; 36: Bulpitt C. Implications for trials in progress of
148–59. publication of positive results. Lancet, 1993;
38. Genest C, Zidek JV. Combining probability dis- 342: 653–7.
tributions: a critique and an annotated biblio- 54. Freedman LS, Spiegelhalter DJ, Parmar MKB.
graphy (with discussion). Statist. Sci. 1986; 1: The what, why and how of Bayesian clinical trials
114–48. monitoring. Stat. Med. 1994; 13: 1371–83.
39. Chaloner K. Elicitation of prior distributions. In: 55. Fayers PM, Ashby D, Parmar MKB. Bayesian
Berry DA, Stangl DK (eds). Bayesian Bio- data monitoring in clinical trials. Stat. Med.
statistics. New York: Marcel Dekker, 1996, 1997; 16: 1413–30.
pp. 141–56. 56. O’Neill RT. Conclusions: 2. Stat. Med. 1994;
40. Kadane JB, Wolfson LJ. Priors for the design 13: 1493–9.
and analysis of clinical trials. In: Berry DA, 57. Heitjan DF. Bayesian interim analysis of phase
Stangl DK (eds). Bayesian Biostatistics. New ii cancer clinical trials. Stat. Med. 1997; 16:
York: Marcel Dekker, 1996, pp. 157–84. 1791–802.
Bayesian Methods 291
58. Cornfield J. Recent methodological contribu- 80. Freedman LS, Lowe D, Macaskill P. Stopping
tions to clinical trials. Am. J. Epidemiol. 1976; rules for clinical trials incorporating clinical
104: 408–21. opinion. Biometrics 1984; 40: 575–86.
59. Freedman LS, Spiegelhalter DJ. Application of 81. Fleming TR Watelet LF. Approaches to mon-
Bayesian statistics to decision making during a itoring clinical trials. J. Natl Cancer Inst. 1989;
clinical trial. Stat. Med. 1992; 11: 23–35. 81: 188–93.
60. Hughes MD. Reporting Bayesian analyses of 82. Korn EL. Projection from previous studies. A
clinical trials. Stat. Med. 1993; 12: 1651–63. caution [letter; comment]. Controlled Clinical
61. Louis TA. Using empirical Bayes methods in Trials 1990; 11: 67–9.
biopharmaceutical research. Stat. Med. 1991; 10: 83. Hutton JL, Owens RG. Bayesian sample size
811–29. calculations and prior beliefs about child sexual
62. Brophy JM, Joseph L. Placing trials in context abuse. Statistician 1993; 42: 399–404.
using Bayesian analysis – GUSTO revisited by 84. Adcock CJ. The Bayesian approach to determi-
Reverend Bayes. JAMA 1995; 273: 871–5. nation of sample sizes – some comments on the
63. Kadane JB. Prime time for Bayes. Controlled paper by Joseph, Wolfson and Du Berger. J.
Clinical Trials 1995; 16: 313–18. Roy. Statist. Soc. Series D 1995; 44: 155–61.
64. Pocock SJ, Hughes MD. Estimation issues in 85. Jennison C, Turnbull BW. Statistical approaches
clinical trials and overviews. Stat. Med. 1990; 9: to interim monitoring of medical trials: a review
657–71. and commentary. Statist. Sci. 1990; 5: 299–317.
65. Armitage PA. Inference and decision in clinical 86. Collins R, Peto R, Flather M, and ISIS-4 Col-
trials. J. Clin. Epidemiol. 1989; 42: 293–9. laborative Group. ISIS-4 – a randomized factorial
66. Lilford RJ, Thornton JG, Braunholtz D. Clinical trial assessing early oral captopril, oral mono-
trials and rare diseases – a way out of a conun- nitrate, and intravenous magnesium- sulfate in
drum. Br. Med. J. 1995; 311: 1621–5. 58,050 patients with suspected acute myocardial
67. Matthews JNS. Small clinical trials – are they all infarction. Lancet 1995; 345: 669–85.
87. Berry DA. Monitoring accumulating data in a
bad? Stat. Med. 1995; 14: 115–26.
clinical trial. Biometrics 1989; 45: 1197–211.
68. Spiegelhalter DJ, Freedman LS, Parmar MKB.
88. Carlin BP, Chaloner K, Church T, Louis TA,
Applying Bayesian ideas in drug development
Matts JP. Bayesian approaches for monitoring
and clinical trials. Stat. Med. 1993; 12: 1501–17.
clinical trials with an application to toxoplasmic
69. Berry DA. A case for Bayesianism in clinical
encephalitis prophylaxis. Statistician 1993; 42:
trials. Stat. Med. 1993; 12: 1377–93.
355–67.
70. Berry DA. Decision analysis and Bayesian
89. George SL, Li CC, Berry DA, Green MR.
methods in clinical trials. Cancer Treatment Res.
Stopping a clinical trial early – frequentist and
1995; 75: 125–54. Bayesian approaches applied to a CALGB trial
71. Urbach P. The value of randomization and in non-small-cell lung cancer. Stat. Med. 1994;
control in clinical trials. Stat. Med. 1993; 12: 13: 1313–27.
1421–31. 90. Dersimonian R. Meta-analysis in the design and
72. Whitehead J. The case for frequentism in clin- monitoring of clinical trials. Stat. Med. 1996; 15:
ical trials. Stat. Med. 1993; 12: 1405–19. 1237–48.
73. Kadane JB. Progress toward a more ethical 91. Halperin M, Lan KKG, Ware JH, Johnson NJ,
method for clinical trials. J. Med. Philos. 1986; DeMets DL. An aid to data monitoring in long-
11: 385–404. term clinical trials. Controlled Clinical Trials
74. Papineau D. The virtues of randomization. Br. J. 1982; 3: 311–23.
Philos. Sci. 1994; 45: 437–50. 92. Spiegelhalter DJ, Freedman LS, Blackburn PR.
75. Hutton JL. The ethics of randomised controlled Monitoring clinical trials: Conditional of pre-
trials: a matter of statistical belief? Health Care dictive power? Controlled Clinical Trials 1986;
Analysis 1996; 4: 95–102. 7: 8–17.
76. Berry SM, Kadane JB. Optimal Bayesian ran- 93. Choi SC, Pepple PA. Monitoring clinical trials
domization. J. Roy. Statist. Soc. Series B. Meth- based on predictive probability of significance.
odological 1997; 59: 813–19. Biometrics 1989; 45: 317–23.
77. Pocock S. Statistical and ethical issues in mon- 94. Quian J, Stangl DK, George S. A Weibull model
itoring clinical-trials. Br. Med. J. 1992; 305: for survival data: Using prediction to decide
235–40. when to stop a clinical trial. In: Berry DA,
78. Lilford RJ, Jackson J. Equipoise and the ethics Stangl DK (eds). Bayesian Biostatistics. New
of randomization. J. R. Soc. Med. 1995; 88: York: Marcel Dekker, 1996, Chapter 6.
552–9. 95. Ware JH, Muller JE, Braunwald E. The futility
79. Palmer CR. Ethics and statistical methodology index: an approach to the cost-effective termina-
in clinical trials. J. Medical Ethics 1993; 19: tion of randomized clinical trials. Am. J. Med.
219–22. 1985; 78: 635–43.
292 The Advanced Handbook of Methods in Evidence Based Healthcare
96. Korn EL, Simon R. Data monitoring committees 113. Simon R, Dixon DO, Friedlin B. Bayesian sub-
and problems of lower than expected accrual or set analysis of a clinical trial for the treatment
event rates. Controlled Clinical Trials 1996; 17: of HIV infections. In: Berry DA, Stangl DK
526–35. (eds). Bayesian Biostatistics. New York: Marcel
97. Berry DA, Wolff MC, Sack D. Decision making Dekker, 1996, pp. 555–76.
during a phase III randomized controlled trial. 114. Skene AM, Wakefield JC. Hierarchial models
Controlled Clinical Trials 1994; 15: 360–78. for multicentre binary response studies. Stat.
98. Ho CH. Some frequentist properties of a Med. 1990; 9: 919–29.
Bayesian method in clinical trials. Biometric. J. 115. Stangl D. Hierarchical analysis of continuous-
1991; 33: 735–40. time survival models. In: Berry DA, Stangl DK
99. Lewis RJ, Berry DA. Group sequential clinical (eds). Bayesian Biostatistics. New York: Marcel
trials – a classical evaluation of Bayesian Dekker, 1996, pp. 429–50.
decision-theoretic designs. J. Am. Statist. Assoc. 116. Hornberger J, Eghtesady P. The cost-benefit of a
1994; 89: 1528–34. randomized trial to a health care organization.
100. Lewis RJ. Bayesian hypothesis testing: Interim Controlled Clinical Trials 1998; 19: 198–211.
analysis of a clinical trial evaluating phenytoin 117. Detsky AS. Using economic-analysis to deter-
for the prophylaxis of early post-traumatic mine the resource consequences of choices made
seizures in children. In: Berry DA, Stangl DK in planning clinical-trials. J. Chron. Dis. 1985;
(eds). Bayesian Biostatistics. New York: Marcel 38: 753–65.
Dekker, 1996, pp. 279–96. 118. Claxton K, Posnett J. An economic approach to
101. Grossman J, Parmar MKB, Spiegelhalter DJ, clinical trial design and research priority-setting.
Freedman LS. A unified method for monitoring Health Economics 1996; 5: 513–24.
and analysing controlled trials. Stat. Med. 1994; 119. Lilford R, Royston G. Decision analysis in the
13: 1815–26. selection, design and application of clinical and
102. Armitage P, Mcpherson K, Rowe BC. Repeated health services research. J. Health Serv. Res.
Policy 1998; 3: 159–66.
significance tests on accumulating data. J. Roy.
120. Zelen M. Play the winner rule and the controlled
Statist. Soc. A 1969; 132: 235–44.
clinical trial. J. Am. Statist. Assoc. 1969; 64:
103. Cornfield J. A Bayesian test of some classical
131–46.
hypotheses – with applications to sequential
121. Tamura RN, Faries DE, Andersen JS, Heiligen-
clinical trials. J. Am. Statist. Assoc. 1966; 61:
stein JH. A case-study of an adaptive clinical
577–94.
trial in the treatment of out-patients with depres-
104. Armitage P. Discussion of Breslow (1990).
sive disorder. J. Am. Statist. Assoc. 1994; 89:
Statist. Sci. 1990; 5: 284–6.
768–76.
105. Berry DA. When is a confirmatory randomized 122. Armitage P. The search for optimality in clinical
clinical trial needed? J. Natl Cancer Inst. 1996; trials. Int. Statist. Rev. 1985; 53: 15–24.
88: 1606–7. 123. Selwyn MR, Dempster AP, Hall NR. A
106. Parmar MKB, Ungerleider RS, Simon R. Bayesian approach to bioequivalence for the
Assessing whether to perform a confirmatory 2 3 2 changeover design. Biometrics 1981; 37:
randomized clinical trial. J. Natl Cancer Inst. 11–21.
1996; 88: 1645–51. 124. Racine A, Grieve AP, Fluhler H, Smith AFM.
107. Kass RE, Greenhouse JB. Comments on ‘Invest- Bayesian methods in practice – experiences in
igating Therapies of potentially great benefit: the pharmaceutical industry. Appl. Statist. J.
ECMO’ by JH Ware. Statist. Sci. 1989; 4: Roy. Statist. Soc. Series C 1986; 35: 93–150.
310–17. 125. Grieve AP. Evaluation of bioequivalence studies.
108. Hughes MD. Practical reporting of Bayesian Eur. J. Clin. Pharmacol. 1991; 40: 201–2.
analyses of clinical trials. Drug Information J. 126. Baudoin C, O’Quigley J. Symmetrical intervals
1991; 25: 381–93. and confidence intervals. Biometrical J. 1994;
109. Greenhouse JB, Wasserman L. Robust Bayesian 36: 927–34.
methods for monitoring clinical trials. Stat. Med. 127. Grieve AP. Bayesian analyses of two-treatment
1995; 14: 1379–91. crossover studies. Statist. Methods Med. Res.
110. Carlin BP, Sargent DJ. Robust Bayesian ap- 1994; 3: 407–29.
proaches for clinical trial monitoring. Stat. Med. 128. Forster JJ. A Bayesian approach to the analysis
1996; 15: 1093–106. of binary crossover data. Statistician 1994; 43:
111. Donner A. A Bayesian approach to the inter- 61–8.
pretation of subgroup results in clinical trials. J. 129. Albert J, Chib S. Bayesian modelling of binary
Chron. Dis. 1982; 35: 429–35. repeated measures data with application to
112. Simon R. Statistical tools for subset analysis in crossover trials. In: Berry DA, Stangl DK
clinical trials. Rec. Res. Cancer Res. 1988; 111: (eds). Bayesian Biostatistics. New York: Marcel
55–66. Dekker, 1996, pp. 577–600.
Bayesian Methods 293
130. Zucker DR, Schmid CH, McIntosh MW, using conditional independence models. Am. J.
Agostino RB, Selker HP, Lau J. Combining Epidemiol. 1993; 138: 430–42.
single patient (n-of-1) trials to estimate popula- 147. Bernardinelli L, Clayton D, Pascutto C,
tion treatment effects and to evaluate individual Montomoli C, Ghislandi M, Songini M. Bayesian
patient responses to treatment. J. Clin. Epi- analysis of space-time variation in disease risk.
demiol. 1997; 50: 401–10. Stat. Med. 1995; 14: 2433–43.
131. Simon R, Freedman LS. Bayesian design and 148. Ashby D, Hutton JL. Bayesian epidemiology.
analysis of two 3 two factorial clinical trials. In: Berry DA, Stangl DK (eds). Bayesian Bio-
Biometrics 1997; 53: 456–64. statistics. New York: Marcel Dekker, 1996,
132. Miller MA, Seaman JW. A Bayesian approach pp. 109–38.
to assessing the superiority of a dose combina- 149. Pocock S. The combination of randomised and
tion. Biometrical J. 1998; 40: 43–55. historical controls in clinical trials. J. Chron.
133. Racine-Poon A, Wakefield J. Bayesian analysis Dis. 1976; 29: 175–88.
of population pharmacokinetic and instantan- 150. Ryan LM. Using historical controls in the analy-
eous pharmacodynamic relationships. In: Berry sis of developmental toxicity data. Biometrics
DA, Stangl DK (eds). Bayesian Biostatistics. 1993; 49: 1126–35.
New York: Marcel Dekker, 1996, pp. 321–54. 151. Tarone RE. The use of historical control infor-
134. Wakefield J, Bennett J. The Bayesian modeling mation in testing for a trend in proportions.
of covariates for population pharmacokinetic Biometrics 1982; 38: 215–20.
models. J. Am. Statist. Assoc. 1996; 91: 917–27. 152. Dempster AP, Selwyn MR, Weeks BJ. Com-
135. O’Quigley J, Pepe M, Fisher L. Continual re- bining historical and randomized controls for
assessment method – a practical design for assessing trends in proportions. J. Am. Statist.
phase I clinical trials in cancer. Biometrics 1990; Assoc. 1983; 78: 221–7.
46: 33–48. 153. Berry DA, Hardwick J. Using historical controls
136. Goodman SN, Zahurak ML, Piantadosi S. Some in clinical trials: application to ECMO. In: Berger
practical improvements in the continual re- JO, Gupta S (eds). Statistical Decision Theory
assessment method for phase I studies. Stat.
and Related Topics V. New York: Springer-
Med. 1995; 14: 1149–61.
Verlag, 1993, pp. 141–56.
137. Mehta CR, Cain KC. Charts for the early stop-
154. Goldstein H, Spiegelhalter DJ. Statistical aspects
ping of pilot studies. J. Clin. Oncol. 1984; 2:
of institutional performance: league tables and
676–82.
their limitations (with discussion). J. Roy. Statist.
138. Thall PF, Simon R. A Bayesian approach to
Soc. A 1996; 159: 385–444.
establishing sample size and monitoring criteria
155. Normand SL, Glickman ME, Gatsonis CA.
for phase ii clinical trials. Controlled Clinical
Statistical methods for profiling providers of
Trials 1994; 15: 463–81.
139. Stallard N. Sample size determination for medical care: Issues and applications. J. Am.
phase II clinical trials based on bayesian decision Statist. Assoc. 1997; 92: 803–14.
theory. Biometrics 1998; 54: 279–94. 156. Marshall EC, Spiegelhalter DJ. League tables of
140. Whitehead J. Designing phase II studies in the in-vitro fertilisation clinics: how confident can
context of a program of clinical research. Bio- we be about the rankings? Br. Med. J. 1998; 317:
metrics 1985; 41: 373–83. 1701–4.
141. Brunier HC, Whitehead J. Sample sizes for phase 157. Hedges LV. Bayesian meta-analysis. In: Everitt
II clinical trials derived from Bayesian decision BS, Dunn G (eds). Statistical Analysis of
theory. Stat. Med. 1994; 13: 2493–502. Medical Data: New Developments. London:
142. Byar DP, Simon RM, Friedewald WT, Schles- Arnold, 1998, pp. 251–76.
selman JJ, DeMets DL, Ellenberg JH, Gail MH, 158. DuMouchel W, Berry DA. Bayesian meta-
Ware JH. Randomized clinical trials. Perspec- analysis. In: DuMouchel W, Berry DA (eds).
tive on some recent ideas. N. Engl. J. Med. Statistical Methodology in the Pharmaceutical
1976; 295: 74–80. Sciences. New York: Marcel Dekker, 1989,
143. Eddy DM, Hasselblad V, Shachter R. Meta- pp. 509–29.
Analysis by the Confidence Profile Method: the 159. Smith TC, Spiegelhalter DJ, Thomas A.
Statistical Synthesis of Evidence. San Diego: Bayesian approaches to random-effects meta-
Academic Press, 1992. analysis: a comparative study. Stat. Med. 1995;
144. Zelen M, Parker RA. Case control studies and 14: 2685–99.
Bayesian inference. Stat. Med. 1986; 5: 261–9. 160. Abrams K, Sansó. Approximate Bayesian infer-
145. Ashby D, Hutton J, McGee M. Simple Bayesian ence for random effects meta-analysis. Stat.
analyses for case-control studies in cancer epi- Med. 1998; 17: 201–18.
demiology. The Statistician 1993; 42: 385–97. 161. Raudenbush SW, Bryk AS. Empirical Bayes
146. Richardson S, Gilks WR. A Bayesian approach meta-analysis. J. Educational Statist. 1985; 10:
to measurement error problems in epidemiology 75–98.
294 The Advanced Handbook of Methods in Evidence Based Healthcare
162. Stijnen T, Van Houwelingen JC. Empirical cancer screening. Technical Report 95-02,
Bayes methods in clinical trials meta-analysis. Department of Epidemiology and Public Health,
Biometrical J. 1990; 32: 335–46. University of Leicester, 1995.
163. Carlin JB. Meta-analysis for 2 3 2 tables – 168. Shachter R, Eddy DM, Hasselblad V. An influ-
a Bayesian approach. Stat. Med. 1992; 11: ence diagram approach to medical technology
141–58. assessment. In Oliver RM, Smith JQ (eds).
164. Black N. Why we need observational studies to Influence Diagrams, Belief Nets and Decision
evaluate the effectiveness of health care. Br. Analysis. Chichester: Wiley, 1990, pp. 321–50.
Med. J. 1996; 312: 1215–18. 169. Weinstein MC, Fineberg HV. Clinical Decision
165. General Accounting Office. Cross design syn- Analysis. Philadelphia: W.B. Saunders, 1980.
thesis: a new strategy for medical effectiveness 170. Parmigiani G, Samsa GP, Ancukiewicz M,
research. Washington, DC: General Accounting Lipscomb J, Hasselblad V, Matchar DB. Assess-
Office, 1992. ing uncertainty in cost-effectiveness analyses:
166. Abrams K, Jones DR. Meta-analysis and the application to a complex decision model. Medical
synthesis of evidence. IMA J. Mathematics Appl. Decision Making 1997; 17: 390–401.
Med. Biol. 1995; 12: 297–313. 171. International Conference on Harmonisation.
167. Smith TC, Abrams KR, Jones DR. Using hierar- Statistical Principles for Clinical Trials, 1998.
chical models in generalised synthesis of evi- Available on http://www.ipfma.org/pdfifpma/
dence: an example based on studies of breast e9.pdf.
17
Methods for Evaluating Organisation- or
Area-Based Health Interventions
the level of health service organisation or in 3 Individual-level outcomes are often corre-
geographical or administrative areas. Evaluation lated within organisational clusters.
of organisation-based interventions presents a
number of problems and the use of randomisa-
tion is currently neither well accepted nor Including Adequate Numbers of
widely practised. This chapter reviews the ways Organisational Units
that the problems associated with organisation-
based evaluation may be addressed; it is particu- Organisation-based evaluations typically include
larly concerned with quantitative studies that only small numbers of organisational clusters.
measure the effectiveness of organisational However, because differences between clusters
interventions by evaluating their outcomes. must be taken into account in evaluating the
effect of an intervention, the validity of the
study will be more dependent on the number of
clusters than on the number of individuals. One
PROBLEMS OFORGANISATION-BASED of the challenges presented by organisation-
EVALUATIONS based evaluations is to include a sufficient num-
ber of organisational units in order to allow the
Hierarchical Structure of Data in study to give valid, generalisable results.
Health Services Research
Distinction Between Cluster-Level and
In clinical research, individual patient outcomes Individual-Level Evaluation
are usually considered as independent observa-
tions at a single level. Data from organisation- A key problem in organisation-based evaluation
based evaluations often have a hierarchical is that the level of measurement is often differ-
structure. Observations may be made at two ent from the level of intervention and unit of
or more levels. Responses at lower levels are assignment. Organisation-level interventions are
nested or clustered within higher levels, and easier to evaluate when inferences from evalu-
individual outcomes are often correlated within ation are intended at organisation level because
clusters. the organisational clusters can then be con-
Clustering is an almost universal feature of sidered as the units of analysis. For example,
social organisation. When considering com- the consequences of different methods of fund-
munity based health interventions, individual ing primary care services might be evaluated
healthy subjects can be considered to be clus- according to whether certain types of practice
tered within health authorities, towns, schools, had better facilities. However, organisation-level
workplaces or neighbourhoods. For healthcare interventions are often designed to modify indi-
interventions, individual patients can be con- vidual level outcomes. For example, a study
sidered to be clustered within general practices, might aim to determine whether setting up a
hospital clinics or hospital wards. Sometimes general practice asthma clinic resulted in fewer
there are several levels of clustering. For ex- symptoms and better respiratory outcomes in the
ample, children are clustered within classes, practice’s patients. In this instance intervention
which in turn are nested within schools, and is at cluster – that is at practice – level, but
schools within local authorities. Patients are inferences are intended at individual level.1 This
clustered according to their general practitioner, leads to the ‘unit of analysis problem’ where
general practitioners are nested within general standard individual level analyses are performed
practices, which in turn are nested within health inappropriately in an organisation-based evalu-
authorities. ation. Whiting-O’Keefe et al. found that this
type of error was present in 20 out of 28
healthcare evaluations which they studied.2
Three Problems
The hierarchical structure of data from Correlation of Responses Within
organisation-based evaluations is associated with Organisations
three problems:
Standard statistical methods are based on the as-
1 It may be difficult to study an adequate sumption that individual responses are independ-
number of organisational clusters. ent of one another. In the context of organisation-
2 The level of intervention (and unit of based evaluations, responses are rarely inde-
assignment) may differ from the level of pendent because areas and organisations tend to
evaluation. contain individuals who are more similar to each
Organisation- or Area-Based Evaluations 297
other than they are to individuals in other areas best evidence into a series of methodological
or organisations. In other words, because indi- recommendations.10
vidual responses within an organisational cluster Two areas of study were particularly relevant
usually show some correlation, variation among to the present review. From the field of cardio-
individuals in different organisations may be vascular disease prevention, work on the design
greater than variation among individuals within and analysis of community intervention trials
the same organisation.1–3 There are at least three had particular relevance to the design of inter-
reasons why this might be the case.3–6 Firstly, vention studies in health services research.11
subjects may select the organisational cluster to From the field of educational research, work on
which they belong. For example, people with the analysis of hierarchical data in the assess-
similar characteristics may tend to move into ment of school performance was particularly
similar neighbourhoods. Patients may choose relevant to the evaluation of existing healthcare
their doctor according to characteristics such as services.12
age, gender or ethnic group. Secondly, organisa- The review is divided into three sections. In
tion level variables might influence all cluster the next section we review factors affecting the
members in a similar direction and each cluster choice of study design. We then discuss sample
will be subject to distinct influences. For ex- size estimation for organisation-based evalu-
ample, patient outcomes may differ systematic- ations. Finally, we review different approaches
ally among surgeons performing the same pro- to analysis of data from organisation-based
cedure.7 Thirdly, individuals within clusters may evaluation.
interact, influence each other and tend to con-
form, as might happen in a school classroom or
a psychotherapeutic group.
In general, because individual members of the A CLASSIFICATION OF STUDY DESIGNS
same cluster are unlikely to be independent with
respect to health outcomes,8 standard statistical Healthcare interventions may be evaluated either
methods for calculating sample sizes, assessing in the context of experimental studies, some-
power and analysing data are not appropriate for times called intervention studies, or in the
use in organisation-based evaluations. Use of context of observational designs (Figure 1).
standard methods may result in the required The former should be used to evaluate new
sample size being underestimated and standard interventions and the latter are more appro-
methods of analysis will give confidence inter- priately used to evaluate the quality of existing
vals that are too narrow and P-values that are services and to generate hypotheses for evalu-
too small.9 ation in randomised studies. In its epidemio-
logical sense the term ‘observational’ refers to
studies in which there is no intervention. In
METHODS order to avoid confusion, it is important to make
the distinction between healthcare interventions
and experimental interventions. Observational
The focus for this review included both random- studies of healthcare interventions do not incor-
ised and non-randomised evaluations of existing porate experimental interventions initiated by
services, but we did not include qualitative the investigator.
methods nor methods of health economic evalu-
ation. The literature search was restricted to the
English language. References were retrieved Minimum Number of Clusters
from personal collections, from hand-searches
of relevant journals, from the electronic data- Health services researchers are often asked to
bases Medline, EMBASE and ERIC (Education evaluate interventions implemented in single
Resources Information Centre). Limited searches clusters. In the field of coronary heart disease
were also made on the world wide web for prevention, the North Karelia project provides a
literature and multilevel modelling software. well-known example.13 This study, carried out
Because the search process identified a very in Finland, aimed to reduce coronary heart dis-
large number of citations, these were initially ease risk factors by a range of health promotion
screened for relevance to the focus of the interventions carried out in a single intervention
review. Relevant papers were retrieved, assessed county in comparison with a single control
against conventional epidemiological and stat- county. This type of evaluation is sometimes
istical principles and qualitative judgements strengthened by including more than one control
were made of their validity. We synthesised the cluster, as illustrated in a recent evaluation of a
298 The Advanced Handbook of Methods in Evidence Based Healthcare
Uncontrolled studies are therefore not recom- biased by unknown confounders or by residual
mended for use in health technology assessment confounding. For this reason random allocation
(see Figure 1 and Table 1). of clusters is to be preferred in health tech-
Non-randomised controlled studies with nology assessment, and convincing justification
before- and after- evaluation in both the inter- should always be given for the use of non-
vention and control groups (sometimes known randomised designs. A detailed discussion of
as ‘quasi-experimental’ studies) can be used to non-randomised designs was provided in Cook
separate the effects of intervention from secular and Campbell’s book19 and a range of examples
trends. The major limitation of such studies is are provided in our full report.20
that baseline characteristics may differ between In the simplest form of randomised design, all
the intervention and control groups, thus con- clusters are independently randomised to inter-
founding both the intervention effect and the vention or control groups. We noted earlier that
secular trend. The effects of confounding can be organisation-based evaluations often include
reduced by using restricted forms of allocation, relatively small numbers of clusters. Whereas
for example, by matching or stratifying clusters individually randomised trials often include
according to important prognostic factors before sufficiently large numbers to preclude import-
allocating them to intervention and control ant imbalances between treatment and control
groups. Allowance can also be made for con- groups, this is rarely the case for organisation-
founding at the time of analysis. Nevertheless, based evaluations. Simple randomisation is un-
estimated effects of intervention may still be likely to produce groups that are balanced with
300 The Advanced Handbook of Methods in Evidence Based Healthcare
respect to baseline prognostic variables and needs to be sufficiently large to offset the reduc-
restricted randomisation may then be used to tion in degrees of freedom.25 The power of a
produce groups which are more balanced.21 study will inevitably be reduced by matching on
a variable that is unrelated to the outcome if a
matched analysis is used. If the number of
Restricted Randomisation clusters is small, then an unmatched randomised
design will usually be more efficient because of
The aim of restricted randomisation is to pro- the loss of degrees of freedom associated with
duce intervention and control groups that are the matched analysis.25 Diehr et al.16 suggested
more evenly balanced with respect to prognostic that in studies with small numbers of clusters, if
variables than if simple randomisation were pre-intervention matching seems to be required
used. Clusters are first divided into strata ac- at the outset, an unmatched analysis will usually
cording to prognostic characteristics and then be more powerful than a matched analysis.
randomised within strata. This increases the However, strong correlates of the outcome may
statistical power, since estimates of the outcome make matching worthwhile for very small
are more precise.1,3,11,22–24 Clusters should only studies, as was described in a study of preven-
be stratified on variables that are known to be tion of HIV transmission in Tanzania. In this
highly correlated with the outcome of inter- study, intervention communities were pair mat-
est,16,21–23,25 for example, baseline levels of the ched with control communities according to
outcome variable.26 baseline indicators of risk of transmission of
Restricted allocation may also reduce bias in HIV and other sexually transmitted diseases.17
quasi-experimental studies.11,23 Murray recom- Finally, Klar and Donner23 cautioned that the
mended that non-random assignment of a small matched pairs design may be less suitable for
number of poorly matched or unmatched units organisation-based evaluations because of the
to intervention groups is inadvisable and should difficulty of separating the effect of intervention
be avoided.27 Thus, some form of restricted from between-cluster variation within pairs, and
allocation is mandatory for non-randomised hence in estimating the value of the intracluster
intervention studies. correlation. Thompson et al.29 recently proposed
that random effects meta-analysis can be applied
to the analysis of paired cluster randomised
Paired Designs designs. One variance component can be used to
allow for both natural variation between clusters
The paired design is a special case of the strat- and variation in the intervention effect from
ified design in which there are only two clusters stratum to stratum. Natural variation between
per stratum and one cluster is randomised to clusters is not estimated explicitly, but is still
each group. In matched pairs designs, clusters allowed for in the analysis. This approach will
are paired with respect to baseline character- be restricted to studies with the relatively large
istics that are associated with the outcome and number of cluster pairs required to estimate vari-
randomly allocated within pairs to intervention ance components with reasonable precision.30
and control groups. The matched pairs design
has important limitations which should be rec-
ognised. Firstly, it may be difficult to find Stratified Designs
matching variables that can be used to create
distinct pairs, particularly for a large number of Under the stratified design, two or more clusters
clusters.23,25 A second potential problem is that from each of several strata are allocated to each
in the event that a particular cluster drops out of intervention group.31 As for the matched pairs
the study, the other cluster (and thus the entire design, the clusters are grouped into strata
stratum) must be eliminated, causing a reduction according to one or more variables which are
in study power. Such a problem was described related to the outcome, but the stratified design
in a recent community randomised trial of pre- differs qualitatively from the matched pairs
vention of human immune deficiency virus design because there is replication of clusters
(HIV) transmission.28 Thirdly, because a paired within each intervention–stratum combination.
analysis considers the differences between clus- It is therefore possible to obtain an estimate
ters within pairs, the power of the study will of between-cluster variation from a stratified
depend on the number of pairs rather than the organisation-based trial as the cluster effect can
number of clusters. This loss of degrees of be separated from both the intervention effect
freedom means that a matched analysis may be and the stratum effect.
less powerful than an unmatched analysis when Using the completely randomised design as
the number of pairs is small. For pairing to be the benchmark in their simulations, Klar and
successful the decrease in experimental error Donner23 found that compared to the matched
Organisation- or Area-Based Evaluations 301
pairs design there is little loss in power for in relation to the repeated cross-sectional
stratified designs when strata are not related to one.32,36–38 In deciding between the two designs,
outcome. Another advantage of the stratified the relative costs should also be considered.39
design over the matched design is that for studies
in which there are a large number of clusters it is
easier to construct meaningful strata.1,23 In the SAMPLE SIZE REQUIREMENTS
light of these advantages, Klar and Donner23
suggest that the stratified design is under-utilised
in comparison to the matched pairs design and When estimating the sample size it is important
should be considered more often. The stratified to take into account the planned method of
design requires at least some strata to have two analysis. If the intention is to analyse cluster-
or more clusters assigned to each intervention. level outcomes, such as the proportion of gen-
eral practices running a diabetic clinic, then a
conventional sample size calculation may be
Design Issues for Follow-Up performed to estimate the required number of
clusters, using an estimate of the variance of the
Two designs can be used for sampling indi- cluster-level outcome. If individual level analy-
viduals from within clusters in organisation- ses are to be performed, it will be necessary to
based studies with follow-up. Under the repeated allow for both within- and between-cluster vari-
cross-sectional design a fresh sample of individ- ation by multiplying the sample size estimate
uals is drawn from each of the clusters at each obtained using standard formulae by an appro-
measurement occasion. Under the cohort design, priate variance inflation factor or design effect.
data are collected from the same sample of
individuals at each time point. When the ratio of Within-Cluster and Between-Cluster
sample size to population size is high, a cohort Variation
will be generated within the framework of a
repeated cross-sectional design as individuals In studies at the individual level, variation
may be sampled repeatedly by chance.21 The is between individuals. In organisation-based
overlap of repeated cross-sectional samples may studies, because of the correlation of individual
be engineered deliberately, and some com- level responses within clusters, there are two
munity intervention studies have set out to separate components of variation, within-cluster
achieve this mixed design.21,32 If the main objec- variation and between-cluster variation. Stand-
tive is to determine how an intervention changes ard statistical methods for sample size estima-
the behaviour of individuals, then a cohort tion and analysis do not recognise these two
design should be used. Cohort studies provide components of variation in the outcome and
the opportunity to link behavioural changes cannot be applied directly to organisation-based
directly to individual level prognostic factors. studies with evaluation at the individual level.
When the main objective is to determine how Standard sample size calculations, carried out
an intervention affects a cluster-level index of using estimates of between-individual variation,
health such as risk-factor prevalence, then a will lead to sample size estimates that are too
repeated cross-sectional design should be used small. Appropriate methods for sample size cal-
as it will generate data which are representative culation and analysis need to include estimates
of the study clusters throughout the study of within- and between-cluster variation.
period. Cohorts are subject to ageing, migration,
death, loss to follow-up and other factors which Estimating within- and between-cluster
affect representativeness.21,24,33 In choosing the
most suitable design for follow-up, the aim of components of variance
the intervention needs to be clearly defined so The overall variation in subject responses may
that the appropriate choice is made. be decomposed into its two constituent com-
When cohort or repeated cross-sectional ponents, within-cluster and between-cluster
designs are considered equally appropriate and variation, using analysis of variance. Variance
there are no biases, cohort designs will generally components may be estimated using a one-
offer greater statistical power because measure- way analysis of variance of the outcome with
ments made on the same subject over time are the clustering variable included as a random
often correlated and estimates of the inter- effect. The algebraic form of the appropriate
vention effect have a smaller variance than model is
those calculated from repeated cross-sectional
data.21,24,33–35 The greater the correlation be- yij 5 a 1 bj 1 eij (1)
tween measurements made on the same individ- where yij is the response of the ith individ-
ual the more powerful the cohort design will be ual within the jth cluster, a is the mean of the
302 The Advanced Handbook of Methods in Evidence Based Healthcare
responses, bj is the random effect of the jth respect to the outcome, then the intra-class cor-
cluster, and eij is the random error component relation is 1. In the context of organisation-
for the ith individual in the jth cluster. The bj are based evaluations, the intra-class correlation
independently and Normally distributed with coefficient will usually assume small positive
zero mean and constant variance, s2b. The eij are values; negative values can usually be attributed
independently and Normally distributed with to sampling error.
zero mean and constant variance s2w. Estimates The analysis of variance model can also be
of the between- and within-cluster variance used to calculate the intra-class correlation co-
components are calculated from the analysis of efficient when the outcome is dichotomous. In
variance table this case, the resulting estimate is a version of
ŝ2b 5 (MSB 2 MSW) / n0 (2) the kappa coefficient, which itself is an intra-
class correlation coefficient.43,44
ŝ2w 5 MSW (3)
where MSB is the between-cluster mean square
and MSW is the within-cluster mean square. n0 is The Design Effect
a measure of average cluster size which is
appropriate when clusters are variable in size:40 In order to take account of between cluster
F On G
variation, standard sample size formulae applied
2
1 j at the individual level need to be increased by
n0 5 N2 (4) the design effect. The design effect (Deff ) is
J21 N
defined as the ratio of the variance of the esti-
where J is the number of clusters, N is the total mated outcome under the cluster sampling strat-
number of individuals and nj is the number of egy (s2c ) to the variance that would be expected
individuals in the jth cluster. for a study of the same size using simple ran-
The analysis of variance approach can be dom sampling (s2srs) and is given by:45,46
extended to more complex designs which in-
clude other categorical and continuous variables s2c
that are important in the design of the study Deff 5 . (7)
from which variance components have to be s2srs
estimated. Thus, there may be two intervention
The design effect has also been referred to as
groups, several strata, and several time points at
the variance inflation factor.3,9 In evaluations
which measurements are made. Further details
are provided in standard texts.41 of organisation-based interventions the design
effect will usually be greater than unity because
of between-cluster variation. An interpretation
The intra-class correlation coefficient (r) of design effect is the number of times more
Within-cluster and between-cluster components subjects an organisation-based evaluation should
of variance may be combined in a single statis- have compared with one in which individuals
tical measure of between cluster variation, the are randomised, in order to attain the same
intraclass correlation coefficient. The intraclass power.
correlation coefficient (r) is defined as the pro- It can be shown that the design effect is given
portion of the true total variation in the outcome approximately by
that can be attributed to differences between the
clusters. Algebraically r is given by: Deff 5 1 1 (n 2 1)r (8)
then be used to calculate the design effect of total number of individuals required by the aver-
proposed studies at the planning stage. age cluster size. When it is feasible to sample
individuals within clusters, the power of the
study may be increased either by increasing the
Total Number of Individuals number of clusters or the number of individuals
sampled within clusters. Examples of appro-
When the aim of the study is to estimate the priate calculations were given by Thompson et
difference between intervention group means, al.29 and Kerry and Bland.53
the standard normal deviate sample size formula Increasing the number of clusters rather than
for comparing two independent groups should the number of individuals within clusters has
be multiplied by the design effect.48 The several advantages.29 The results of the study
required number of individuals per group, N, is will usually appear to be more generalisable if
given by the intervention has been implemented in a
number of different clusters. A larger number of
2 M s2 clusters also allows more precise estimation of
N5 x [1 1 (n 2 1)r] (9)
d2 the intra-class correlation coefficient and more
flexible approaches to analysis.29 As a general
where n is the cluster size, r is the intra-class rule, individual-level or multi-level approaches
correlation coefficient, s2 is the variance of to analysis will require at least 20 clusters (at
the outcome measure under simple random least 10 per group). Furthermore, there is a limit
sampling, and the d is the true difference to the extent to which power may be increased
between group means that can be detected. The solely by increasing the number of individuals
value of M depends on the levels of power within clusters because increasing the cluster
(1 2 b) and significance (a) and whether the test size also increases the design effect.29 The
is to be one-sided or two-sided. Values for M are relative cost of increasing the number of clusters
given in many texts, for example Ref. 49. When in the study, rather than the number of individ-
unequal cluster sizes are anticipated, Donner uals within clusters, will be an important con-
et al.48 suggested the use of either the expected sideration when deciding on the final structure
average cluster size, or more conservatively the of the study.
expected maximum cluster size.
When the study will estimate the difference
between group proportions, an adapted version Obtaining Estimates of the Intra-Class
of the standard sample size for comparing two Correlation Coefficient
proportions may be used.8,48 The required num-
ber of individuals per group, N, is: In order to carry out sample size calculations it
is necessary to estimate the intra-class correla-
M [pT (1 2 pT) 1 pC (1 2 pC)] tion coefficient from previous studies which are
N5
(pT 2 pC)2 (10) as similar as possible in terms of design to the
x [1 1 (n 2 1)r] one being planned. That is, studies that used
similar clusters, contained similar types of indi-
where N is the number of individuals per group, viduals, and used the same outcome. We report
r is the intra-class correlation coefficient with a wide range of components of variance and
respect to the binary trait, pT is the expected intra-class correlations elsewhere.20,54 Some
event rate in the treatment group, pC is the examples are given in Table 2. Since the intra-
expected event rate in the control group and n is class correlation coefficient will be known
the cluster size. M has the values defined above. imprecisely, it is sensible to take this imprecision
An equivalent method is given by Cornfield.9 into account and carry out a sensitivity analysis
Sample size formulae appropriate for matched on the estimated sample size.55,56 If confidence
pairs and stratified designs are given by Hsieh,50 limits of the intra-class correlation coefficient
Thompson et al.,29 Shipley et al.51 and Donner.52 are available, these may be used to perform a
sensitivity analysis.
Where possible, covariates other than design
Number of Clusters variables should be allowed for when estimating
the intra-class correlation coefficient.54,57–60 By
In practice, when designing organisation-based controlling for covariates the size of r, the
evaluations the investigator often needs to esti- design effect and the required sample size can
mate both the required total number of individ- be reduced; cluster-level rather than individual-
uals and the required number of clusters. If the level covariates will usually be more important
cluster size is predetermined, the required num- in this regard.58,60 In order to translate this into a
ber of clusters can be estimated by dividing the power advantage, the same covariates must also
304 The Advanced Handbook of Methods in Evidence Based Healthcare
Health authority Coronary heart Public Health Men , 65 198,419 105 0.0000283 6.62
disease mortality Common Data years
Set 1995
Postcode sector Systolic blood Health Survey Adults . 15 17 711 0.0192 1.31
pressure for England years
199493
Household Eats fruit at least Health Survey Adults . 15 1.72 9053 0.328 1.24
once a day for England years
199493
Elderly care unit Very satisfied RCP Wound Patient with 96.27 25 0.042 5.00
with care Care Audit94 wound
General practice Has symptoms Premaratne Patients 299.1 42 0.0096 3.86
of asthma et al. 199795 registered
with GPs
be controlled for when evaluating the inter- included in the study then there would be suffi-
vention effect of the study. cient power to detect a difference in case-mix of
this magnitude. However, because the subjects
An example: sample size calculation were to be sampled from a number of different
clinics it would also be necessary to allow for
allowing for clustering between-cluster variation. We estimated that the
To provide a numerical example we consider the average number sampled per clinic (in other
planning of an audit of clinical care. Suppose an words the cluster size) would be about 50. We
audit was to be carried out in the offices of did not know the intra-class correlation coeffi-
single-handed general practitioners (GP) and at cient, so we carried out a sensitivity analysis
clinics held in primary care health centres (HC). using values of 0.01, 0.05 and 0.1. The results
We might propose to find out whether the pro- are shown in Table 3. After allowing for
portion of all attenders who were taking anti- between-cluster variation, the sample size
hypertensive treatment was the same in the two requirement was inflated to 608 if the ICC was
clinical settings. 0.01 or 2,407 if the ICC was 0.1. Thus, 12
Suppose it were assumed that about half clinics per group would suffice if the ICC was
(50%) of the health centre attenders and 40% of 0.01, but 48 clinics per group would be needed
GP attenders were taking antihypertensive medi- if the ICC was as high as 0.1. If the number
cation. How many subjects would be required to sampled per clinic could be increased to 100,
detect this difference? Using a standard sample then eight clinics would be required if the ICC
size calculation with a 5 0.05 and power 5 was 0.01, but 44 clinics would be needed if the
0.80, if 408 subjects from each setting were ICC was 0.1. In deciding on the final design of
the study it would be necessary to consider the usually give similar results to adjusted individ-
feasibility and costs of increasing the number of ual analysis but may be less efficient if the
clinics in the study rather than the number of precision of estimates can be increased by con-
individuals sampled per clinic. trolling for individual-level covariates. When
cluster sizes are unequal, weighted versions of
tests can be used.
METHODS OF ANALYSIS Regression analyses may be carried out using
the cluster-specific estimates as observations,
Choice of Analytical Approach but this approach is often limited by the small
number of clusters. Cluster-level confounders
When observations are made at the cluster-level, may be included directly in these analyses.
standard statistical methods may be used. For Individual-level confounders cannot be included
example, the effectiveness of smoking control directly, but cluster-specific estimates may be
policies might be evaluated using the proportion standardised for individual level characteristics
of smoke-free workplaces as an outcome. Stand- such as age and sex.
ard statistical methods would not be appropriate
for the analysis of individual-level outcomes,
such as whether individuals gave up smoking, Univariate Individual-Level Analyses
because the individual responses are likely to be Adjusted for the Design Effect
correlated within organisational clusters such as
workplaces. Univariate individual level analyses consider the
There are three appropriate approaches to outcome values for individual subjects as the
analysis of individual-level data: observations. In order to allow for clustering of
data, the design effect is incorporated into stand-
1 Cluster-level analysis with the cluster means ard statistical formulae for hypothesis testing and
or proportions as the unit of analysis. confidence interval estimation. The univariate
2 Individual level univariate analysis with approach may often be less appropriate for non-
standard errors adjusted for the design effect. randomised designs. The intra-class correlation
3 Regression analysis using methods for clus- coefficient may be estimated from the study data
tered data, allowing for both individual and and about 20 clusters are required to calculate r
cluster-level variation. with reasonable precision.61
The choice of analytical approach may be The approach can be illustrated with reference
restricted by the design of the study and by the to the estimation of a difference between two
number of clusters and individuals included in proportions. The standard error used to construct
the sample. Univariate methods of analysis will confidence intervals for the difference between
not usually be appropriate for non-randomised proportions P̂1 2 P̂2 can be adapted to give a
studies because it will be necessary to allow for method which is appropriately used when the
confounding by using multiple regression analy- groups include clusters of equal or varying
sis. Analysis of data at the individual level, size:
either by means of adjusted univariate tests
or regression methods for clustered data, will
require data from 20 to 25 clusters in order to
allow the estimation of between-cluster vari-
SÊ(P̂1 2 P̂2) 5 F
Deff1P̂1Q̂1
N1
1
N2
G
Deff2P̂2Q̂2 1/2
(11)
carry out standard individual-level regression intercept, b is the regression coefficient for
analyses with clustered data. As for the adjusted covariate x, xij is the covariate value for the ith
individual-level univariate tests, the variances of individual in the jth cluster, uj is the random
the regression coefficients need to be increased effect for the jth cluster and eij is the level one
by the design effect. In the past, standard regres- residual for the ith individual in the jth cluster.
sion methods were adapted to account for clus- This model differs from the standard linear
tering, either by including an indicator variable regression model by the incorporation of the
as a fixed effect for each cluster in individual- random effect, uj. This represents the amount by
level analyses, or by aggregating data to the which the intercept for the jth cluster differs
cluster-level, but both these approaches have from the overall mean value a. Through uj the
well documented shortcomings.63–67 An approach dependence between observations within the
that is being used increasingly to adapt standard same cluster is modelled explicitly. When there
regression methods for clustering is the esti- is no variation between clusters the random
mation of robust standard errors (or robust effects model is equivalent to standard regres-
variance estimates). This approach allows the sion analysis. Random effects models may be
requirement that the observations be independ- generalised to allow one or more regression
ent to be relaxed, and leads to corrected standard coefficients to vary randomly between clusters.
errors in the presence of clustering. Robust Random effects models may be estimated
standard errors may be derived using the linear, using the statistical packages SAS73 and Stata.68
logistic or other regression commands in the In addition, there are a number of specialist
statistical package Stata.68 packages available of which the most popular
in the UK is MLn/MLwiN.74 Random effects
Regression Methods for Clustered Data models can also be implemented in a Bayesian
framework.75
A number of approaches to the regression analy- The estimation of random effects models for
sis of clustered data have been developed over non-normally distributed outcomes (for example,
the past few years, and these provide a useful random effects logistic regression) presents
alternative to cluster-level analysis and use of considerable computational difficulties. Because
adjusted individual-level univariate tests. There full maximum-likelihood estimation procedures
are two common approaches to the regression are not available, approximations (e.g. penalised
analysis of clustered or multi-level data, random quasi-likelihood76) are used. These may give rise
effects modelling and marginal modelling using to biased estimates and practical difficulties, for
Generalised Estimating Equations. example, non-convergence. This is an area of
active development.77–79 For binary outcomes
Random effects models the method of Generalised Estimating Equations
presents fewer difficulties (see below).
The random effects model is an extension of the It is important to emphasise the substantial
generalised linear model which can be used to sample sizes that are required for random effects
allow for the clustering of data within areas or modelling. Duncan et al. suggest that a minimum
organisations.64,69–72 The term random effects
of 25 individuals in each of 25 clusters may
modelling refers to the fact that the mean
be required for analysis at a single level of
response is treated as varying randomly between
clustering.80 The requirement for a reason-
clusters. Several other terms, for example,
ably large number of clusters may often limit
multi-level modelling69 and hierarchical linear
modelling70 are commonly used and emphasise the application of random effects models to
the multi-level, hierarchical nature of the data intervention studies in healthcare, but random
being analysed. A full discussion of random effects modelling of observational data on health
effects models is provided by Goldstein.69 service performance will often be appropriate.
Random effects models allow for the corre- For example, Leyland and Boddy81 used this
lation of responses within organisations by approach to study whether mortality of patients
explicitly modelling between-cluster variation. with myocardial infarction varied in different
This approach also facilitates the inclusion in hospitals.
the analysis of cluster-level confounders in
addition to individual-level confounders. With Generalised Estimating Equations
normally distributed outcomes, the form of a (GEEs)
random effects model is:
The method of generalised estimating equations
yij 5 a 1 bxij 1 uj 1 eij (12) is an alternative approach to the regression analy-
where yij is the outcome response of the ith sis of clustered data.82–85 The method uses what is
individual in the jth cluster, a is the constant or sometimes referred to as a population averaged
Organisation- or Area-Based Evaluations 307
or marginal modelling approach. Unlike the ran- modelled as nested within hospitals or health
dom effects model, the GEEs method does not authorities. This type of analysis in which
model the within-cluster dependence explicitly. organisational units are treated as random
Instead, the GEEs method assumes a correlation effects, leads to cluster-specific estimates which
matrix which describes the nature of the associ- are shrunk towards the population mean with
ation within clusters and obtains estimates of confidence intervals that are more appropriate
standard errors which are corrected for cluster- when compared to a conventional fixed effects
ing. Robust variance estimates which produce analysis. Over the past few years there has been
corrected standard errors even if the within increasing interest in monitoring the perform-
cluster correlation structure has been specified ance of health services by ranking organisational
incorrectly are generally used. The statistical units into league tables. Goldstein and Speigel-
packages Stata68 and SAS73 can be used to halter12 showed that regression methods for
implement GEEs. clustered data find a direct application in the
In the context of organisation-based evalu- analysis of this kind of data. In particular, an
ations the GEEs approach suffers from the dis- appropriate analysis can be used to generate
advantages that it requires a large number of confidence intervals for the rankings of organ-
clusters,84 it can only be used to model data with isational units as well as for the cluster-specific
one level of clustering, and it is not possible to outcomes. The confidence intervals for ranks
allow associations to vary between clusters. In will often be very wide, and even when there are
spite of these disadvantages the GEEs approach large differences in outcomes between units,
is useful because correct standard errors and ranks will often be imprecise and unstable indica-
P values may be obtained without use of tors of performance.88 Thus, league table rankings
complex modelling of variance structures, as in should not be used without additional information
random effects modelling. When the outcome is concerning their degree of imprecision.
binary, GEEs may provide the most feasible
approach to the regression analysis of data from An example: analysis allowing for
organisation-based evaluations.
clustering
To illustrate how an analysis might be per-
Applications of Multi-Level Analysis to formed in order to allow for clustering, we
Health Data analysed data obtained from an audit of clinical
care (Table 4). The aim of this analysis was to
The practical application of multi-level analysis see whether there was a difference between the
to the evaluation of healthcare is still at an early proportion of attenders who were receiving anti-
stage of development. Some applications up to hypertensive treatment at health centre clinics
1996 were reviewed by Rice and Leyland,64 and (HC) compared with attenders at single-handed
up to 1998 by Duncan et al.80 For example, this general practitioners (GPs).
approach may be applicable if cluster sampling Initial inspection of the data showed that there
is used to recruit individuals from different were 26 clinics, including 18 health centres and
clinics.86 In a study of neonatal mortality in eight single-handed GPs. The number of patients
relation to activity level in neonatal units, GEEs sampled per clinic was quite variable. The pro-
were used to adjust for the correlation of infant portion of patients treated for hypertension
outcomes within hospitals.87 seemed to be higher at the health centre clinics,
Duncan et al.80 make the point that random but the health centre patients were older than
effects models will be more useful than GEEs those attending the GPs. The value for r esti-
when the evaluator wants to distinguish the mated from the 26 clinics was 0.259, but it
effects of the organisational or geographical might not be justified to assume a common
context from the characteristics of individuals value for r at both types of clinic, since the
sampled within the organisation or area. This estimated value for r in the HC clinics was
distinction is often important in observational 0.107 while for the GP clinics it was 0.013.
evaluations of existing health services. For However, the number of GP clinics was small
example, in comparing the performance of dif- and possibly not sufficient to obtain a stable
ferent institutions after allowing for differences estimate for r.
in case mix,81 or comparing healthcare processes These observations suggest that certain
and outcomes in different geographical areas.80 approaches to analysis will be more appropriate
Goldstein and Speigelhalter12 discussed some than others. Because age is associated with both
of the statistical issues in comparisons of institu- blood pressure and clinic type it will be neces-
tional performance. They showed that it is sary to carry out analyses which allow for differ-
important to carry out the correct hierarchical ences in age among clinic types. Donner and
analysis in which individual responses are Klar61 pointed out that adjusted individual level
308 The Advanced Handbook of Methods in Evidence Based Healthcare
Table 4 Data from clinical audit. Figures are incorrect because, as we noted above, between
frequencies (% of row total) clinic variation is present. In order to construct
CIs, the design effect should be incorporated
Number Mean into the formula for the standard error for the
Clinic of On BP age difference in two proportions as shown in
type patients treatment (years) Women Equation (11). The design effect for the health
centre clinics is 6.66, and for the single-handed
HC 62 47 (76) 63 46 (74) general practitioners is 1.90. The standard error
HC 51 44 (86) 62 34 (67) of the difference in proportions was 0.0458. The
HC 43 41 (95) 65 33 (77) difference was still 0.453, but the corrected CIs
HC 43 32 (74) 64 27 (63) were from 0.363 to 0.543. This result showed
HC 35 28 (80) 63 26 (74) that the difference between the two audits was
HC 21 9 (43) 69 7 (33) estimated less precisely after allowing for
HC 83 69 (83) 66 61 (73) between-clinic variation. The corresponding
HC 68 24 (35) 56 43 (63) adjusted chi-squared test61 yielded a x2 of 113.0
HC 59 28 (47) 60 34 (58) (df 1). Note however that this test assumes a
HC 74 39 (53) 59 45 (61) common value for r.
HC 67 53 (79) 64 52 (78) An alternative approach to the analysis is to
HC 44 27 (61) 59 28 (64) use logistic regression. A conventional logistic
HC 63 44 (70) 62 40 (63) regression analysis showed that the relative odds
HC 76 39 (51) 63 49 (64) of patients being on antihypertensive medication
HC 36 15 (42) 55 27 (75) in the health centre clinics compared with the
HC 66 41 (62) 60 42 (64) GP clinics were 7.56 (5.91 to 9.66). The z
HC 37 24 (65) 64 27 (73) statistic obtained from this model was 16.1. This
HC 47 32 (68) 57 42 (89) analysis was modified to allow for clustering
GP 45 5 (11) 52 28 (62) within clinics by estimating robust standard
GP 43 6 (14) 52 29 (67) errors giving the same odds ratio 7.56, but the
GP 36 9 (25) 54 31 (86) 95% CIs were from 4.88 to 11.69 with a z
GP 89 14 (16) 46 69 (78) statistic of 9.1. Analyses were adjusted for age
GP 120 19 (16) 46 81 (68) and sex and the relative odds were reduced to
GP 21 8 (38) 53 13 (62) 5.46 (CIs 3.66 to 8.15). The method of GEEs
GP 120 31 (26) 54 87 (73) offered an alternative approach to regression
GP 84 19 (23) 47 49 (58) analysis. This was carried out by specifying a
HC subtotal 975 636 (65) 61 663 (68) logistic link function, an exchangeable correla-
GP subtotal 558 111 (20) 49 387 (69) tion structure, and robust standard errors to
allow for clustering within clinics. The analysis
Total 1533 747 (49) 57 1050 (68) adjusted for age and sex gave an odds ratio of
BP, blood pressure; HC, health centre; GP, general 5.06 (CIs 3.32 to 7.70) which was similar to the
practitioner. result obtained using logistic regression with
robust standard errors, but this analysis adjusted
the estimate for clustering as well as the stand-
hypothesis tests which assume a common value ard error. We used the Stata software to per-
for the design effect across groups will be less form both the logistic regression and GEEs
applicable to non-randomised data. analyses.68
The results of several different methods of A third approach to the analysis of these data
analysis are shown in Table 5. Initially, cluster- is to use random effects logistic regression to
level analyses were performed using the cluster- model the variation in response at clinic level.
specific proportions as observations. A non- We initially used a simplified model which
parametric rank sum test shows a significant assumed the same variation at level 2 (that is,
difference between the two clinical settings with clinic level) for both settings. The adjusted analy-
an approximate z statistic of 3.95. The two- sis gave a higher odds ratio (6.16) and slightly
sample t-test gave a mean difference in the wider CIs (3.39 to 11.18), when compared to the
cluster-specific proportions of 0.439 (95% con- logistic regression and GEE analyses. The level
fidence interval (CI) 0.307 to 0.571). Analyses 2 variance on the log odds scale was 0.387 with
were also performed at the individual level. An a standard error of 0.139. We used MLn89 to fit
analysis which made no allowance for the clus- the random effects logistic regression models
tering of responses gave a difference between with second-order PQL estimation.
the two proportions of 0.453 with 95% CI from The incorrect individual level analyses with
0.409 to 0.498 and a z statistic of 17.1. How- no adjustment for clustering gave narrower con-
ever, a standard statistical analysis would be fidence intervals and larger z statistics than do
Organisation- or Area-Based Evaluations 309
Difference
Cluster-level analysis (95% CI for difference)
Rank sum test – 3.95
Two-sample t-test 0.439 (0.307 to 0.571) t 5 6.85
Individual analysis – difference of proportions
Standard univariate test 0.453 (0.409 to 0.498) 17.1
Univariate test adjusted for design effect 0.453 (0.363 to 0.543) x2 5 113.0 (df 1)*
Odds ratio
Univariate individual analysis – logistic regression methods (95% CI for odds ratio)
Standard logistic regression – univariate 7.56 (5.91 to 9.66) 16.1
– adjusted for age and sex 5.46 (4.22 to 7.08) 12.9
Logistic regression, robust standard errors – univariate 7.56 (4.88 to 11.69) 9.1
– adjusted for age and sex 5.46 (3.66 to 8.15) 8.3
Generalised estimating equations – univariate 7.32 (4.62 to 11.57) 8.5
– adjusted for age and sex 5.06 (3.32 to 7.70) 7.6
Random effects logistic regression – univariate 8.47 (4.51 to 15.91) 6.6
– adjusted for age and sex 6.16 (3.39 to 11.18) 6.0
any of the methods which allow for clustering studies will be enhanced by adopting approaches
of responses. Carrying out a two-sample t-test to study design and analysis that explicitly
using the cluster-specific proportions as obser- acknowledge the hierarchical structure of data
vations provided the most accessible method of from evaluations of health services.2,3,12,20,64,80,90
analysis and gave a valid result. However, as it
was also necessary to adjust for individual level
confounders, one of the regression methods for
clustered data was to be preferred. Ordinary
ACKNOWLEDGEMENTS
logistic regression with robust standard errors
gave a similar result to that obtained using This work was funded by a contract from the
generalised estimating equations in these data. UK National Health Service Health Technology
Logistic regression was more accessible, and Assessment Programme. The views expressed
was computationally less intensive with shorter are those of the authors and do not necessarily
analysis times, but the GEE approach adjusted reflect those of the NHS Research and Develop-
both the estimate and the standard error for ment Programme. The authors thank Dr D
clustering and was to be preferred for this Mahabir for permission to present the audit
reason. Random effects logistic regression might data.
have allowed additional modelling of the vari-
ance at cluster level which could have been
useful if the analysis was to focus on the per-
formance of particular primary care clinics.
REFERENCES
11. Koepsell TD, Diehr PH, Cheadle A, Kristal A. intervention trials with community as the unit
Invited commentary: symposium on community of randomization. Am. J. Epidemiol. 1995; 142:
intervention trials. Am. J. Epidemiol. 1995; 142: 587–93.
594–9. 25. Martin DC, Diehr P, Perrin EB, Koepsell TD. The
12. Goldstein H, Spiegelhalter DJ. League Tables and effect of matching on the power of randomised
Their Limitations: Statistical Issues in Compar- community intervention studies. Stat. Med. 1993;
isons of Institutional Performance. J. Roy. Statist. 12: 329–38.
Soc., A 1996; 159: 385–443. 26. Kirkwood BR, Morrow RH. Community-based
13. Puska P, Salonen JT, Tuomilehto J, Nissinen A, intervention trials. J. Biosocial Sci. 1989; 10:
Kottke TE. Evaluating community-based pre- 79–86.
ventive cardiovascular programs: problems and 27. Murray DM. Design and analysis of community
experiences from the North Karelia project. J. trials: lessons from the Minnesota Heart Health
Community Health 1983; 9: 49–64. Program. Am. J. Epidemiol. 1995; 142: 569–75.
14. Nicholl J, Turner J. Effectiveness of a regional 28. Kelly JA, Murphy DA, Sikkema J, McAuliffe
trauma system in reducing mortality from major TL, Roffman RA, Solomon LJ, Winett RA,
trauma: before and after study. Br. Med. J. 1997; Kalichman SC, the Community HIV Prevention
315: 1349–54. Research Collaborative. Randomised, controlled,
15. Puska P, Holland WW, Detels R, Knox G (eds). community-level HIV prevention intervention for
Oxford Textbook of Public Health. 12, Interven- sexual-risk behaviour among homosexual men in
tion and experimental studies. Oxford: Oxford US cities. Lancet 1997; 350: 1500–5.
Medical Publications; 1991, pp. 177–87. 29. Thompson SG, Pyke SD, Hardy RJ. The design
16. Diehr P, Martin DC, Koepsell T, Cheadle A. and analysis of paired cluster randomised trials:
Breaking the matches in a paired t-test for com- an application of meta-analysis techniques. Stat.
munity interventions when the number of pairs is Med. 1997; 16: 2063–79.
small. Stat. Med. 1995; 14: 1491–504. 30. Donner A, Hauck W. Estimation of a common
17. Hayes R, Mosha F, Nicoll A, Grosskurth H, odds ratio in case-control studies of familial
Newell J, Todd J, Killewo, J, Rugemalila J, aggregation. Biometrics 1988; 44: 369–78.
Mabey D. A community trial of the impact of 31. Donner A, Donald A. Analysis of data arising
improved sexually transmitted disease treatment from a stratified design with the cluster as unit of
on the HIV epidemic in rural Tanzania: 1. Design. randomization. Stat. Med. 1987; 6: 43–52.
AIDS 1995; 9: 919–26. 32. Diehr P, Martin DC, Koepsell T, Cheadle A,
18. Black N. Why we need observational studies to Psaty BM, Wagner EH. Optimal survey design
evaluate the effectiveness of health care. Br. Med. for community intervention evaluations: cohort or
J. 1996; 312: 1215–18. cross-sectional? J. Clin. Epidemiol. 1995; 48:
19. Cook TD, Campbell DT. Quasi-experimentation. 1461–72.
Design and analysis issues for field settings. 33. Gail MH, Mark SD, Carroll RJ, Green SB, Pee D.
Chicago: Rand McNally College Publishing On design considerations and randomization-
Company, 1979. based inference for community intervention trials.
20. Ukoumunne OC, Gulliford MC, Chinn S, Sterne Stat. Med. 1996; 15: 1069–92.
J, Burney PGJ. Methods for evaluating area-wide 34. Gail MH, Byar DP, Pechacek TF, Corle DK.
and organisation-based interventions in health Aspects of statistical design for the Commun-
and health care. Health Technology Assessment ity Intervention Trial for Smoking Cessation
1999; 3(5). (COMMIT). Controlled Clinical Trials 1992; 13:
21. Koepsell TD, Wagner EH, Cheadle AC, Patrick 6–21.
DL, Martin DC, Diehr PH, Perrin EB, Kristal AR, 35. Koepsell TD, Martin DC, Diehr PH, Psaty
Allan-Andrilla CH, Dey LJ. Selected methodo- BM, Wagner EH, Perrin EB, Cheadle A. Data
logical issues in evaluating community-based analysis and sample size issues in evaluations of
health promotion and disease prevention programs. community-based health promotion and disease
Annu. Rev. Public Health 1992; 13: 31–57. prevention programs: a mixed-model analysis of
22. Freedman LS, Green SB, Byar DP. Assessing variance approach. J. Clin. Epidemiol. 1991; 44:
the gain in efficiency due to matching in a com- 701–13.
munity intervention study. Stat. Med. 1990; 9: 36. Murray DM, Hannan PJ. Planning for the appro-
943–52. priate analysis in school-based drug-use preven-
23. Klar N, Donner A. The Merits of Matching: A tion studies. J. Consult. Clin. Psychol. 1990; 58:
Cautionary Tale. Stat. Med. 1997; 16: 1573–64. 458–68.
24. Green SB, Corle DK, Gail MH, Mark SD, Pee D, 37. Donner A, Klar N. Statistical considerations in
Freedman LS, Graubard, BI, Lynn WR. Interplay the design and analysis of community interven-
between design and analysis for behavioural tion trials. J. Clin. Epidemiol. 1996; 49: 435–9.
312 The Advanced Handbook of Methods in Evidence Based Healthcare
38. Feldman HA, McKinlay SM. Cohort versus and analysis of community trials: intraclass corre-
cross-sectional design in large field trials: pre- lations from the Minnesota Heart Health Pro-
cision, sample size, and a unifying model. Stat. gram. Epidemiology 1994; 5: 88–95.
Med. 1994; 13: 61–78. 57. Mickey RM, Goodwin GD, Costanza MC. Esti-
39. McKinlay SM. Cost-efficient designs of cluster mation of the design effect in community inter-
unit trials. Prevent. Med. 1994; 23: 606–11. vention studies. Stat. Med. 1991; 10: 53–64.
40. Armitage P, Berry G. Statistical Methods in 58. Murray DM, Short B. Intraclass correlation
Medical Research. 2nd edn. Oxford: Blackwell among measures related to alcohol use by young
Science, 1987. adults: estimates, correlates and applications in
41. Snedecor GW, Cochran WG. Statistical methods. intervention studies. J. Studies Alcohol. 1995; 56:
6th edn. Ames, Iowa: Iowa State University Press, 681–94.
1967. 59. Siddiqui O, Hedeker D, Flay BR, Hu FB. Intra-
class correlation estimates in a school-based
42. Donner A. The analysis of intraclass correlation
smoking prevention study – outcome and mediat-
in multiple samples. Ann. Hum. Genet. 1985; 49:
ing variables, by sex and ethnicity. Am. J. Epi-
75–82.
demiol. 1996; 144: 425–33.
43. Fleiss JL, Cohen J. The equivalence of weighted
60. Raudenbush SW. Statistical analysis and optimal
kappa and the intraclass correlation coefficient design for cluster randomized trials. Psychol.
as measures of reliability. Education Psychol. Methods 1997; 3: 173–85.
Measure 1973; 33: 613–19. 61. Donner A, Klar N. Methods for comparing event
44. Fleiss JL. Statistical methods for rates and rates in intervention studies when the unit of
proportions. 2nd edn. Chichester: Wiley; 1981, allocation is a cluster. Am. J. Epidemiol. 1994;
pp. 211–36. 140: 279–89.
45. Moser CA, Kalton G. Survey Methods in Social 62. Donner A, Klar N. Confidence interval construc-
Investigation. Aldershot: Dartmouth Publishing tion for effect measures arising from cluster ran-
Company Limited, 1993, pp. 61–78. domization trials. J. Clin. Epidemiol. 1993; 46:
46. Kish L. Survey Sampling. London: John Wiley & 123–31.
Sons, 1965, pp. 148–81. 63. Aitkin M, Longford N. Statistical modelling
47. Donner A. An empirical study of cluster random- issues in school effectiveness studies. J. Roy.
ization. Int. J. Epidemiol. 1982; 11: 283–6. Statist. Soc. A 1986; 149: 1–43.
48. Donner A, Birkett N, Buck C. Randomization by 64. Rice N, Leyland A. Multilevel models: applica-
cluster. Sample size requirements and analysis. tions to health data. J. Health Services Res. Policy
Am. J. Epidemiol. 1981; 114: 906–14. 1996; 1: 154–64.
49. Pocock SJ. Clinical trials. A practical approach. 65. Zucker DM. An analysis of variance pitfall: The
Chichester: John Wiley & Sons, 1983. fixed effects analysis in a nested design. Educa-
50. Hsieh FY. Sample size formulae for intervention tion Psychol. Measurement 1990; 50: 731–8.
studies with the cluster as unit of randomization. 66. Singer JD. An Intraclass Correlation Model for
Stat. Med. 1988; 7: 1195–201. Analysing Multilevel Data. J. Exp. Education
51. Shipley MJ, Smith PG, Dramaix M. Calculation 1987; 55: 219–28.
of power for matched pair studies when random- 67. Brown RL, Baumann LJ, Cameron L. Single-
ization is by group. Int. J. Epidemiol. 1989; 18: level analysis of intervention studies with hier-
457–61. archically structured data: a cautionary note.
Nursing Res. 1996; 45: 359–62.
52. Donner A. Sample size requirements for stratified
68. Stata Corporation. Stata statistical software.
cluster randomization designs. Stat. Med. 1992;
release 5. College Station, Texas. Stata Press,
11: 743–50.
1997.
53. Kerry SM, Bland JM. Sample size in cluster
69. Goldstein H. Multilevel Statistical Models.
randomisation. Br. Med. J. 1998; 316: 549. London: Arnold, 1996.
54. Gulliford MC, Ukoumunne OC, Chinn S. Com- 70. Bryk AS, Raudenbush SW. Hierarchical Linear
ponents of variance and intraclass correlations Models. London: Sage, 1992.
for the design of community-based surveys and 71. Longford NT. Random Coefficient Models.
intervention studies. Data from the Health Survey Oxford: Oxford University Press, 1995.
for England 1994. Am. J. Epidemiol. 1999; 149: 72. Arnold CL. An introduction to hierarchical linear
876–83. models. Measurement and Evaluation Counsel
55. Mickey RM, Goodwin GD. The magnitude and Develop. 1992; 25: 58–90.
variability of design effects for community inter- 73. SAS Institute Inc. SAS/STAT User Guide. Version
vention studies. Am. J. Epidemiol. 1993; 137: 6. Volume 2. 4th edn. Cary, North Carolina: SAS
9–18. Institute, 1990, pp. 1127–34.
56. Hannan PJ, Murray DM, Jacobs DR, Jr., 74. Kreft IGG, de Leeuw J, van der Leeden R.
McGovern PG. Parameters to aid in the design Review of five multilevel analysis programs:
Organisation- or Area-Based Evaluations 313
BMDP-5V, GENMOD, HLM, ML3, VARCL. to 1991. Aust. N.Z. J. Public Health 1997; 21:
Am. Statist. 1994; 48: 324–35. 257–64.
75. Speigelhalter DJ. Bayesian graphical modelling: a 87. Phibbs CS, Bronstein JM, Buxton E, Phibbs RH.
case study in monitoring health outcomes. Appl. The effects of patient volume and level of care at
Statist. 1998; 47: 115–33. the hospital of birth on neonatal mortality. JAMA
76. Breslow E, Clayton DG. Approximate inference 1996; 276: 1054–9.
in generalized linear mixed models. J. Am. Statist. 88. Marshall EC, Speigelhalter DJ. Reliability of
Assoc. 1993; 88: 9–25. league tables of in vitro fertilisation clinics: retro-
77. Guo SW, Lin DY. Regression analysis of multi- spective analysis of live birth rates. Br. Med. J.
variate grouped survival data. Biometrics 1994; 1998; 316: 1701–5.
50: 632–9. 89. Paterson L. Woodhouse G. Multilevel modelling
78. Hedeker D, Gibbons RD. A random-effects applications. A guide for users of MLn. London:
ordinal regression model for multilevel analysis. Institute of Education, 1996.
Biometrics 1994; 50: 933–44. 90. Ukoumunne OC, Gulliford MC, Chinn S, Sterne
79. Ten Have TR, Landis JR, Hartzel J. Population- J, Burney P, Donner A. Evaluation of health
averaged and cluster-specific models for clustered interventions at area and organisation level. In:
ordinal response data. Stat. Med. 1996; 15:
Black N, Brazier J, Fitzpatrick R, Reeves B (eds).
2573–88.
Methods for Health Care Services Research.
80. Duncan C, Jones K, Moon G. Context, composi-
London: BMJ Books, 1998.
tion and heterogeneity: using multi-level models
91. Taylor FC, Ramsay ME, Renton A, Cohen H.
in health research. Soc. Sci. Med. 1998; 46:
Methods for managing the increased workload
97–117.
in anticoagulant clinics. Br. Med. J. 1996; 312:
81. Leyland AH, Boddy FA. League tables and acute
myocardial infarction. Lancet 1998; 351: 555–8. 286.
82. Liang K, Zeger SL. Longitudinal data analysis 92. Pan XR, Li GW, Hu YH et al. Effect of diet and
using generalized linear models. Biometrika 1986; exercise in preventing NIDDM in people with
73: 13–22. impaired glucose tolerance. The Da Qing IGT
83. Zeger SL, Liang KY. Longitudinal data analysis and Diabetes Study. Diabetes Care 1997; 20:
for discrete and continuous outcomes. Biometrics 537–44.
1986; 42: 121–30. 93. Colhoun H, Prescott-Clarke P. Health survey for
84. Zeger SL, Liang KY, Albert PS. Models for England 1994. London: HMSO, 1996.
longitudinal data: a generalized estimating equa- 94. Royal College of Physicians Research Unit. Older
tion approach. Biometrics 1988; 44: 1049–60. people’s programme. National chronic wound
85. Diggle PJ, Liang KY, Zeger SL. Analysis of audit. London: Royal College of Physicians,
Longitudinal Data. Oxford: Oxford University 1997.
Press, 1996. 95. Premaratne UN, Sterne JAC, Webb J, et al.
86. Reid SE, Simpson JM, Britt HC. Pap smears in Clustered randomised trial of an intervention to
general practice: a secondary analysis of the improve the management of asthma: Greenwich
Australian Morbidity and Treatment Survey 1990 asthma study. Br. Med. J. 1999; 318: 1251–5.
18
Handling Uncertainty in Economic
Evaluation
A N D R E W H . B R I G G S and A L A S T A I R M . G R A Y
single methodological scenario for each article new therapy compared to the standard therapy.
with 333 associated baseline results. These Where one therapy is simultaneously cheaper
results were converted to a common cost base and more effective than the other (quadrants II
year and rank ordered to give a comprehensive and IV on the CE plane), it is clearly the
‘league table’ of UK results. Of the 333 results, treatment of choice and is said to dominate.
61 had an associated full range of values to However, where one intervention is both more
represent uncertainty. Alternative rankings effective and more costly (quadrants I and III),
based on the high or low values from this range then the decision is no longer clear. Rather, a
showed that there could be considerable disrup- judgement must be made concerning whether
tion to the rank order based on the baseline point the difference in costs resulting from a switch in
estimates only. therapy is justified by the difference in effective-
The review of patient-level cost data showed ness that such a switch would bring about. If it
that 53 of the 492 studies on the database had is possible to define some maximum acceptable
patient-level cost data, and that just 15 of these value for the incremental cost–effectiveness
had reported some measure of cost variance.
ratio (ICER), then this ‘ceiling’ value of the
Only four studies had calculated 95% confi-
dence intervals for cost. A review of five avail- ratio can be used to judge whether the treatment
able cost data sets showed that cost data were in question is cost-effective. The ceiling value of
not normally distributed, and in two cases the ICER can be represented by the (slope of
showed substantial skewness. the) dashed line on the CE plane of Figure 1. If
The review of methods for calculating con- the incremental costs and effects are plotted to
fidence intervals for cost–effectiveness ratios the right of this line on the CE plane, then the
results in the recommendation that the parametric treatment is considered cost-effective, while
Fieller’s method or non-parametric bootstrap points to the left of this line represent cost-
approach be used. Cost–effectiveness accept- ineffective interventions.
ability curves are identified as a method of Cost–effectiveness analysis is essentially a
presenting uncertainty in studies that has several comparative methodology and, in the absence of
advantages over the presentation of confidence a clearly defined ceiling ICER appropriate for
intervals. decision-making, results have meaning pri-
marily in terms of comparison to other reported
cost–effectiveness results. In this review, we
INTRODUCTION therefore concentrate on economic evaluations
that report cost per life year or cost per quality-
adjusted life year (QALY) results, since it is
Health economics is a burgeoning discipline, possible to compare these results across differ-
particularly in the area of the economic evalu- ent disease/treatment areas. It is common for
ation of healthcare interventions. During the point estimates of cost–effectiveness to be
1980s, much of the literature concerning econ- grouped together and presented in rank order in
omic analysis related to justifying the applica- so-called cost–effectiveness league tables. The
tion of economic thinking to the healthcare implication of constructing such league tables
sector. In the 1990s, with a much wider accept- is that those interventions with low cost per
ance of the need to evaluate economically new (QA)LY results should take precedence over
and existing healthcare interventions, the focus interventions with higher cost per (QA)LY
has become more methodological – with much
results.
more attention being placed on the appropriate
The use of point estimates in constructing
methods to employ when evaluating health-
care interventions. The most popular evaluative such tables raises problems. Consider Figure 2,
technique for economic evaluation is cost– where the point estimates for seven different
effectiveness analysis,1 and many cost–effective- healthcare interventions are plotted in increasing
ness studies present only point estimates of order of magnitude (this is adapted from Table
the value for money of particular healthcare II by Petrou and colleagues4). The ‘I’ bars on
interventions, although it is clear that there is this figure represent the estimated range of
often considerable uncertainty in the evaluative uncertainty around the point estimates. It is
process. immediately apparent that the range of un-
The cost–effectiveness (CE) plane2,3 is often certainty is so wide that it would be possible to
employed to show how decisions can be related rank these interventions in a completely differ-
to both costs and effects. The CE plane (Figure ent order than that obtained from the point
1) is divided into four quadrants indicating four estimates. Thus, if policy-makers are to be fully
possible situations in relation to the additional informed when making decisions based on the
costs and additional health outcome effects of a results of economic evaluations, it is imperative
316 The Advanced Handbook of Methods in Evidence Based Healthcare
that analysts estimate and present the level of The standard method for handling uncertainty
uncertainty inherent in their results. due to sampling variation is statistical analysis.
Uncertainty in economic evaluation is per- Where patient specific resource use and health
vasive, entering the evaluative process at every outcome data have been collected (for example,
stage. It is useful to distinguish uncertainty as part of a prospective clinical trial) in a
related to the data requirements of a study (the so-called stochastic analysis,6 then statistical
resource use and health outcome consequences techniques have been developed to calculate
of a particular intervention and the data required confidence intervals around point estimates
to value those consequences) from uncertainty of cost–effectiveness (although the methods
related to the process of evaluation. Uncertainty required to estimate confidence limits around a
in the data requirements of a study arises ratio statistic are less straightforward than for
through natural variation in populations, which many other statistics).
means that estimates based on samples drawn In practice, however, a relatively small pro-
portion of all economic evaluations are con-
from that population will always be associated
ducted alongside clinical trials. Instead, data are
with a level of uncertainty which is inversely
synthesised from a number of different sources
related to sample size. Examples of uncertainty – including reviews of the literature, hospital
due to the evaluative process include: the need records and even clinical judgement – in a
to extrapolate when conducting evaluations, for deterministic analysis. Hence, standard statis-
example from a clinical outcome measure (such tical methods cannot be employed. Moreover,
as cholesterol lowering) to a health outcome even where a stochastic analysis is possible, the
measure (reduced morbidity and mortality due remaining levels of uncertainty that are not
to heart disease); uncertainty related to general- related to sampling variation need exploration
ising from the context of the study to other and quantification. To do this, a technique
contexts and to different patient populations; known as sensitivity analysis is used, which
and the uncertainty related to choice of ana- involves systematically examining the influence
lytical methods when conducting economic of the variables and assumptions employed in an
evaluation (e.g. whether to include indirect costs evaluation for the estimated cost–effectiveness
in the analysis).5 results.
Handling Uncertainty in Economic Evaluation 317
There are three main forms of sensitivity literature. The review is conducted at a number
analysis.7 A one-way sensitivity analysis exam- of levels of detail. The overall review process is
ines the impact of each variable independently, described on pp. 318–22, including the search
while holding all other variable constant. By and identification methods used and the criteria
contrast, an extreme scenario analysis involves for inclusion in the final sample. Descriptive
setting each variable to simultaneously take the statistics for the overall review are then pre-
most optimistic (pessimistic) value from the sented, including the methods employed by ana-
point of view of the intervention under evalu- lysts to quantify uncertainty in their study. The
ation in order to generate a best (worst) case results of a detailed review of all studies report-
scenario. In real life the components of an ing results for UK healthcare interventions is
evaluation do not vary in isolation, nor are they reported on pp. 322–7. This includes the devel-
perfectly correlated; hence a one-way sensitivity opment of a relational cost–effectiveness data-
analysis might be expected to underestimate base with the ability to match study results by
uncertainty, while an extreme scenario analysis the underlying methods employed by analysts.
might be expected to overestimate uncertainty. A subset of studies for which patient level cost
A third technique, known as probabilistic information are available is examined in detail
sensitivity analysis, involves allowing the vari- on pp. 327–31 in order to see how healthcare cost
ables in the analysis to vary simultaneously data are reported. To aid the discussion, actual
across a plausible range according to predefined healthcare cost information from a number of
distributions. Probabilistic analysis will nor- available data sets is reported. The particular
mally produce a result lying between the ranges problem of calculating confidence intervals for
implied by one-way sensitivity analysis and cost–effectiveness ratios and reviews recently
extreme scenario analysis, and therefore may proposed methods for estimating confidence
produce a more realistic interval.8 limits for cost–effectiveness ratios are examined
In the following sections, we report the results on pp. 331–4. Finally, guidelines are proposed
of a major review of the economic evaluation based on the findings of this review.
318 The Advanced Handbook of Methods in Evidence Based Healthcare
Figure 5 Studies on the database by country to which the results are applicable.
study designs were based on modelling-type ap- may produce more studies reporting standard
proaches, it is not surprising that the majority of statistical confidence intervals around cost–
analyses used some form of sensitivity analysis, effectiveness ratios. Of the 24 studies on the
most commonly a simple one-way sensitivity database that incorporated some form of statis-
analysis. Some form of statistical analysis was tical analysis of uncertainty, only three attemp-
attempted in just 5% of studies, although this ted any kind of statistical analysis of the cost–
was rarely related to the cost–effectiveness ratio effectiveness ratio itself. This may be due in part
itself. Of some concern was that 17% of analysts to the complexities of calculating confidence
failed to provide any attempt to quantify the intervals for a ratio, a problem addressed in
inherent uncertainty in their results. recent research – an overview of which is given
on pp. 331–4.
It is worrying that 17% of studies do not
Discussion attempt to quantify in any way the uncertainty in
their analysis, especially given the proliferation
Very few studies identified in our review are of published reviews and guidelines for analysts,
based on primary data collection, the majority reviewers and editors in recent years which
relying instead on modelling techniques to syn- emphasise the importance of sound evaluative
thesise data from a number of different sources. methods, including the use of sensitivity analy-
This pattern may change over the coming years sis. To examine whether there is any evidence
with the increasing tendency to incorporate an that this situation has improved over time, Fig-
economic component into clinical trials, and this ure 8 shows the percentage of articles on the
Handling Uncertainty in Economic Evaluation 321
database which failed to employ any methods compared to a statistical 95% confidence inter-
for quantifying uncertainty, by year of publica- val.7 By contrast, extreme scenario analysis,
tion over the past 10 years. Although no clear which might be expected to overestimate such
trend is discernible over the whole period, there an interval and could therefore be seen as a more
would appear to be some improvement from conservative approach, was only used by 6% of
1992 onwards, giving grounds for optimism that analysts. The methods of probabilistic sensitiv-
methodological standards are improving. How- ity analysis may offer a compromise between
ever, the reliance on simple one-way sensitivity these two methods,8 but have so far been under-
analysis is less encouraging, as this method is utilised in the economic evaluation literature.
likely to underestimate the level of uncertainty A further concern is that the use of one-way
322 The Advanced Handbook of Methods in Evidence Based Healthcare
sensitivity analysis often seemed less than com- clear that many such baseline results can be
prehensive, with only select variables subjected presented for each scenario, relating to the dis-
to the analysis. We deal directly with this prob- ease in question, the type of intervention, and
lem in the next section, as we consider the the clinical characteristics and age/sex distribu-
detailed review of all studies reporting cost– tion of the patient population. For example,
effectiveness results for the UK. Dusheiko and Roberts15 evaluated interferon
treatment for both hepatitis B and C; Daly and
colleagues16 presented baseline results for three
A DETAILED REVIEW OF UK different combinations of hormone replacement
COST–EFFECTIVENESS STUDIES therapy for menopausal women; and Fenn and
colleagues17 presented baseline results for
Performing a reliable economic evaluation is not thrombolytic therapy by age and by sex sub-
a simple task. Firstly, the underlying methodo- groups of patients attending with symptoms of
logical framework is still being debated among acute myocardial infarction. For the third stage,
health economists. Secondly, there is often all ranges of values reported around each
uncertainty surrounding data and assumptions, baseline result to represent uncertainty were
and a lack of consensus over how to handle this. recorded. For example, Freemantle and col-
Finally, there is subjectivity in presenting and leagues18 reported that their baseline estimate of
interpreting results. What to one analyst is £51 717 per life-year saved by the use of selec-
clearly ‘a highly cost-effective use of resources’ tive serotonin re-uptake inhibitors for the pre-
may to another analyst seem poor value for vention of suicide could vary from £31 000 to
money. £155 000 per life-year saved when each para-
The recent development of agreed guidelines meter was varied separately across a plausible
for analysts, reviewers, editors and decision- range of values (a so-called one-way sensitivity
makers should help to standardise methodolo- analysis), and from £19 000 to £173 000 per
gies and improve techniques for handling un- life-year saved when these parameters were
certainties over data and assumptions. Here we varied simultaneously (a so-called multi-way
concentrate on the presentation and interpreta- analysis).
tion of results, and suggest an approach derived The final stage of the review was to record all
from a structured review of published UK cost– references to other cost–effectiveness results
effectiveness studies identified from the overall made by analysts. The number of results quoted
database reported in the previous section.a A was recorded, along with the value of each
total of 60 studiesb were identified that had result, the source, and the therapeutic area. As
presented cost per life-year or cost per QALY well as providing an opportunity to assess
results in a UK context – the bibliographic whether the external results chosen by analysts
references for these studies are listed in full in for comparison truly reflected the range of
Appendix 3 of the full report.12 values found in the main review, this process
provided a useful check on the identification
The Structured Review process of the overall review – references not
already on the database were obtained and
Each of the 60 studies was reviewed with reviewed.
respect to: (i) the methods employed in the
study; (ii) the baseline results reported; and
(iii) the range of values reported to represent Results from the Database
uncertainty. Additionally, any external cost–
effectiveness results quoted by authors to com- A summary of the entries on each level of the
pare with their own results was also recorded. database is given in Table 1. A total of 60
The first stage of the review process was studies which reported cost per (QA)LY results
to identify each ‘methodological scenario’ were identified and reviewed, and from these a
deployed by the analysts. For example, Field total of 106 methodological scenarios, 548
and colleagues,14 in their evaluation of health baseline results, 209 ranges of values and 248
promotion strategies in primary care presented externally quoted ICERs were recorded. The
results for two scenarios: firstly, where health mean, minimum and maximum values for each
outcomes generated by the programmes were level in relation to the next level in the database
discounted at the same rate as costs; and sec- hierarchy are also given. A range of values to
ondly, where those health outcomes were not represent uncertainty for at least one of the
discounted. In the second stage, the baseline baseline results was recorded for 66% of studies,
results given by the analysts were then reviewed and 61% of studies quoted an external ICER for
and attached to the appropriate scenario. It is comparison with their own results.
Handling Uncertainty in Economic Evaluation 323
325
326 The Advanced Handbook of Methods in Evidence Based Healthcare
Figure 10 Histogram of the distribution of (a) baseline ICERs from the database
analysis and (b) external ICERs quoted in studies for comparison.
from the rank ordering exercise. Histograms for to date been a common way of making a com-
these two distributions are presented in Figure 10, parison against other treatments routinely pro-
emphasising that the cost–effectiveness results vided in the health service, i.e. other treatments
chosen by analysts to compare with their own which by inference society/healthcare providers
results tend to be higher than those recorded in are already revealing a willingness to provide at
our structured review of overall results.f the calculated cost–effectiveness ratio. How-
ever, league tables – or other lists of external
comparators – are seldom comprehensive. In our
Discussion review, the 61% of studies which referred to
external comparators quoted on average only six
Cost–effectiveness analysis is a comparative UK cost–effectiveness results. Thus, assessment
methodology, in which results have meaning of the relative cost–effectiveness of an inter-
primarily in relation to opportunity cost as vention is generally made on the basis of very
measured by other cost–effectiveness results. It incomplete information. This is especially so
follows that the greater the number of results when analysts are trying to get some grip on
available for comparison, the more accurately mean or median values – what is routinely
can the relative cost–effectiveness of any par- provided and regarded as a currently accept-
ticular study be judged. The ‘league table’ has able norm across the health sector. This scope
Handling Uncertainty in Economic Evaluation 327
for subjective opinion and interpretation of non-RCT study. In total, therefore, 53 of the 492
cost–effectiveness results has been an import- studies in the database (11%) reported collect-
ant factor in the controversies over research ing patient-specific resource/cost data in a way
sponsorship.24 that allowed some measure of variance to be
Assessing the results of a cost–effectiveness reported.g
analysis against a structured review, such as that Details were extracted from this set of studies
reported here of all UK studies published up on the study design; whether or not patient-level
to 1996, should allow a better assessment of data had been collected on resource use and on
relative cost–effectiveness. For example, rather cost (and if so the sample size used); the size of
than vague and inevitably subjective comments the trial population from which any resource
that a particular result ‘compares well with sample was drawn; the number of arms in a
existing uses of healthcare resources’, analysts trial; the number of centres from which resource
(or the reader) could place results within a or cost data were collected; and the mean
distribution – ‘below the median figure for all follow-up period. Also recorded was whether
published UK cost–effectiveness results’, or the study reported a resource use mean, variance
‘within the first decile of published UK cost– or distributional form; the number of cost com-
effectiveness results’. Of course, it should be ponents reported, and for the control and inter-
noted that our database – as with league tables – vention arm(s) the reported mean or median
is based purely on published results, and does cost, the measures of cost variance reported, any
not imply that the interventions are necessarily mention of the distributional form of the cost
in use. A published ICER should not be equated data, any significance tests for cost differences.
with a currently accepted use of healthcare Of the 53 studies, 17 reported some statistical
resources. The acceptance of a more broadly measure of variance concerning use of resources,
based but methodologically similar set of com- and 25 reported some statistical measure of
parisons across the full spectrum of health- variance concerning costs. Three studies men-
related interventions, as proposed here, might tioned the distributional form of resource use
encourage adherence to standards, and mean- data (two normal, one positively skewed). In
while should help to reduce the amount of terms of the measures of cost variance reported,
subjectivity involved in interpreting results. five articles gave a standard error, seven a
standard deviation, four gave 95% confidence
intervals, two gave an interquartile range, and
DISTRIBUTION AND VARIANCE OF COST DATA 11 gave a range. In addition, one study reported
an indeterminate measure – of the form
With many more economic analyses being plan- ‘mean 6 x’ – which was probably either a stand-
ned prospectively, analysts are increasingly in ard error or standard deviation, but could not be
possession of patient-level data on the health- classified. In only four cases did articles report
care costs of alternative interventions. A com- more than one measure of variance: standard
mon belief is that such healthcare cost data are deviation and range (two), standard error and
(positively) skewed rather than following a nor- interquartile range (one), standard error and
mal distribution, and concerns have been raised range (one).
about what this may mean for standard statis- In the 26 RCT-related articles, 12 articles
tical tests.25 In this section we first examine how reported the value of the mean difference in
analysts reporting patient-level data on health- costs between trial arms, and three of these
care costs have reported the distribution of that articles reported some measure of variance in
data. Secondly, we examine available healthcare mean cost difference (95% confidence inter-
cost data to see to what extent they are in fact vals). Eight articles either reported the P-value
skewed, and we discuss the implications that for a significance test of mean cost difference
this may have for the statistical analysis of cost (four articles) or reported that the difference in
data. mean costs was not statistically significant
(four). In the 27 non-RCT-related articles, none
reported a mean cost difference with associated
Review of Studies Reporting variance between the patient groups or inter-
Patient-Level Cost Data ventions being compared (one reported a mean
cost difference with no measure of variance),
From the full set of 492 studies identified in and only one article reported a significance test
pp. 318–22, 26 studies were found which had for mean cost difference.
collected patient specific resource/cost data as In summary, therefore, only a tiny fraction of
part of a randomised controlled trial (RCT), and the published cost–effectiveness results retrieved
a further 27 were identified which had collected (15/492, or 3%) reported some conventional
patient-specific resource/cost data as part of a measure of variance (SD, SE, 95% CI) around
328 The Advanced Handbook of Methods in Evidence Based Healthcare
the mean cost estimates they used to calculate The non-normality of the data presented
cost–effectiveness ratios. above may cause problems for parametric statis-
From the 15 articles identified above as tical tests for the equality of two means. The
reporting patient-specific mean cost data and standard t-test is based on an assumption of
conventional statistical measures of variance, a normality of the underlying data and on the
total of 32 mean cost figures with associated assumption that the two population variances are
standard error, standard deviation or confidence equal. Although the t-test is known to be robust,
interval were extracted (for example, a three- i.e. that moderate failure to meet the assump-
arm trial with data for each patient group gave tions of normality and equal variance will not
three observations). The mean cost was 37 836 affect the results very much,26 it is not clear just
(in different denominations), the mean standard how great the departures from normality/equal
deviation was 40 150 and the corresponding variance must be before the t-test becomes in-
mean of the coefficients of variation was 1.01. appropriate. It is clear from the central limit
For the 28 out of 32 observations for which a theorem, for example, that although the under-
sample size was reported, the mean sample size lying data may not be normally distributed, the
was 78 (SD 82, median 37, range 2 to 279). sampling distribution of the difference between
two means, like the means themselves, will
approximate the normal distribution for large
Five Data Sets Describing Patient-Level sample sizes.27 A useful result is that the skew
Cost Data coefficient of the population will be reduced by
a factor of Î n in the sampling distribution of the
In addition to the review described above, we mean of that population, where n is the sample
obtained and analysed five healthcare cost data- size.h However, it is still not clear what value
sets to examine the variance and distributional of skew in a sampling distribution is ‘large
form of the data, and to explore the potential to enough’ to cause concern, such that some of the
employ data transformation techniques when large skewness observed in the five data sets
analysing and presenting the data. A description described above may be problematic despite the
of each of the five data sets can be found in the sample sizes concerned. We therefore consider
full report. Briefly, each data set contained two alternatives to the standard parametric
patient-level resource use data for a treatment approach that might be employed for the analy-
and control arm of a clinical trial, and these sis and presentation of the above data: the use of
resource data were weighted by unit cost infor- transformations to an alternative scale of meas-
mation to give patient-level costs associated urement in order that the data more closely
with the interventions under evaluation. Histo- conform to the standard statistical assumptions;
grams for each arm of the five data sets is given and the non-parametric approach of boot-
in Figure 11. It is clear that these patient-level strapping, which makes no assumptions con-
cost data do not seem to follow a normal dis- cerning the distributional form of the underlying
tribution, with many showing evidence of sub- data.
stantial skewness.
In these circumstances, the median is often
used as a measure of central tendency rather
Transformations
than the mean. However, although median cost The transformation of data from one scale to
provides useful descriptive information (particu- another can be used to overcome problems asso-
larly when presented alongside the mean), it is ciated with non-normality of data and unequal
inappropriate to use median costs in a cost variance. Fortunately, transformations which
analysis, as we are interested both in the average normalise data will often also provide more
per patient cost of a particular treatment and the equal variances.28 In the final report we consider
total cost of care for a patient group. Multi- three of the most commonly used transfor-
plying the median cost by the number of patients mations: the (natural) log transformation; the
treated will not give the total cost of treatment square-root transformation; and the reciprocal
for that patient group. Since, ultimately, some- transformation, and present the results from the
one will have responsibility for a budget from five data sets on the transformed scales.
which total costs of care will have to be met, the Although, in a statistical sense, transformed
appropriate statistic for analysts in economic scales of measurement are no less appropriate
analyses (and decision-makers applying the for analysing differences between two arms of a
results of such analyses) is the mean cost per study than the untransformed scale, transforma-
patient. A further problem with median statistics tion can lead to problems of interpretation for
is that convenient measures of dispersion, such economic analyses. It is not clear what the
as standard deviations or standard error do not magnitude of differences mean on the trans-
exist as they do for the mean. formed scales; hence analysts are likely to want
Handling Uncertainty in Economic Evaluation 329
Figure 11 Histograms showing the distribution of costs for the control and
treatment arms for the five data sets. Overlaid are normal distributions with the
same mean and variance as the data.
to back-transform point estimates of central transformed scale.29 Unfortunately, the cost dif-
tendency and confidence limits from the trans- ferences observed following a square-root trans-
formed scale to the original scale. This is a formation or reciprocal transformation are not
simple process of using the inverse function for interpretable when back-transformed.30 Cost dif-
point estimates of the original cost data on the ferences under a log transformation do have an
330 The Advanced Handbook of Methods in Evidence Based Healthcare
Normal F[(costdiff–m)/s]
Fraction
0.75
.1
Anticoagulation 0.50
0.25
0 0.00
–40 –20 Costdiff 0 20 0.00 0.25 0.50 0.75 1.00
.2 Empirical P(i)=i/(N+1)
1.00
Normal F[(costdiff–m)/s]
Fraction
0.75
Antiviral .1
0.50
0.25
0 0.00
–200 –100 0 50 0.00 0.25 0.50 0.75 1.00
Costdiff
Empirical P(i)=i/(N+1)
.2
1.00
Normal F[(costdiff–m)/s]
Leg ulcer
Fraction
0.75
.1
0.50
0.25
0 0.00
–500 –250 0 250 500 0.00 0.25 0.50 0.75 1.00
Costdiff
Empirical P(i)=i/(N+1)
.2
1.00
Normal F[(costdiff–m)/s]
Prostate
Fraction
0.75
.1
0.50
0.25
0 0.00
0 300 600 0.00 0.25 0.50 0.75 1.00
Costdiff
Empirical P(i)=i/(N+1)
Psychiatric 1.00
Normal F[(costdiff–m)/s]
.2
Fraction
0.75
0.50
.1
0.25
0 0.00
–100 –50 0 50 0.00 0.25 0.50 0.75 1.00
Costdiff Empirical P(i)=i/(N+1)
interpretation when back-transformed since the the prostate example and the psychiatric ex-
difference of two log values is equal to the log ample, the bootstrap estimate of the sampling
of their ratio. The antilog transformation of a log distribution closely approximates a normal dis-
value gives the geometric mean; hence the anti- tribution. Hence, for these studies we are un-
log of the mean log difference gives the ratio likely to want to transform results onto an
of the geometric mean in the control group to alternative scale, since we can be happy that the
the geometric mean in the treatment group. sample sizes are sufficient that the untrans-
Similarly, the antilog transformation of the con- formed t-test results are robust.
fidence limits for the difference in log values By contrast, for the anticoagulation example
gives the confidence limits to this ratio of the and the leg ulcer example, the bootstrap results
geometric means in each group.30 Hence, a sig- predict that some non-normality remains in the
nificant result after back-transformation is indi- sampling distribution due to the relatively small
cated by a confidence interval that excludes sample sizes in relation to the extreme non-
unity (i.e. where the ratio of the geometric normality of the underlying data. In these ex-
means is the same). amples, there may be a case for transforming the
However, although an interpretation is pos- data to ensure that the assumptions underlying
sible after back-transforming results on a log-
the t-test results are met. However, the bootstrap
transformed scale, an economic interpretation is
method can also be used to estimate confidence
still problematic for two reasons. Firstly, the
geometric mean is not the appropriate summary limits for the cost difference, which may be
statistic for economic analyses, for the same preferable to employing parametric techniques
reason the median was argued to be inappropri- on a transformed scale. The most straightfor-
ate above. Secondly, the back-transformation ward method of bootstrap confidence interval
does not result in the original scale of measure- estimation is the percentile method, by choosing
ment. If the aim is to combine information on the 2.5th and 97.5th percentile values from the
cost with information on effect to produce a vector of cost differences to represent the 95%
cost–effectiveness ratio, we need to express confidence interval limits. The bootstrap 95%
cost on the original scale (e.g. in pounds and percentile confidence interval for the anticoagula-
pence). tion data cost difference is 2£21.91 to £8.03,
while for the leg ulcer data, the 95% percentile
confidence interval is 2£258.36 to £242.96.
Non-parametric bootstrapping i These intervals are not symmetrically positioned
The bootstrap approach is a non-parametric around the point estimate of cost difference,
method that makes no distributional assump- reflecting the non-normality of the bootstrap
tions concerning the statistic in question. estimate of the sampling distribution.
Instead, it employs the original data and com-
puting power in a re-sampling exercise in order
to give an empirical estimate of the sampling
distribution of that statistic. Successive random CONFIDENCE INTERVALS FOR
samples of size n are then drawn from x with COST–EFFECTIVENESS RATIOS
replacement j to give the bootstrap re-samples.
The statistic of interest is then calculated for
Analysts will increasingly be in possession of
each of these re-samples, and these bootstrap
replicates of the original statistic make up the patient-level data on both cost and health out-
empirical estimate of the statistic’s sampling come as more economic analyses are conducted
distribution. alongside clinical trials. The overall review in
The results of bootstrapping the sampling pp. 318–22 found that just 24 studies had
distribution of the cost differences between the attempted any form of statistical analysis in
treatment and control arms of the five datasets looking at their data, and in only three cases was
are presented in Figure 12. What the bootstrap this analysis related to the ICER itself. Standard
estimates of the sampling distributions show is statistical methods of confidence interval esti-
the extent to which, in each case, the sample mation do not apply to the ICER since the
size of the study is sufficiently large that the variance of a ratio statistic is intractable. There
assumption of a normal sampling distribution has been much recent research into possible
(from the central limit theorem) is justified. methods for estimating confidence intervals for
Where the sampling distribution approximates the ICER stemming from a seminal paper by
the normal distribution, we need not be too O’Brien and colleagues.6 In this section, we
concerned about the violations of the assump- offer an overview of this recent literature in
tions underlying the t-test results reported order to elucidate the strengths and weaknesses
above. It is clear that, for the antiviral example, of the alternative approaches. We also consider
332 The Advanced Handbook of Methods in Evidence Based Healthcare
the use of an alternative approach to represent- including methods based on confidence boxes,6
ing uncertainty in cost–effectiveness informa- confidence ellipses,31 the Taylor series expan-
tion and discuss the continuing role of sensitiv- sion,6 Fieller’s theorem32,33 and non-parametric
ity analysis for handling uncertainty not related bootstrapping.32,34,35 Each of the approaches out-
to sampling variation. lined above are illustrated using patient-specific
datak obtained from a clinical trial that compared
Confidence Interval Estimation Methods a drug therapy treatment group with a placebo
control group.36 The representation of uncer-
In recognition of the problem associated with tainty based on each of the methods outlined
standard statistical methods for estimating con- above is shown in Figure 13 for the clinical trial
fidence limits for the ICER, a number of data. The horizontal and vertical ‘I’ bars rep-
analysts have proposed alternative methods, resent the 95% confidence intervals on incre-
Table 3 Comparing methods of confidence within the interval. This simulation is then
interval estimation: a numerical example repeated a large number of times and the propor-
tion of times the true value of the ICER is
Method Lower Upper Interval contained within the interval is used to judge
limit limit length the coverage properties of the confidence inter-
val calculation methods. The results of two
Confidence box:
such articles have shown that Fieller’s method
Standard 25.2 606.7 611.9
Revised 49.3 290.8 241.5
and bootstrap methods outperform the other
Taylor series 221.89 295.94 317.83
methods.38,39
Confidence ellipse 248.7 523.4 572.1 Given that the Monte Carlo evaluation studies
Fieller’s theorem 28.3 380.6 388.9 show the bootstrap method to be a good overall
Bootstrap percentile 25.9 392.0 397.9 method, the bootstrap can be used to show the
inaccuracies of the inferior methods from Table
3 and Figure 13. For the Taylor series method,
6.8% of the bootstrap replications gave ICERs
that were above the upper estimate of the con-
mental effects and cost separately, which inter- fidence limit, while 1.5% were below the lower
sect at the ICER point estimate of 137.18. The limit, suggesting an overall confidence interval
dashed box is the confidence box generated by of 91.7%. The ellipse method had an overall
these intervals. The ellipse shown joins points of confidence interval of 98.4% (0.74% boot-
equal probability (assuming a joint normal dis- strap replications below the lower limit and
tribution) such that the ellipse covers 95% of the 0.84% above the upper limit), a greater over-
integrated probability. Also shown are 5000 specification than the standard box method at
bootstrap replications of the incremental effect 96.9% (2.56% below and 0.56% above). The
and incremental cost pairs and the rays whose potential pit-falls of the revised box approach
slopes correspond to the upper and lower 95% are plainly apparent (note that cost and effect
confidence limits of the ICER, as calculated by differences in the numerical example used in
the Taylor, Fieller and bootstrap (percentile) Table 3 are almost independent), since the
methods. The confidence limits for each of the method gives just an 80% confidence interval:
different methods, together with the interval 12.6% of the bootstrap replications fell below
length, are shown in Table 3. A full review and the lower limit and 7.3% were above the upper
details for calculating these confidence limits limit.
can be found in the final report.
What is immediately apparent from this analy-
sis is that the symmetrical nature of the Taylor Cost–Effectiveness Acceptability Curves
series method gives a very different balance to
the interval than the other methods. It is also In the final report we identify a number of
interesting to observe that, in this example, the potential problems for confidence intervals as a
interval from the ellipse method is not internal method for representing uncertainty in the cost–
to the interval from the box method, a result effectiveness of healthcare interventions. These
that contradicts recent representations in the relate to the interpretation of negative ratios, the
literature.8,37 one-sided nature of the decision problem, un-
Figure 13 and Table 3 emphasise the differ- certainty concerning the maximum acceptable or
ences in confidence intervals that can be ob- ceiling ICER appropriate for decision making,
tained for the same data when employing differ- and the adherence to conventional levels of
ent methods. However, it is not possible to judge statistical significance.
which method is most appropriate on the basis Cost–effectiveness acceptability curves offer
of a single example. Recall that the definition of an alternative approach to presenting and think-
a 95% confidence interval is that the true popu- ing about uncertainty in cost–effectiveness
lation parameter will be included in the interval analysis. Although originally proposed by van
95% of the time in repeated sampling. In order Hout and colleagues,31 they have received com-
to estimate the accuracy of the intervals ob- paratively little attention in the recent literature.
tained by different methods, simulation experi- If we define a ceiling cost–effectiveness ratio
ments, where the true population ICER is appropriate for decision-making, this can be
defined, can be undertaken. Such experiments represented on the cost–effectiveness (CE) plane
involve drawing samples of costs and effects (see Figure 1) as a ray passing through the
from predefined distributions, calculating con- origin, with slope equal to the ICER ceiling
fidence intervals for the estimated ICER, and value. This divides the CE plane into two, and
recording whether the true value is contained van Hout and colleagues refer to the area of the
334 The Advanced Handbook of Methods in Evidence Based Healthcare
CE plane lying to the right of the ceiling ratio as the data requires a Bayesian approach to statis-
the acceptability surface. They advocate the con- tical analysis.
struction of a cost–effectiveness acceptability This section has been concerned primarily
curve on the basis of the probability an inter- with handling uncertainty in cost–effectiveness
vention is cost-effective in relation to all pos- analysis arising from sampling variation. How-
sible values of the ceiling ratio from 0 through ever, it is important to remember that sensitivity
to ∞ (equivalent to rotating the ceiling ratio line analysis methods have a continuing role along-
on the CE plane from the horizontal through to side statistical methods for handling other im-
the vertical). portant types of uncertainty in economic evalu-
The cost–effectiveness acceptability curve for ation. A further advantage of cost–effectiveness
the clinical trial data is shown in Figure 14. For acceptability curves is that a range of different
all possible values of the ceiling ratio (x-axis), curves arising from a number of sensitivity
the probability that the intervention is cost- analyses may be easier to interpret than a corres-
effective is shown (y-axis). Note that the point ponding number of confidence intervals.
of inflexion on the cost–effectiveness accept-
ability curve occurs at the midpoint of the prob-
ability function and corresponds to the baseline
cost–effectiveness data for the clinical trial data GUIDELINES ARISING FROM THE REVIEW
of 137.18. The point at which the curve crosses
the vertical axis corresponds to the P-value In this concluding section we present the guide-
associated with the incremental cost of the inter- lines arising from the findings in our structured
vention, and the limit to which the curve is review. Full details and justification for these
tending corresponds to the P-value associated guidelines are given in the full report.
with the incremental effect of the intervention.l
Although cost–effectiveness acceptability curves 1 Analysts should aim to present results
are closely related to P-value functions, strictly using a methodological reference case in
their interpretation of showing the probability order to increase the comparability of
that an intervention is cost-effective given that results between studies.
Handling Uncertainty in Economic Evaluation 335
2 Analysts should be aware of the potential interpretation and use of transformations, and
for the incremental cost–effectiveness ratio provided comments on the review of confidence
to vary at the margin. interval methods reported. Simon Dixon of the
3 Analysts should avoid selective comparison Sheffield Health Economics Group provided the
of their results with the results from other leg ulcer data used as one of the examples, and
studies. Professor Paul Fenn and colleagues provided
4 Analysts should ensure that they consider access to their data for another of the examples.
the potential implications of uncertainty for We are grateful also to the editors and anony-
the results of their analysis. mous referees of Journal of Health Service
5 Interval estimates should accompany each Research & Policy and Health Economics who
point estimate presented. provided feedback on articles to be published
6 Where sensitivity analysis is employed to based on parts of this report. This feedback has
estimate an interval, analysts should be contributed to the revision of this final report.
comprehensive in their inclusion of all vari- The Institute of Health Sciences library staff
ables in the analysis. provided help in obtaining many of the articles
7 When reporting sensitivity analysis, ana- included in the review. Revision of the report
lysts should be aware of the probabilistic and updating of the database to include the 1996
nature of the reported range. publication year was made possible through a
8 When reporting patient-level cost informa- joint MRC/Anglia and Oxford Region Special
tion, analysts should make more use of Training Fellowship held by A.B. Responsibility
descriptive statistics. for the views expressed in this paper, and for all
9 Even when data are skewed, economic remaining errors, is our own.
analyses should be based on means of
distributions.
10 When reporting statistical tests of cost-
differences, analysts should be aware that NOTES
significance tests may be more powerful on
a transformed scale, but that confidence a. Although we focus on UK studies, the general
limits should be reported on the original principles employed in this section are internationally
scale.
applicable.
11 Where patient-level data on both cost and
b. One of the specified criteria for inclusion on
effect are available, the parametric ap-
the overall database was that studies were published
proach based on Fieller’s theorem or the
journal articles. We made two exceptions to this rule
non-parametric approach of bootstrapping
by including the Forrest Report40 on breast cancer
should be employed to estimate a con-
screening and the Standing Medical Advisory Com-
fidence interval for the cost–effectiveness
ratio. mittee report41 on cholesterol testing. Both were felt to
12 Sensitivity analysis has a continuing role in be high-quality reports that had been influential from
handling uncertainty not related to sampling a policy perspective without being published in a
variation. journal.
13 Consideration should be given to using cost– c. The UK Treasury has recently revised their
effectiveness acceptability curves to present recommendation concerning the rate of discount for
uncertainty in stochastic cost–effectiveness health outcomes from 1.5 to 2%, while the recom-
studies. mended rate for costs stays at 6%. We adopt the
previous recommendation here as that was applicable
when the review was started, and when most analysts
were conducting their studies.
ACKNOWLEDGEMENTS d. We defined a ‘full range’ as being two-sided and
excluded those results for which the sensitivity analy-
This report was funded by the NHS Executive sis range was ‘one-sided’ in either direction such that
through the National Co-ordinating Centre for the baseline result was also a high or low value for the
Health Technology Assessment. A number of range.
people have contributed directly to the produc- e. The reported distribution of the ICER results and
tion of this report. Professor Paul Fenn and other statistics excluded 10 of the 333 baselines in
Professor Alistair McGuire provided helpful which the intervention in question was dominated by
comment on the design of the research, and the comparator.
Professor Fenn also made constructive com- f. Overall, external ICERs quoted by analysts tend
ments on the review of methods for calculating to under-represent the low values and over-represent
confidence intervals. Dr David Smith and Sir the mid-range values. However, analysts appear not to
David Cox provided statistical advice on the quote the very highest ICER results available.
336 The Advanced Handbook of Methods in Evidence Based Healthcare
g. A further 12 studies (three RCT, nine non-RCT) 9. Light RJ, Pillemer DB. Summing up: the science
were identified which had collected patient-specific of reviewing research. Cambridge: Harvard Uni-
information in the course of the analysis, but analysed versity Press, 1984.
it in such a way (aggregating resource use in each arm 10. Chalmers I, Altman DG. Systematic reviews.
of a trial, applying unit costs and then dividing by n, London: BMJ, 1995.
or variants of this approach) that it was not possible to 11. Gerard K. Cost–utility in practice: a policy
calculate inter-patient cost variance. maker’s guide to the state of the art. Health Policy
h. This is a precise result, under standard para- 1992; 21: 249–79.
metric assumptions, that can be derived from the 12. Briggs AH, Gray AM. Handling uncertainty when
definition of the skew coefficient. performing economic evaluation of health care
i. In the final report, we also consider other com- interventions: a structured review of published
monly used non-parametric rank order tests. However, studies with special reference to the distributional
we argue that these are not appropriate for cost form and variance of cost data. Health Tech-
analysis where difference in mean cost is the prime nology 1999; 3.
concern. 13. Blades CA, Culyer AJ, Walker A. Health service
j. Of course, sampling from x without replacement efficiency: appraising the appraisers – a critical
n times, would simply yield x itself. review of economic appraisal in practice. Soc.
k. This is the same antiviral cost dataset from Sci. Med. 1987; 25: 461–72.
pp. 327–31, but this time we include an analysis of 14. Field K, Thorogood M, Silagy C, Normand C,
health outcome. O’Neill C, Muir J. Strategies for reducing coronary
l. Note that due to the one-sided nature of the curve, risk factors in primary care: which is most cost
the P-values associated with a two-sided hypothesis effective? Br. Med. J. 1995; 310: 1109–12.
test would be twice the values from Figure 12. 15. Dusheiko GM, Roberts JA. Treatment of chronic
type B and C hepatitis with interferon alfa: an
economic appraisal [see comments]. Hepatology
1995; 22: 1863–73.
REFERENCES 16. Daly E, Roche M, Barlow D, Gray A, McPherson
K, Vessey M. HRT: an analysis of benefits, risks
1. Backhouse M, Backhouse RJ, Edey SA. Economic and costs. Br. Med. Bull. 1992; 48: 368–400.
evaluation bibliography. Health Economics 1992; 17. Fenn P, Gray AM, McGuire A. The cost-
1(Suppl.): 1–235. effectiveness of thrombolytic therapy following
2. Anderson JP, Bush JW, Chen M, Dolenc D. acute myocardial infarction. Br. J. Clin. Pract.
Policy space areas and properties of benefit-cost/ 1991; 45: 181–4.
utility analysis. JAMA 1986; 255: 794–5. 18. Freemantle N, House A, Song F, Mason JM,
3. Black WC. The CE plane: A graphic representa- Sheldon TA. Prescribing selective serotonin re-
tion of cost-effectiveness. Med. Decis. Making. uptake inhibitors as strategy for prevention of
1990; 10: 212–14. suicide. Br. Med. J. 1994; 309: 249–53.
4. Petrou S, Malek M, Davey PG. The reliability of 19. Gerard K, Mooney G. QALY league tables:
cost–utility estimates in cost-per-QALY league handle with care. Health Economics 1993; 2:
tables. PharmacoEconomics 1993; 3: 345–53. 59–64.
5. Briggs AH, Sculpher MJ, Buxton MJ. Un- 20. Mason J, Drummond M, Torrance G. Some
certainty in the economic evaluation of health guidelines on the use of cost effectiveness league
care technologies: the role of sensitivity analysis. tables. Br. Med. J. 1993; 306: 570–2.
Health Economics 1994; 3: 95–104. 21. Drummond M, Torrance G, Mason J. Cost-
6. O’Brien BJ, Drummond MF, Labelle RJ, Willan effectiveness league tables: more harm than
A. In search of power and significance: issues good? Soc. Sci. Med. 1993; 37: 33–40.
in the design and analysis of stochastic cost- 22. Gold MR, Siegel JE, Russell LB, et al. Cost-
effectiveness studies in health care. Med. Care effectiveness in health and medicine. New York:
1994; 32: 150–63. Oxford University Press, 1996.
7. Briggs AH. Handling uncertainty in the results of 23. Department of Health. Hospital and Community
economic evaluation. OHE Briefing Paper No. 32, Health Services pay and price index. London:
1995. Department of Health, 1996.
8. Manning WG, Fryback DG, Weinstein MC. 24. Hillman AL, Eisenberg JM, Pauly MV, Bloom
Reflecting uncertainty in cost–effectiveness BS, Glick H, Kinosian B, Schwartz JS. Avoiding
analysis. In: Gold MR, Siegel JE, Russell LB, bias in the conduct and reporting of cost-
Weinstein MC (eds). Cost-effectiveness in health effectiveness research sponsored by pharmaceut-
and medicine. New York: Oxford University ical companies. N. Engl. J. Med. 1991; 324:
Press, 1996. 1362–5.
Handling Uncertainty in Economic Evaluation 337
clinical trials and clinical practice. Currently, proving with, for example large databases of
there are a large number of different QoL ques- therapeutic trials in the Cochrane Collaboration.
tionnaires, indices and scales being used and Increasingly, QoL is being chosen as a major
translated for many countries. However, many end-point in clinical trials, and thus is one of the
of these instruments are either based on some- main determinants of the size of a sample. It is
what general classifications of health, or are now recognised that QoL measures often have
restricted to one physical illness. De Haes and very skewed distributions and the usual sample
Van Knippenberg1 found that the philosophy size calculations are not valid.4–6 Thus, we
and definition of QoL differed from one instru- decided to review the distributional form of
ment to another, and that many of these some commonly used QoL measures and meth-
instruments also had other methodological prob- ods used to determine sample sizes in clinical
lems, particularly with respect to their reliability, studies.
validity and design. This chapter considers the following issues:
The majority of early instruments for assess-
ing QoL are: 1 To discover which QoL measures are being
used in cancer studies.
(a) lengthy and complicated for both physicians 2 To obtain an understanding of the distribu-
and patients; tion of the score of three regularly used QoL
(b) difficult to score, often requiring trained instruments.
personnel; 3 To provide guidelines for methods of sample
(c) restricted to physical function, with little size estimation for QoL data.
regard to social and emotional issues;
(d) not validated or tested for reliability; and
(e) not able to differentiate between people
with different levels of QoL. METHODS
Literature Search
WHY THIS ISSUE IS IMPORTANT
The literature search was conducted using the
EMBASE section contained within the Bath
Quality of life is particularly important in palli- Information Data Services (BIDS) database. The
ative medicine because, for many cancers, little most general strategy was used; search expres-
improvement in cure rates or survival times has sions being entered into the ‘title, abstract or
been achieved in recent years. keywords’ option. The keywords used were
Sample size issues are important for the plan- ‘sample size’, ‘effect size’, ‘quality of life’,
ning of clinical trials. Studies that are either too ‘sample size 1 quality of life’ and ‘cancer 1
small or too large may be judged unethical,2 clinical trial’. This led to a vast number of QoL
although there is some disagreement over this.3 instruments that were being used in all types of
For example, a study that is too large could have studies and populations. Because of the large
demonstrated clinical improvement before the
number of QoL instruments initially found, it
end of the trial had been reached, and so some
was decided that this project should concentrate
patients may have been randomised to a treat-
ment that could have been shown to be inferior. on the three QoL questionnaires that have been
On the other hand, provided that it is not a life- used by the British Medical Research Council
or-death situation, large trials give the opportu- in their cancer clinical trials. These are the
nity for examining a variety of end-points, and Hospital Anxiety and Depression Scale (HADS),7
possibly subgroups of patients. A trial that is too the Rotterdam Symptom Checklist (RSCL)8 and
small will have no power to demonstrate a the European Organisation for Research and
meaningful effect, and so may be regarded as a Treatment in Cancer Questionnaire (EORTC
waste of resources; moreover, patients are being QLQ-C30).9
put through the potential trauma of a trial for no Concentrating on these three instruments, a
tangible benefit. Inconclusive trials can also second and more precise search was carried out,
make the conduct of subsequent trials difficult if seeking papers that cited the original paper
a marginal benefit that may have arisen by describing the questionnaire. This has shown a
chance is shown. However, it could also be steady increase in the number of cancer clinical
argued that any evidence is better than none, and trials using these three questionnaires as out-
provided that the results are reported they may come measures. This can be seen in more detail
be included in a formal synthesis by overview. in the results section within this chapter.
The problem here is that it is difficult to publish Although the searches revealed an increase in
inconclusive results, although matters are im- the number of citations, and hence an apparent
340 The Advanced Handbook of Methods in Evidence Based Healthcare
increase in usage of these three QoL ques- The Hospital Anxiety and Depression
tionnaires, it is not certain that the actual fre- Scale (HADS)
quency of their usage has been determined.
Certainly papers will not have been found due to This was developed by Zigmond et al.7 in 1983.
the keywords not being in their title or abstract, It was designed to measure two psychological
or the authors not being cited. Other clinical dimensions of QoL, those of anxiety and depres-
trials may not have been published because they sion, in patients who were physically ill, and
were conducted in the pharmaceutical industry, therefore it excluded somatic symptoms that
which frequently does not publish its results in might be due to illness. It is a self-rating ques-
peer-reviewed journals. More might have been tionnaire which a patient completes in the wait-
missed due to the ‘negative finding bias’ of ing room before meeting a doctor in order to
researchers and journals. Another factor which reflect how they have felt during the past week.
could not be measured is the possible lag effect The questionnaire has 14 items which split
in the use of QoL instruments. The measure equally into the two subscales; each question
being used is the number of published articles has four possible responses with scoring levels
using QoL measures. However, the usage may of between 0 and 3. For this question, the
be more widespread as it takes time to design, responses are ‘most of the time’ (3), ‘a lot of the
conduct, report and publish a clinical trial, time’ (2), ‘time to time’ or ‘occasionally’ (1),
resulting in a lag between the use of QoL and ‘not at all’ (0).
measures and the article quoting them. If these The total score for the seven questions in each
factors remain constant, however, the apparent of the two subscales lies therefore between 0
increase in use should reflect reality. and 21. Patients can be categorised into one of
We have not investigated whether these ques- three clinically predefined categories. A total
tionnaires are being used by oncologists and score of 0–7 implies normality, 8–10 is border-
physicians in clinical practice in the UK. In a line, and a score of 11–21 suggests that a patient
recent paper10 on QoL perspectives of 60 oncol- has significant anxiety or depression.
ogists in America and Canada, only 7% stated The advantages of the HADS are that it is
that they formally both collected and used QoL short, has a simple scoring system, is accepted
information. An important question that needs to by patients, is easy to understand, and is quick
be answered is, ‘How to really incorporate QoL to complete and administer. Silverstone et al.18
issues into clinical practice?’ suggested that the HADS does not accurately
diagnose the presence of major depressive dis-
orders in either medical or psychiatric patients.
Data Sets Nonetheless, they concluded that the HADS
should continue to be used as a clinical indicator
Five data sets were obtained from the Medical
of the possibility of a depressive disorder.
Research Council Cancer Trials Office in Cam-
Moorey et al.19 have carried out factor analysis
bridge. These studies have already been pub-
on the structure and stability of the HADS using
lished.11–13 One data set will be explored in
principal-components analysis on the two sub-
detail here.11 These data provide scores for dif-
scales separately. Their results suggest that the
ferent QoL dimensions and subgroups of lung
items on the two subscales do separate the
cancer patients receiving either a standard or a
concepts of anxiety and depression extremely
new treatment. The QoL questionnaires were
well, and they concluded that the HADS is a
completed together at different times over 6
useful instrument for measuring these dimen-
months. The sample size estimates from these
sions in cancer patients. They also felt that for
data are not generalisable to other studies, unless
patients with early stage cancer the repeated
the patients in these studies are sufficiently
administration of its two subscales was justified,
similar. However, the methods of calculation
and provided the physician with practical facts
and the issues raised by the application of these
concerning the patient’s progress.
methodologies are generalisable
Sample size calculations have been described
in a number of recent papers and a book,14–17 and The Rotterdam Symptom Checklist
one based on this work published recently.6 (RSCL)
This was designed by De Haes et al.8 in 1990,
initially as an instrument to measure the symp-
BACKGROUND toms reported by cancer patients participating
in clinical research. It incorporates 38 items,
Descriptions of the three QoL questionnaires each rated on a 4-point scale (between 0 and 3)
under review are given below. and measures both physical and psychological
Quality of Life Measures in Cancer Clinical Trials 341
dimensions. The psychological dimension con- or medical treatment caused you financial diffi-
tains eight questions, and hence has a total score culties?’ These replies and their corresponding
ranging from 0 to 24. It has two clinically scores are ‘very much’ (4), ‘quite a bit’ (3), ‘a
predefined categories where a total score of little’ (2) and ‘not at all’ (1).
0–10 is considered to be a Non-Case and a In addition to the core, modules are used
patient scoring 11–24 is considered a Case of which relate to specific cancers, some of these
being psychologically distressed. Conversely, are still in the development stage. The core also
the interpretation of the physical dimension has 18 translations available at present, plus six
(activities of daily living), which has seven pending translations and four English versions
questions with a total score ranging from 0 to for English-speaking countries. Nine of these
21, is reversed, with a higher score implying a translations have a separate male and female
better physical condition. The four possible version. Each is a patient self-rating question-
responses and scores to each question are ‘able’ naire where the format changes according to the
(3), ‘with difficulty’ (2), ‘only with help’ (1) and module being used. The core questionnaire is
‘unable’ (0). distributed at clinics by nurses and takes less
The RSCL advantages are that it is clearly than 10 min to complete. It is easy to score and
laid out, is easy to score, is readily administered produces both a total score and scores on each
by nurses, it takes 5–10 min to complete, and subscale. Whilst the scale has been carefully
it can be divided into distinct subscales scores developed on the basis of existing instruments,
for the physical and psychological areas. This its reliability and validity have still to be prop-
instrument was carefully designed from a pool erly assessed, as the core has only been on
of items based on existing questionnaires and release since 1993 and it still has to be checked
involved check-lists from interviews with cancer against ‘gold standard’ interviews to see how it
patients. Patients indicate how much they have performs. Further details of the development of
experienced particular symptoms over the past the EORTC QLQ-C30 and the EORTC studies
week. Maguire et al.20 considered that the RSCL that use it have been given recently.27
was a good, clear and simple questionnaire The relative merits of the three questionnaires
which has been validated against ‘gold standard’ are given in Table 1.
interviews. They found it to have both high
sensitivity and specificity in measuring the psy-
chological dimensions. It has been employed
successfully within busy clinics, and appears
RESULTS OF LITERATURE REVIEW
as effective for patients with both early or
advanced cancer, and seems to measure the In 1990, 27% of all deaths in England and
physical and social dimensions equally well. Wales were due to cancer.21 The numbers of
The reliability of the RSCL has again been cancer QoL clinical trial citations using just
assessed using principal component analyses, these three questionnaires as outcome measures
and details of the three studies used for this per 1000 cancer deaths are displayed as rates
validation can be found in De Haes and Van and are presented in Table 2. This table gives an
Knippenberg.1 overall citation rate for all cancer types, and
displays the rates for specific cancer QoL cita-
tions per 1000 specific cancer deaths.
The European Organisation for The reason why the lung cancer rate is lower
Research and Treatment in Cancer in comparison to the others, even though it is
Questionnaire (EORTC QLQ-C30) highly cited, is because of the large number of
deaths from lung cancer. Conversely, the skin
This was designed by Aaronson et al.9 in 1993, cancer rate appears high because of the low
and is a 30-item core questionnaire containing number of deaths: only about 2% of skin cancer
multi- and single-item scales that reflect the patients actually die from skin cancer. There-
multi-dimensionality of QoL. It contains five fore, if one quoted the rates as QoL citations per
functional scales, three symptom scales, a global 1000 cancer incident cases, the magnitude of
health QoL scale, and six individual items. The these figures may alter.
core QLQ-C30 questionnaire has a scale which Since the development of the HADS ques-
is easy to score and is designed to evaluate tionnaire in 1983 there has been a steady
several multi-dimensions of emotional, physical increase in citations both from generic and can-
and social issues for heterogeneous groups of cer trials, to total nearly 500. The frequency of
cancer patients. The responses for the physical citations is displayed in Figure 1 and indicates
items are ‘yes or no’ and for the symptom items clearly an increase in popularity of the HADS in
categorical. An example of the social function- cancer clinical trials, whilst the cancer citations
ing dimension is ‘Has your physical condition as a percentage of all HADS citations are shown
342 The Advanced Handbook of Methods in Evidence Based Healthcare
Table 2 The rate of cancer QoL citations per cancer patients since its development, including
1000 cancer deaths all those of the HADS questionnaire, plus gas-
tric, prostate and mixed cancer trials.
No. of No. of It is too recent for the EORTC QLQ-C30 core
Cancer deaths citations Rate questionnaire to have been widely cited since its
development in 1993, but we identified a total of
All 76 000 233 3.07 52 citations in its first two years of use, see
Lung 39 000 21 0.54 Figure 4. The EORTC QLQ-30 has been used in
Breast 15 000 49 3.27 several types of cancer trials including bladder,
Skin 500 4 8.00 penis, brain and oesophagus.
Pancreas 6000 6 1.00
Figure 1 Number of citations per year for the HADS questionnaire in cancer
studies.
Figure 2 Number of citations per year for the HADS questionnaire in cancer
studies as a proportion of all HADS citations per year.
344 The Advanced Handbook of Methods in Evidence Based Healthcare
i51
3
i
ferent number of patients to one that tried to With this equation, m is the total number of
decrease it. patients required in each group, OR is the odds
Quality of Life Measures in Cancer Clinical Trials 347
ratio of a subject being in a given category of a limited number of values if the control group
less in one group compared to the other group, k already has a high success rate.
is the number of categories and pi is the mean
proportion expected in category i, i.e. pi 5
(pAi 1 pBi)/2 where pAi and pBi are the propor- Sample Size Comparison
tions expected in category i for the two groups
A and B respectively. This equation is quite The final outcome of the trial is summarised in
complicated to use, but rather than go through Table 3 for the HADS and the RSCL, as a
details here we refer the reader to two papers standardised difference in mean scores or as an
and a book that discuss its application in more odds ratio for either a Non-Case for the HADS
detail.6,15,16 and RSCL or for a borderline Case for the
Equation (2) makes no assumptions about the HADS alone. It can been seen that using either
distribution of data, but it does assume propor- outcome the effect sizes are very small, and
tional odds, where the odds of being in a given essentially there is little difference in the out-
category or less, in one treatment group are comes. However, for the HADS the assumption
fixed relative to the odds of being in the same of a constant odds ratio would appear to be true.
category or less in the other treatment group. In Table 4 we ask the question ‘Suppose we
Thus the ORs are identical for each pair of wished to design a new trial in similar patients
adjacent categories throughout the QoL scale. It as before, where the HADS anxiety dimension is
is this assumption of a constant odds ratio (pro- to be the prime outcome. We assume the dis-
portional odds) that allows one to generalise the tribution of baselines was as in the earlier trial,
OR calculated from a predetermined cut-off to and postulate as an effect size the observed
the remaining categories of the scale. effect in the earlier trial. What size sample
The calculation is based on the proportions would be required, with 80% power at 5%
and cumulative distribution in each QoL cat- significance?’ The effect sizes as measured by a
egory across dimensions. The application of the difference in means or as an odds ratio both
proportional odds in this formula assumes that refer to the same data set, and so are attempting
when the distribution of the outcome for only to summarise the same quantity, which is the
one of the treatment arms is known, the expec- difference in the distribution of the outcome
ted cumulative proportions for the other arm can measure under the two treatments. The sample
be obtained and from these the expected propor- size was estimated in three ways: (i) using
tions derived for each pair of categories. Hence Equation (1) and the difference in means; (ii)
with prior knowledge of the distribution of just for the HADS dichotomising the data at two
one treatment group, the OR and the sample size separate cut-off points and using the two separate
estimate can be calculated. This calculation ORs obtained in Equation (2) (a methodology
gives different results for risks which are equal, often applied by researchers aware of the non-
but either represents an improvement (OR) or a Normality of the data); and (iii) for the HADS
deterioration (1/OR). This is sensible since it is taking the average of the two ORs and using all
harder to demonstrate an improvement than a three categories again using Equation (2). For
deterioration in an outcome variable which takes the RSCL there was only one score value for
separating Cases and Non-Cases, and so there is the sign of the expected effect size. However,
no way of extending the sample size calcula- the non-parametric techniques, shown in Equa-
tions to ordinal data. The results are quite tion (2), for the QoL dimensions give markedly
instructive. For the HADS, using a binary cut- different sample sizes according to the direction
off leads to a much higher sample size than that of the expected effect size.
using the parametric methods. Methods for Our results have shown that many researchers
ordered categorical data result in a marginal quote means and standard deviations and imply
saving of the sample size, compared with simple a Normal distribution. We have demonstrated
binary cut-offs. One might imagine that because that this assumption in sample size calculations
the parametric method utilises much more infor- can lead to unrealistically sized studies, and we
mation, the sample sizes would always be much have highlighted the discrepancies between
smaller than for the binary method. However, parametric and non-parametric techniques. We
for the RSCL this was not the case, with the recommend that all researchers publish the dis-
parametric method requiring more subjects. This tribution of the scores at baseline and at final
was because of the lack of Normality in the outcome.
RSCL, so that the parametric assumptions are Our recommendations are to use the non-
invalid. parametric formula, Equation (2), for the sample
From this exercise we learned that: (i) lack size estimations. This will allow health pro-
of Normality seriously affects the parametric fessionals involved in the care of cancer patients
method; (ii) a constant odds ratio model is to develop interventions which maximise out-
reasonable for HADS for this data set; and (iii) come in terms of both a patient’s overall sur-
different outcome measures (e.g. the HADS and vival and their QoL.
the RSCL) have markedly different sample size
requirements.
INTERPRETATION OF TREATMENT EFFECTS
DISCUSSION AND CONCLUSIONS We have not been able to determine a clinically
relevant treatment effect because we were deal-
The European Guidelines for the number of ing with retrospective data sets which show
patients needed when conducting a clinical trial what might be achieved, but not what is desir-
are given in a paper by Lewis et al.24 The sample able to achieve. It is important that a QoL
size should always be large enough to yield a measure can be responsive to change, and it is
reliable conclusion, and be the smallest neces- the change from baseline that is often the most
sary to reach this conclusion and the intended important outcome. Guyatt et al.25 discuss the
aim of the trial. If the sample size is too small, interpretation of QoL scores. They suggest that
the usefulness of the results will be reduced even if the mean difference between a treatment
because of the inadequate number of patients and a control is appreciably less than the small-
and clinically significant changes may be con- est change that is important, treatment may have
cealed. However, if the sample size is too large an important impact on many patients. They
the trial will incur a waste of resources and suggest that the smallest change in a categorical
patients may continue to receive a less effective variable that is important to a patient is 0.5 per
or even dangerous treatment. These reasons to question, with 1.0 being a moderate change and
calculate an adequate sample size before con- 1.5 being considered large. They describe a
ducting a clinical trial are well known. method by which these can be converted into a
Fayers and Machin17 reported that it is also proportion of patients who benefit from a treat-
known that many randomised clinical trials with ment, which can be compared to the number-
treatment comparisons are of an insignificant needed-to-treat (NNT) provided with the usual
size to place weight behind the trial conclusions. clinical outcomes. We would not advise their
Additionally, they recommend that one should method for the parallel group trial however,
calculate the sample size for a number of likely because it requires an assumption that responses
estimates of the effect size, to try and give an within a patient are likely to be independent.
opinion as to whether the proposed trial is likely Stockler et al.26 looked at the responsiveness to
to be impractical. change of the EORTC QLQ-C3027 and com-
It is evident that the methodology which pared it with a more specific instrument in
assumes a symmetric Normal distribution for the prostate cancer. They concluded that the QLQ-
QoL dimensions give symmetric sample sizes, C30 is responsive to changes in health-related
as the parametric techniques for Normal data, quality of life, and suggest that the QLQ–30 is
shown in Equation (1), depend only on the more responsive to global perceptions of health
absolute value of the difference regardless of than the more specific questionnaire. However
Quality of Life Measures in Cancer Clinical Trials 349
as Stucki et al.28 point out, numerically equal justify the size of clinical trials in papers. How-
gains in an ordinal score may differ in their ever, one of the most prevailing errors in clinical
meaning depending on the baseline health status. trials is a lack of justification of the sample size
They recommend that the distribution of base- and one paper has concluded ‘the reporting
line health status measures and the distribution of statistical power and sample size needs to
of responders by baseline status be reported in be improved’.24 The identification of correct
evaluative studies. sample size estimates for use in clinical trials
The sample size is most critically dependent circumvents the unethical under- or over-
on the estimation of the treatment effect size. If recruitment of patients to a particular trial.
you halve the effect size, then you quadruple the We have provided recommendations and pro-
sample size.17 The effect size is commonly taken duced a table which supplies evidence of the
as a difference which is the minimum to be advantage of the non-parametric technique used
clinically meaningful, but its interpretation and for the sample size calculation with QoL out-
specification are very subjective. The definition comes. This will allow researchers to estimate
of a clinically significant difference is based on more accurately sample sizes for future clinical
such things as experience of the QoL measure of trials involving these three commonly used QoL
interest and the population in which the QoL questionnaires as outcome measures in clinical
measure is being applied. This is not immedi- trials and in clinical practice.
ately straightforward, for different populations
give different distributions for the QoL meas-
ures and different interpretations of what is a
clinically significant effect size. However, there FURTHER RESEARCH
is at present limited experience of QoL meas-
ures in different populations, when compared to Five suggestions for further research are listed
clinical end-points. below:
The specification of the treatment effect size
is also important as one must also be able to put 1 To investigate patients’ change in score over
any results observed into a clinical context. A time. In particular, in many cancer patients,
median reduction of 3 units must have a mean- QoL scores may change quite rapidly, indi-
ing regardless of statistical significance. If, using cating a more restricted QoL and this is
HADS as an example, a median reduction of 3 quickly followed by death. Research could
was observed, the interpretation of this reduc- be carried out to compare these scores with
tion would have a different context if most the scores of those who have survived, and
subjects were classified as ‘normal’ in con- then base the sample size estimate on the
trast to if most subjects were classified as effect size over time. This might improve the
‘depressed’. Patients who have higher scores on sensitivity of QoL questionnaires to detect a
average have a higher potential for reduction. significant treatment difference in longitudi-
Previously published work is an obvious nal studies.
source of information of what is a clinically sig- 2 To quantify the effect of using covariates on
nificant treatment effect. However, one should sample size calculations. The consideration
not fall into the trap of simply setting the effect of covariates in calculating the clinically
size to be what has been observed previously, meaningful difference to be considered may
without concurrent information as to whether it be different in different subgroups of
is meaningful to the patients. patients, and this will need to be corrected
One way of determining a clinically signifi- for in the sample size estimation.
cant difference when there are no randomised 3 The European Guidelines24 state that the
trials available is through elicitation of pos- sample size should consider the possible
sible differences using the opinion of individual number of patients that drop out of the trial,
clinicians with experience of the QoL meas- as it is the number of patients at the end of
ures.29 Thus, one does not have a single estimate each trial which constitutes the sample size.
of an effect size, but a range or distribution of Thus, further factors that need to be con-
possible effects. This would lead to a range of sidered are the take-up rates (usually quite
possible sample sizes. low) and drop-out rates (normally quite
high) in a palliative care study. Rates for
different types of patients and different
scales could be provided from which infla-
IMPLICATIONS FOR THE RESEARCHERS tion factors for the total sample size can be
derived. Another area to consider is what is
It is now almost mandatory to include a sample known as informative missing data. Data
size calculation in research protocols and to may be missing because, for example,
350 The Advanced Handbook of Methods in Evidence Based Healthcare
patients are too ill to complete the form and Treatment of Cancer QLQ-C30: a quality of life
so simply treating these data as missing at instrument for use in international clinical trials
random, and so ignoring patients with miss- in oncology. J. Natl Cancer Inst. 1993; 85:
ing data, will give an unduly optimistic 365–76.
picture of the quality of life of a group. 10. Taylor KM, Macdonald KG, Bezjak A, Ng P,
Similarly, patients who are immobile may DePetrillo AD. Physicians’ perspective on quality
simply ignore questions about pain on mov- of life – An exploratory study of oncologists.
ing, so the fact that the response is missing is Quality of Life Research 1996; 5: 5–14.
related to the condition of the patients and 11. Bleehen NM, Girling DJ, Hopwood P, Lallemand
should not be ignored. G, Machin D, Stephens RJ, Bailey AJ. Medical
4 As previously stated, the frequency of use of Research Council Lung Cancer Working Party:
these QoL instruments in clinical practice in randomised trial of four-drug vs less intensive
the UK by specialist such as oncologists and two-drug chemotherapy in the palliative treatment
physicians is unknown. Clinicians’ percep- of patients with small-cell lung cancer (SCLC)
tions of the usefulness of these instruments and poor prognosis. Br. J. Cancer 1996; 73:
in making decisions about various treatment 406–13.
options for their patients could also be 12. Bleehen NM, Bolger JJ, Girling DJ, Hasleton PS,
investigated. Hopwood P, MacBeth FR, Machin D, Moghissi
5 To investigate the effect on sample size K, Saunders M, Stephens RJ, Thatcher N, White
considerations in other therapeutic areas, the RJ. A Medical-Research Council (MRC) random-
use of different instruments, and especially ized trial of palliative radiotherapy with two frac-
the SF-36 and EuroQoL (EQ-5D), should be tions or a single fraction in patients with inoper-
undertaken. able non-small-cell lung-cancer (NSCLC) and
poor performance status. Br. J. Cancer 1992; 65:
934–41.
13. Bleehan NM, Girling DJ, Machin D, et al. A
REFERENCES randomized trial of three or six courses of etopo-
side, cyclophosphamide, methotrexate and vin-
1. De Haes JCJM, Van Knippenberg FCE. The cristine or six courses of etoposide and ifosfamide
quality of life of cancer patients – a review of the in small-cell lung-cancer (SCLC) 2. Quality-of-
literature. Soc. Sci. Med. 1985; 20: 809–17. life. Br. J. Cancer 1993; 68: 1157–66.
2. Altman DG. Statistics and ethics in medical 14. Whitehead J. Sample size calculations for ordered
research III – How large a sample? Br. Med. J. categorical data. Stat. Med. 1993; 12: 2257–73.
1980; 281: 1336–8. 15. Machin D, Campbell MJ, Fayers P, Pinol A.
3. Edwards SJL, Lilford RJ, Braunholtz D, Jackson Statistical Tables for the Design of Clinical
J. Why ‘underpowered’ trials are not necessarily Studies, 2nd edn. Oxford: Blackwell Science,
unethical. Lancet 1997; 350: 804–7 1997.
4. Julious SA, George S, Campbell MJ. Sample size 16. Campbell MJ, Julious SA, Altman DG. Estimat-
for studies using the short form 36 (SF36). J. ing sample sizes for binary, ordered categorical,
Epidemiol. Community Health 1995; 49: 642–4. and continuous outcomes in two group compar-
5. Campbell MJ, Julious SA, George SL. Estimating isons. Br. Med. J. 1995; 311: 1145–8.
sample sizes for studies using the SF36 (reply to 17. Fayers PM, Machin D. Sample size – how many
letter). J. Epidemiol. Community Health 1996; 50: patients are necessary? Br. J. Cancer 1995; 72:
473–4. 1–9.
6. Julious SA, George S, Machin D, Stephen RJ. 18. Silverstone PH. Poor efficacy of the Hospital
Sample sizes for randomised trials measuring Anxiety and Depression Scale in the diagnosis of
quality of life in cancer patients. Quality of Life major depressive disorder in both medical and
Research 1997; 6: 109–17. psychiatric patients. J. Psychosom. Res. 1994; 38:
7. Zigmond AS, Snaith RP. The Hospital Anxiety 441–50.
and Depression Scale. Acta Psychiatr. Scand. 19. Moorey S, Greer S, Watson M, Gorman C,
1983; 67: 361–70. Rowden L, Tunmore R, Robertson B, Bliss J. The
8. De Haes JCJM, Van Knippenberg FCE, Neijt JP. factor structure of factor stability of the hospital
Measuring psychological and physical distress in anxiety and depression scale in patients with
cancer patients – structure and application of the cancer. Br. J. Psychiatry 1991; 158: 255–9.
Rotterdam Symptom Checklist. Br. J. Cancer 20. Maguire P, Selby P. Assessing quality of life in
1990; 62: 1034–8. cancer patients. Br. J. Cancer 1989; 60: 437–40.
9. Aaronson NK, Ahmedzai S, Bergman B, et al. 21. Neal A, Hoskin P. Clinical Oncology – A textbook
The European Organisation for Research and for students. Edward Arnold, 1994; 616–992.
Quality of Life Measures in Cancer Clinical Trials 351
22. Gardner MJ, Machin D, Campbell MJ. Use of quality of life instrument (PROSQOL1) with
checklists in assessing the statistical content of analogous scales from the EORTC QLQ C30 and
medical studies. Br. Med. J. 1986; 292: 810–12. a trial specific module. J. Clin. Epidemiol. 1998:
23. Senn SJ. Statistical Issues in Drug Development. 51: 137–45.
John Wiley & Sons, 1997. 27. Kiebert GM, Curran D, Aaronson NK. Quality of
24. Lewis JA, Jones DR, Rohmel J. Biostatistical life as an endpoint in EORTC clinical trials. Stat.
methodology in clinical trials – a European guide- Med. 1998; 17: 561–70.
line. Stat. Med. 1995; 14: 1655–82. 28. Stucki G, Daltroy L, Katz JN, Johannesson M,
25. Guyatt GH, Juniper EF, Walter SD, Griffith LE, Liang MH. Interpretation of change scores in
Goldstein RS. Interpreting treatment effects in ordinal clinical scales and health status measures:
randomised trials. Br. Med. J. 1998; 316: 690–3. the whole may not equal the sum of the parts. J.
26. Stockler MR, Osoba D, Goodwin P, Corey P, Clin. Epidemiol. 1996; 49: 711–17.
Tannock IF. Responsiveness to change in health 29. Spiegelhalter DJ, Freedman LS, Parmar MK.
related quality of life in a randomised clinical Bayesian approaches to randomized trials. J. Roy.
trial: a comparison of the prostate cancer specific Statist. Soc. A 1994; 157: 357–416.
20
Simultaneous Analysis of Quality of
Life and Survival Data
chronic diseases. Although length of survival is a hand-search of the journal Quality of Life
usually the standard end-point for assessing Research.
treatments in such clinical trials, informed clin- The search produced 1127 references in total,
ical decisions generally require quantification of which 361 were considered to be relevant and
and interpretation of quality of life measures, were included in the review.7 The literature
especially with respect to variation over time search aimed to identify all methods proposed
and the inter-relation with length of life. The for the analysis of quality of life and survival
role of quality of life will become even more data. It is not possible to determine if this aim
prominent in the future as, for many diseases, has been achieved, but the broadness of the
especially cancer, improvements in survival due search strategy used, together with the method
to treatment are either unlikely to be dramatic of exploding references, should have ensured a
or likely to be made at the expense of quality reasonably complete coverage of material in the
of life. field.
There has been much research into the devel-
opment of instruments with which to measure
quality of life. This has resulted in a plethora of Illustrative Example: The MIC Study
instruments1 and in substantial amounts of qual-
ity of life data being gathered in trials. Conse-
quently, methods that enable the effective and The methods identified by the literature search
efficient analysis and interpretation of such data are illustrated by application to data from a
are needed. There are a number of previous previously conducted study. The study is the
reviews of the assessment and analysis of qual- second of two concurrent Phase III randomised
ity of life in clinical trials.2–6 trials of MIC (mitomycin, ifosfamide and cis-
In quality of life studies where survival is also platin) chemotherapy in patients with non-small-
an end-point, the patient population will not be cell lung cancer that were conducted at the
stable over time. Patients are often severely ill, Cancer Research Campaign Trials Unit at Bir-
and subjects may have incomplete follow-up of mingham University. The results from both
quality of life for reasons related to disease or trials are reported elsewhere,8 and the worked
treatment, including death. This drop-out pro- examples here use just the quality of life and
cess may be related to the quality of life being survival data from the second trial (MIC2),
experienced, rather than being at random, and is referred to here as the MIC study.
known as informative drop-out. In such situ- Patients in the MIC study were randomised to
ations, statistical analysis can be particularly receive either standard palliative care (PAL),
problematic and this chapter reviews appropriate usually radiotherapy, or MIC chemotherapy to a
methods that yield unbiased and clinically rele- maximum of four courses followed by palliative
vant assessment of health technologies. care (CT). The aim of the study was to compare
treatments in terms of both length of survival
and quality of life. The trial started in 1988 and
METHODS OF REVIEW closed in 1996, during which time 359 patients
were randomised into the study. For practical
Literature Search reasons, quality of life was assessed in a sub-
group of 109 patients from three centres, with
The scientific and medical literature were 67 on the CT arm and 42 on the PAL arm.
searched for relevant methodological articles. Quality of life was measured using a patient-
An article was defined as relevant if it included
completed questionnaire consisting essentially
some sort of quality of life assessment over
time, in circumstances where survival was also of 12 questions on various aspects of quality
an issue, and it used either a known method- of life. Responses were measured on a four-level
ology of interest or a new and clearly detailed ordered categorical scale: ‘none’, ‘a little’,
methodology. ‘quite a bit’ or ‘very much’.
Electronic searches were carried out system- A single measure of quality of life is needed
atically using the Science Citation Index, the to illustrate the methods, and this may be as
Social Science Citation Index and the EMBASE simple or as complex as required to reflect the
database provided by BIDS (Bath Information quality of life of the patients. A malaise ques-
and Data Service). Reference lists from articles tion, which asked if patients had been feeling
defined as being relevant were ‘exploded’ such ‘generally ill’, is chosen as the measure of
that they were checked to identify further quality of life to use in the analysis here. Due to
relevant literature. Articles recommended by the lack of data, the level of malaise, i.e. ‘a
colleagues and from personal collections were little’, ‘quite a bit’ or ‘very much’ has to be
also incorporated. These were supplemented by ignored, resulting in a binary quality of life
354 The Advanced Handbook of Methods in Evidence Based Healthcare
outcome such that a patient either does or does METHODS FOR THE ANALYSIS OF QUALITY OF
not have malaise. Although this measure is very LIFE AND SURVIVAL DATA
simplistic, it nonetheless illustrates the essentials
of the methods and allows insight into the prob-
lems associated with the methods. Methods for analysing quality of life and sur-
Quality of life was assessed only during vival data were found to fall into three broad
the patient’s treatment period. The study was categories according to the research question
designed such that patients on the CT arm underlying the study, which in turn depends on
completed five questionnaires at approximately the disease and treatments under investigation.
3-weekly intervals (one before each course of The first category consists of methods where the
chemotherapy and one post treatment), whilst primary aim of the study is to compare treat-
patients on the RT arm completed four ques- ments in terms of quality of life whilst account-
tionnaires at approximately 3-weekly intervals. ing for informative drop-out. The second con-
There was some variation in the timing, but 99% sists of methods where the primary aim is to
of questionnaires had been completed by 18 compare treatments in terms of survival whilst
weeks and the 18-week period from entry to trial adjusting for the effects of quality of life. The
is defined as the analysis period for the study. third consists of methods for studies where qual-
Patients were severely ill and hence there were a ity of life and survival are both important end-
considerable number of ‘early’ deaths and drop- points for assessing treatments, and the two end-
outs from the quality of life study. Thirty-seven points are to be analysed simultaneously.
patients died within 18 weeks of entry to the If treatment comparison is primarily in terms
trial, 29 of whom died before completing the of quality of life, then standard methods of
full set of assessments. In total, 46 patients longitudinal data analysis9–11 can be applied.
dropped out of the quality of life study. Ten of The application of such methods to quality of
these drop-outs could be attributed directly to life data has been discussed,12–17 but in situations
death in that they died within 3 weeks of the last where informative drop-out is present, they will
completed assessment. The drop-out problem in give biased results since missing data are
the MIC study makes it a realistic example with assumed to be missing at random. Missing data
which to illustrate methods for the assessment of can be imputed using techniques such as ‘last
quality of life and survival data. value carried forward’,18 but this could also
result in bias. Modelling techniques that deal
Assumptions required for analysis of the with informative drop-out have been devel-
oped.19,20 The methods are not readily accessible
MIC data to researchers and, with examples of their appli-
In the MIC study, the survival time within the cation to quality of life data being limited,
18-week analysis period for each patient was further research into their application in this
split into time with malaise and time without field is required.
malaise according to the responses to the ques- If the aim of the analysis is to compare length
tionnaire at each quality of life assessment. of survival for different treatments then it may
Various assumptions are required to categorise be necessary to incorporate changing quality of
patients into health states at all points of their life values over time as a time-dependent co-
survival time using the assessments made at variate. This may be done using standard sur-
discrete time points. If a patient’s state of vival analysis techniques such as Cox regression
malaise changed between two assessments, then modelling.21,22 If assessments of quality of life
the change was assumed to happen at the mid- are infrequent, or data are missing for reasons
point between assessment dates and subjects other than death, then it may be difficult to
were assumed to remain in a specified health adjust for changing quality of life with any
state until a change occurred. The state of degree of accuracy.
malaise that was reported at the last assessment Methods that simultaneously analyse quality
was assumed to carry over until either death or of life and survival data not only provide a
18 weeks, whichever came first. This may not be means of comparing treatments in terms of the
a reasonable assumption for patients who drop- two end-points simultaneously, but also provide
ped out of the study, especially for those who a means for overcoming problems of missing
dropped out early and then remained alive for quality of life data, resulting from informative
the whole of the 18-week period. Sensitivity drop-out, when analysing either quality of life
analysis should be used to examine the robust- or survival as the primary end-point. For this
ness of any conclusions with respect to the reason, these methods are the focus of this
assumptions made. chapter.
Analysis of Quality of Life and Survival Data 355
Three different approaches can be used to was originally developed for a breast cancer
analyse quality of life and survival data simul- application,27 but has since been modified where
taneously. The first approach considered is necessary for use with other diseases such as
quality-adjusted survival analysis wherein treat- rectal cancer29 and AIDS.30 With Q-TWiST,
ments are compared in terms of a composite periods of survival time spent with symptoms of
measure of quality and quantity of life, created disease and toxicity resulting from treatment are
by weighting periods of survival time according each given weights between 0 and 1, rather than
to the quality of life experienced. A second being ignored as they are in TWiST.
approach is to use multi-state models to describe There are two main approaches to quality-
the movement of patients between various adjusted survival analysis, depending on the
health states, defined by levels of quality of life level of aggregation of the quality of life and
and death, and explore how treatments differ in survival data. The subject-based approach com-
terms of these transitions. Finally, a simultan- bines quality of life and survival at the patient
eous modelling approach considers quality of level, thus creating a single end-point for each
life and survival as two simultaneous processes subject on which to compare treatments, whilst
occurring in patients and describes the data in the population-based approach aggregates qual-
terms of two inter-linked models. These meth- ity of life and survival at a (predefined) group
ods will be discussed, and the first two will be level. Both approaches will be discussed and
illustrated using data from the MIC study. illustrated using data from the MIC study.
terms of the study of survival time. It should be relates to completeness of follow-up in terms of
noted that the follow-up in terms of these two quality of life data.
end-points could be different (see Figure 1). In In calculating QALYs for each subject, there
studies of survival time, subjects may either be may be individuals who have censored survival
lost-to-follow-up during the course of the study, times and therefore the number of QALYs for
or may be followed-up for the period of the these individuals will also be censored. It there-
study, but not die. For these patients, although fore seems appropriate to apply standard sur-
the actual survival time is not known, the data vival analysis techniques to this quality-adjusted
provide a lower bound for the value. These are survival time, as for a standard survival end-
known as censored survival times and standard point. This approach, however, may be invalid
survival analysis techniques are specifically and give biased results because the censoring
designed to account for any censored data. The mechanism is related to the QALY end-point,
methods, however, are only valid if the mechan- resulting in informative censoring.23 For ex-
ism that causes the individual’s survival time to ample, if two patients both have the same cen-
be censored is unrelated to the actual unob- sored survival time (say 2 years) and one patient
served survival time. When the censoring mech- spends their survival time in poor quality of life
anism is not independent of survival time, (valued at say 0.5), whilst the other has good
informative censoring occurs and standard meth- quality of life (valued at say 1), then the QALY
ods for survival analysis are invalid.22 It should be time for the first individual will be censored
noted that informative censoring is a different earlier (i.e. 1 year) than the second individual (2
concept to the previously discussed concept of years), purely because of their quality of life.
informative drop-out (see Background), which One method for overcoming this problem is to
Analysis of Quality of Life and Survival Data 357
restrict the analysis to a time period for which 18-week survival to be statistically significant
all patients have an observed survival time. (x2 5 8.07, P 5 0.005).
Then, as there are no censored survival times, Kaplan–Meier survival curves using the
informative censoring will not be a problem and QALW end-point rather than survival are shown
standard survival techniques can be used. If in Figure 3. At any time point the curves rep-
there are a number of survival times censored at resent the proportion of patients having at least
an early time point, then this approach may not that quality-adjusted survival time. It should be
be practical. An alternative population-based noted that the steps do not represent deaths as
approach such as partitioned survival analysis,23 they do with normal survival curves, but rather
which will be discussed later, may be used to termination of a patient’s QALW at that point.
overcome the problem. As the number of QALWs increases, so the
In the MIC study, because the analysis was proportion having at least that value decreases.
restricted to 18 weeks from study entry and all The curves for the different treatments are
patients have complete follow-up during that similar, with the PAL curve decreasing more
time in terms of survival, there was no problem rapidly than that for CT. The areas under the
of informative censoring, and standard survival curves give the mean quality-adjusted life weeks
analysis techniques could be used. accumulated over 18 weeks which was 14.4
Before presenting quality-adjusted survival weeks (standard error 0.46) for CT and 13.0
times, the results of analysing the standard sur- weeks (standard error 0.70) for PAL, indicating
vival end-point should be noted. A standard better quality-adjusted survival with chemo-
analysis of the survival of patients in the MIC therapy. A log-rank test showed no statistically
quality of life study within 18 weeks of trial significant difference between treatment arms
entry showed that chemotherapy increased the (x2 5 1.33, P 5 0.25). Thus, although treatments
length of survival compared to standard palli- differed in terms of survival within 18 weeks,
ative care. The Kaplan–Meier survival curves they were not found to differ significantly in
for the two treatment groups are shown in Fig- terms of quality-adjusted survival. This conclu-
ure 2. The areas under the curves give the mean sion is based on weighting periods of malaise
survival time accumulated over the 18 weeks, with the value of 0.8, and sensitivity analysis
which was 16.2 weeks (standard error 0.48) for should be used to establish the robustness of
CT and 14.4 weeks (standard error 0.74) for conclusions to the choice of weights used to
PAL. A log-rank test showed this difference in calculate the QALW end-point for each subject.
Figure 2 Survival within 18 weeks for patients in the MIC quality of life study.
358 The Advanced Handbook of Methods in Evidence Based Healthcare
Kaplan–Meier survival curves for each treat- period was with malaise and hence have a
ment group using the TWiST end-point are TWiST of 0. The curves show no difference
shown in Figure 4. The low starting point of the between treatments. The areas under the curves
curves represents the proportion of patients gives the mean TWiST accumulated over 18
whose whole survival time during the 18-week weeks which was 6.6 weeks (standard error
Analysis of Quality of Life and Survival Data 359
0.81) for chemotherapy and 6.5 weeks (standard life assessments. Time spent in the health states
error 1.06) for palliative treatment. A log-rank (ti where i is a health state) were combined using
test showed no statistically significant difference a QALY model as defined in Equation (1) where
(x2 5 0.04, P 5 0.85). both good states were allocated weights of 1.
The treatments significantly differ in terms of The two poor states were allocated different
survival, but not in terms of quality-adjusted weights, u1 and u2 (0 # u1 # 1, 0 # u2 # 1), on
survival, possibly because the extra survival the assumption that the quality of life differs
time obtained under chemotherapy is generally depending on whether it is the first or second
spent in poor quality of life. The extra variabil- time in that state.
ity in the quality-adjusted survival data may also
have some impact in changing the significance QALW 5 tGOOD11u1tPOOR11tGOOD21u2tPOOR2 (1)
of the results. The partitioned survival curves for the chemo-
therapy treatment group are shown in Figure
Partitioned survival analysis illustrated 5(a). The curves are standard Kaplan–Meier
with data from the MIC study survival curves with not only time to death
represented, but also time to exiting each of the
It has already been noted that standard survival progressive health states. This partitions the area
analysis of a QALY end-point, calculated for
under the overall survival curve into areas rep-
each individual, may lead to a problem of inform-
resenting time spent in each state. The area
ative censoring. Population-based approaches,
between curves is calculated and represents the
which combine quality of life and survival data
at the population level rather than at the subject mean time spent in the health states, restricted to
level, can overcome the problem of informative the 18-week analysis period. These values can
censoring. This is because survival analysis is be put into the model as specified in Equation
carried out on unweighted survival times and the (1) to give the restricted mean quality-adjusted
quality of life adjustments are applied at a later survival time for the chemotherapy treatment
stage to summary measures of survival time. group as specified in Equation (2). The pro-
Various population-based approaches have been cess is repeated for the palliative group to give
discussed,29,31 but partitioned survival analy- Equation (3).
sis23,28 is the most widely used. In this approach, QALWCT 5 3.37 1 8.10u1 1 3.16 1 1.66u2 (2)
the time spent in each of a finite number of
progressive health states is calculated as a mean QALWPAL 5 4.22 1 7.96u1 1 2.15 1 0.19u2 (3)
for a group rather than on a patient-by-patient
At this point of the analysis there are two main
basis. A weighted sum of these mean times
estimates the quality-adjusted survival time for options. If some knowledge exists regarding the
the group. The method was devised and is values that u1 and u2 should take, then those
generally used for a Q-TWiST end-point with values could be used to calculate the restricted
three progressive health states: time spent with mean quality-adjusted survival time for each
toxicity (TOX); TWiST; and time following treatment group; otherwise a special form of
relapse of disease (REL). Quality of life data do sensitivity analysis called a threshold utility
not generally create such health states, but the analysis may be carried out.
methodology may still be applicable by defining In a threshold utility analysis, all possible
health states based on the quality of life out- pairs of values for the weights are considered,
comes. This approach is illustrated by applica- and values that give no difference between treat-
tion to the MIC data. ments in terms of restricted mean quality-
The first step in the method is to define a adjusted survival time are determined. Weights,
set of progressive health states that the patient u1 and u2, can take any value between 0 and 1.
passes through during their survival time. This is The shaded area in Figure 6 shows all possible
not always straightforward or possible. In the pairs of values that u1 and u2 can take, known as
MIC data, patients appear to fluctuate from one the threshold utility plane. The line in Figure 6,
state to another in no obvious pattern, and it is known as the threshold line, represents pairs of
difficult to define progressive health states that values for which the restricted mean quality-
are clinically meaningful. For illustrative pur- adjusted survival times in the two treatment
poses we chose a set of four progressive health groups are equal. This is calculated by putting
states: a good state (GOOD1) followed by a QALWCT, as specified in Equation (2), equal to
poor state (POOR1) followed by a second good QALWPAL, as specified in Equation (3). The area
state (GOOD2) and a further poor state above the line represents pairs of values for
(POOR2). Good and poor health states were which quality-adjusted survival time is greater
defined as time without and with malaise for CT than PAL, and below the line where
respectively, as determined from the quality of quality-adjusted survival time is less for CT
360 The Advanced Handbook of Methods in Evidence Based Healthcare
compared to PAL. Figure 6 shows that all pos- data. It can be used with any quality of life instru-
sible pairs of values of u1 and u2 fall above ment, providing it is possible to define mean-
the line, so that whatever values of u1 and u2 ingful health states from it. In some situations,
are used in the model, the amount of quality- it may be desirable to carry out separate analy-
adjusted survival time will be greater under ses for each dimension in a multi-dimensional
chemotherapy than under palliative treatment in instrument.
this example. The subject-based approach may suffer from
the problem of informative censoring. This may
Critical appraisal of quality-adjusted be overcome by either restricting the analysis to
an upper time limit, which may occur automatic-
survival analysis ally anyway if the quality of life data collection
Quality-adjusted survival analysis provides a is restricted to a limited time period, or using a
relatively straightforward approach for the simul- population-based approach such as partitioned
taneous analysis of quality of life and survival survival analysis.
Analysis of Quality of Life and Survival Data 361
of model was used for the MIC study (see assuming the transition times follow a specific
Figure 7) with patients categorised as being in distribution with a parametric form for the base-
an alive and well state if they had no malaise line transition rate. The most commonly used
and in an alive and ill state if they had malaise distributions are the exponential and the Weibull
at any level. Patients can move between the distributions, the exponential simply being a
alive states any number of times during their special form of Weibull distribution. If an expo-
follow-up until they finally move to the absorb- nential is assumed, then the underlying baseline
ing state of death. As one would expect, the 37 transition rate is taken as constant; otherwise for
deaths that occurred within the 18-week analysis a Weibull distribution the underlying baseline
period came mostly from the alive and ill state. transition rate is allowed to change over time.
The movement between health states is
described by transition rates. A transition rate is Modelling transition rates in the
the instantaneous potential of transition at any
point in time, and is equivalent to the standard MIC study
hazard rate function for a survival time distribu- To model the transition rates in the three-state
tion. The transition rates in a multi-state model illness–death model for the MIC study (see
can be represented by Cox regression models,41 Figure 7), exact times of transition from one
and are modelled using the transition times for state to another within the 18-week analysis
patients. If exact transition times are not avail- period were needed for each patient. The exact
able then they need to be estimated for this type dates of transition to death were known, but
of analysis. dates of transition between alive states had to be
The transition rate from state i to state j at estimated. Estimation was based on the same
time t, lij(t), can be modelled using a Cox assumptions that were used in the quality-
regression model as follows: adjusted survival analysis, in particular that
lij(t) 5 l0ij(t) exp(b ijT x ij) (4) health state changes occur midway between
assessments (see previous section on Assump-
where l0ij(t) is a baseline transition rate for the tions required for analysis of the MIC data).
transition from i to j, x ij is a vector of covariates The illustrative results presented here (see
specific to that transition and b ij is a vector of Figure 8) relate to a semi-parametric model
unknown regression coefficients. with an underlying semi-Markov process. Other
In modelling the transition rate from state i to models are presented in the full project report.7
state j, simplifying assumptions that may be The PHREG procedure in SAS42 was used to
made include: (i) a Markov process; or (ii) a analyse the data. Figure 8 shows, for each transi-
semi-Markov process. In a Markov process the tion, the ratio of the transition rate for CT
transition rate from state i to state j is dependent compared to the transition rate for PAL. Transi-
only upon the present state occupied; in a semi- tion ratios greater than 1 indicate that the
Markov process the transition rate is also transition rate on CT is greater than that on
dependent upon the duration of time in the PAL. The point estimates suggest that the haz-
present state. ard of moving between the ‘alive’ states is
The transition rates can either be modelled greater on CT compared to PAL, whilst the
semi-parametrically using a Cox regression hazard of death from either of the ‘alive’ states
model, where the underlying baseline transition is reduced with CT compared to PAL. This
rate is left unspecified, or parametrically by suggests that chemotherapy leads to greater
Analysis of Quality of Life and Survival Data 363
Figure 8 Transition ratios (and 95% confidence intervals) for the MIC study
assuming semi-Markov, semi-parametric models for each transition rate.
fluctuations in quality of life whilst alive, but deal with drop-out for other reasons by includ-
reduces the hazard of death. However, the 95% ing a ‘drop-out’ health state. This approach
confidence intervals for the transition ratios allows the drop-out process to be explored.
(shown in brackets) are wide.
Multi-state survival analysis allows the effect
of treatments on different health state transitions Simultaneous Modelling
to be explored. If a single overall statement of
the superior treatment is required explicitly then Simultaneous modelling is a developing area of
some sort of trade-off between these results methodology that currently has not been applied
would be needed. in a quality of life context, but potentially could
provide a powerful approach to analysing qual-
ity of life and survival data. The change in
Critical appraisal of multi-state survival quality of life over time and the time to death
analysis can be considered as two simultaneous processes
Multi-state survival analysis provides a flexible occurring in patients, and can be modelled as
approach for the simultaneous assessment of such. Models are set up for each process, and
quality of life and survival data, and a greater Gibbs sampling may be used to fit the models
biological insight may be gained, though the simultaneously in a single analysis.44,45
method can be problematic. Simultaneous modelling allows survival data
As with quality-adjusted survival analysis, the to be incorporated into the model for quality of
method – in theory – can be used with any life data, thus adjusting for informative missing
quality of life instrument, providing they can be data caused by drop-out due to death. At the
same time, the modelling of survival data with
used to define meaningful health states. For
quality of life as a time-dependent co-variate is
multi-dimensional instruments, separate analy-
enhanced since co-variate values are estimated
ses may be desirable. Defining health states
from the model for quality of life data over time
using quality of life data is subjective, and the
fitted to all subjects. The approach can be gen-
effects of using different definitions should eralised to model the drop-out process rather
be considered as part of a sensitivity analysis. than the survival process and in this way over-
Definitions will differ depending on the quality come the problem of informative drop-out for
of life variable to be used, the number of health reasons other than death.46 It may be possible to
states to be included, and the cut-off values used extend the method to include models of a hier-
to discriminate between health states. archical nature that would account for the multi-
Another difficulty with the method is the need dimensionality of quality of life data.
for exact dates of transition between health
states. These can be estimated and the accuracy
is determined by the frequency of the quality of
life assessments. Alternative methods that do IMPLICATIONS FOR PRACTITIONERS AND
not require exact transition dates are available,43 RESEARCHERS
but these methods require specialised software
that is not readily accessible to researchers. Study Design
One of the main advantages of the method is
that provides a means of dealing with the prob- The method of analysis needs to be decided
lem of informative drop-out. The inclusion of upon at the design stage of a study so that
death as a health state in the model deals with appropriate quality of life data can be collected.
drop-out due to death and it may be possible to The following issues should be considered.
364 The Advanced Handbook of Methods in Evidence Based Healthcare
1 Most methods of analysis are based on a Further research is needed to develop multi-
single measure of quality of life. Using a state models, simultaneous modelling methods
multi-dimensional instrument will result in and hierarchical models in their practical appli-
multiple measures of quality of life each cation to quality of life and survival data, using
potentially requiring a separate analysis. both classical and Bayesian approaches. Con-
Selecting instruments that produce a global sideration should be given to how methods
quality of life score or identifying specific could deal with the multivariate nature of the
dimensions at the design stage for hypothesis quality of life end-point.
testing from a multi-dimensional instrument A full review of available software for meth-
will reduce problems of multiple testing. ods that simultaneously assess quality of life and
2 Quality of life assessments should be fre- survival data is needed to highlight areas requir-
quent enough to ensure accurate estimation ing further development.
of dates of health state transitions.
3 Assessments of quality of life should be
planned such as to minimise non-compliance
and thus reduce the problem of missing ACKNOWLEDGEMENTS
data.
4 Recording the reason for drop-out during
data collection will maximise insight into This research was funded under a National
the drop-out process and enable modelling Health Service Health Technology Assessment
of separate drop-out processes where Methodology Programme grant. We thank the
necessary. Cancer Research Campaign Trials Unit at
5 The number of subjects required for a study Birmingham University for supporting the sec-
will depend on the method of analysis. ondment of Lucinda Billingham to Leicester
Multi-state survival analysis, for example, University for the duration of the project. We
needs a relatively large overall sample size, are grateful to Dr Michael Cullen, Consultant
so that there are sufficient subjects in each and Honorary Reader in Medical Oncology at
transition for adequate modelling of the the Queen Elizabeth Hospital in Birmingham,
data. for allowing us to use the MIC data as the
illustrative example in this chapter. Finally, we
Choosing the Appropriate Method would like to thank the editors for their useful
comments.
The choice of method should be based on the
research question that the study aims to answer.
The advantages and disadvantages of each REFERENCES
method7 should be considered carefully together
with the relevance and interpretability of the
results to clinicians and patients. 1. Bowling A. Measuring health: a review of quality
Methods used to analyse longitudinal quality of life measurement scales. Open University
of life data must consider the problem of Press, 1991.
informative drop-out. 2. Cox DR, Fitzpatrick R, Fletcher AE, Gore SM,
Spiegelhalter DJ, Jones DR. Quality-of-life
Reporting the Analysis assessment: can we keep it simple? J. Roy. Statist.
Soc. A 1992; 155: 353–93.
Methods used should be reported clearly, with 3. Hopwood P, Stephens RJ, Machin D. Approaches
details of definitions and assumptions used in to the analysis of quality of life data: experiences
the analysis. gained from a Medical Research Council Lung
Sensitivity analysis should be carried out to Cancer Working Party palliative chemotherapy
assess the robustness of conclusions to any trial. Quality of Life Research 1994; 3: 339–52.
critical assumptions made in the analysis. 4. Fletcher A, Gore SM, Jones DR, Fitzpatrick R,
Spiegelhalter DJ, Cox DR. Quality of life meas-
ures in health care. II: Design, analysis and inter-
SUGGESTIONS FOR FUTURE RESEARCH pretation. Br. Med. J. 1992; 305: 1145–8.
5. Schumacher M, Olschewski M, Schulgen G.
Further experience in the application of quality- Assessment of quality-of-life in clinical trials.
adjusted survival analysis techniques specific- Stat. Med. 1991; 10: 1915–30.
ally to quality of life data, rather than clinical 6. Olschewski M, Schumacher M. Statistical analy-
data, is needed to enable a proper evaluation of sis of quality of life data in cancer clinical trials.
such methods in this field. Stat. Med. 1990; 9: 749–63.
Analysis of Quality of Life and Survival Data 365
7. Billingham LJ, Abrams KR, Jones DR. Methods of-life-oriented endpoint for comparing therapies.
for the analysis of quality of life and survival data Biometrics 1989; 45: 781–95.
in health technology assessment. Health Technol- 26. Gelber RD, Goldhirsch A. A new endpoint for the
ogy Assessment 1999; 3(10). assessment of adjuvant therapy in postmeno-
8. Cullen MH, Billingham LJ, Woodroffe CM, et al. pausal women with operable breast cancer. J.
Mitomycin, ifosfamide and cisplatin in unresect- Clin. Oncol. 1986; 4: 1772–9.
able non-small cell lung cancer: effects on sur- 27. Goldhirsch A, Gelber RD, Simes RJ, Glasziou P,
vival and quality of life. J. Clin. Oncol. 1999; 17: Coates AS. Costs and benefits of adjuvant therapy
3188–94. in breast cancer: a quality-adjusted survival anal-
9. Matthews JNS, Altman DG, Campbell MJ, Roy- ysis. J. Clin. Oncol. 1989; 7: 36–44.
ston P. Analysis of serial measurements in medical 28. Gelber RD, Cole BF, Gelber S, Goldhirsch A.
research. Br. Med. J. 1990; 300: 230–5. Comparing treatments using quality-adjusted sur-
10. Diggle PJ, Liang K, Zeger SL. Analysis of vival: the Q-TWiST method. Am. Statistician
longitudinal data. Oxford: Clarendon Press, 1994. 1995; 49: 161–9.
11. Goldstein H. Multilevel statistical models. 2nd 29. Gelber RD, Goldhirsch A, Cole BF, Weiand HS,
edn. London: Edward Arnold, 1995. Schroeder G, Krook JE. A quality-adjusted time
12. Korn EL. On estimating the distribution function without symptoms or toxicity (Q-TWiST) analy-
for quality of life in cancer clinical trials. Bio- sis of adjuvant radiation therapy and chemother-
metrika 1993; 80: 535–42. apy for resectable rectal cancer. J. Natl Cancer
13. Hollen PJ, Gralla RJ, Cox C, Eberly SW, Kris Inst. 1996; 88: 1039–45.
MG. A dilemma in analysis: issues in the serial 30. Gelber RD, Lenderking WR, Cotton DJ, Cole BF,
measurement of quality of life in patients with Fischl MA, Goldhirsch A, Testa MA. Quality-of-
advanced lung cancer. Lung Cancer 1997; 18: life evaluation in a clinical trial of zidovudine
119–36. therapy in patients with mildly symptomatic HIV
14. Zwinderman AH. Statistical analysis of longitudi- infection. Ann. Intern. Med. 1992; 116: 961–6.
nal quality of life data with missing measurements. 31. Hwang JS, Tsauo JY, Wang JD. Estimation of
Quality of Life Research 1992; 1: 219–24. expected quality adjusted survival by cross-
15. Beacon HJ, Thompson S. Multi-level models for sectional survey. Stat. Med. 1996; 15: 93–102.
repeated measurement data: application to quality 32. Gelber RD, Goldhirsch A, Cole BF. Parametric
of life data in clinical trials. Stat. Med. 1996; 15: extrapolation of survival estimates with applica-
2717–32. tions to quality of life evaluation of treatments.
16. Beacon HJ. The statistical analysis of self Controlled Clinical Trials 1993; 14: 485–99.
assessed quality of life data in cancer clinical 33. Hinkley DV. Bootstrap methods. J. Roy. Statist.
trials. PhD Thesis. London: London School of Soc. B 1988; 50: 321–37.
Hygiene and Tropical Medicine, University of 34. Cole BF, Gelber RD, Goldhirsch A. Cox regres-
London, 1996. sion models for quality adjusted survival analysis.
17. Zwinderman AH. The measurement of change of Stat. Med. 1993; 12: 975–87.
quality of life in clinical trials. Stat. Med. 1990; 9: 35. Cole BF, Gelber RD, Anderson KM. Parametric
931–42. approaches to quality-adjusted survival analysis.
18. Little RJA, Rubin DB. Statistical analysis with Biometrics 1994; 50: 621–31.
missing data. John Wiley & Sons, 1987. 36. Kay R. Multistate survival analysis: an applica-
19. Little RJA. Modelling the drop-out mechanism in tion in breast cancer. Methods Inf. Med. 1984; 23:
repeated measures studies. J. Am. Statist. Assoc. 157–62.
1995; 90: 1112–21. 37. Hansen BE, Thorogood J, Hermans J, Ploeg RJ,
20. De Stavola BL, Christensen E. Multilevel models van Bockel JH, van Houwelingen JC. Multistate
for longitudinal variables prognostic for survival. modeling of liver transplantation data. Stat. Med.
Lifetime Data Analysis 1996; 2: 329–47. 1994; 13: 2517–29.
21. Cox DR. Regression models and life tables. J. 38. Andersen PK. Multistate models in survival
Roy. Statist. Soc. B 1972; 34: 187–220. analysis: a study of nephropathy and mortality in
22. Collett D. Modelling survival data in medical diabetes. Stat. Med. 1988; 7: 661–70.
research. London: Chapman & Hall, 1994. 39. Marshall G, Jones RH. Multistate models and dia-
23. Glasziou PP, Simes RJ, Gelber RD. Quality betic retinopathy. Stat. Med. 1995; 14: 1975–83.
adjusted survival analysis. Stat. Med. 1990; 9: 40. Gentleman RC, Lawless JF, Lindsey JC, Yan P.
1259–76. Multistate Markov models for analyzing incom-
24. Torrance GW. Utility approach to measuring plete disease history data with illustrations for
health-related quality of life. J. Chron. Dis. 1987; HIV disease. Stat. Med. 1994; 13: 805–21.
40: 593–600. 41. Kay R. The analysis of transition times in multi-
25. Gelber RD, Gelman RS, Goldhirsch A. A quality- state stochastic processes using proportional
366 The Advanced Handbook of Methods in Evidence Based Healthcare
hazard regression models. Communications in ling censored survival data and repeatedly meas-
Statistics – Theory and Methodology 1982; 11: ured covariates: a Gibbs sampling approach. Stat.
1743–56. Med. 1996; 15: 1663–85.
42. SAS Institute Inc. SAS technical report P–229, 45. Berzuini C. Medical monitoring. In: Gilks WR,
SAS/STAT software: changes and enhancements; Richardson S, Spiegelhalter DJ (eds). Markov
Release 6.07. Cary, NC: SAS Institute Inc., Chain Monte Carlo in practice. London: Chap-
1992. man & Hall, 1995; Chapter 18.
43. Kay R. A Markov model for analysing cancer 46. Lindsey JK. Modelling longitudinal measurement
markers and disease states in survival studies. dropouts as a survival process with time-varying
Biometrics 1986; 42: 855–65. covariates. Technical Report, Limburgs Universi-
44. Faucett CL, Thomas DC. Simultaneously model- tair Centrum, Belgium, 1997.
Part V
CONSENSUS, REVIEWS
AND META-ANALYSIS
I N T R O D U C T I O N by A N D R E W S T E V E N S
and K E I T H R . A B R A M S
The research evidence both informing health data needed for synthesised evidence will not
technology assessment (HTA), and consequent always be present. Chapter 24, on consensus as
on it, can be seen as being at three distances a basis for guidelines, reflects both of these
from the end user. Distance 1, primary research points.
(well-designed randomised trials and observa- It is worth rehearsing some of the key features
tional studies) form a core means of providing of the systematic review revolution that have
data for HTA. Primary data collection has been generated their wide usage and continuing meth-
dealt with extensively elsewhere in this book. odological interest. First, systematic reviewing
However, at the heart of HTA is the logic of means systematicity of search. Precise rules for
the need for systematic reviews, i.e. distance 2. how far to search to produce a systematic review
The case for systematic reviews has been estab- have not been universally agreed. The elements
lished beyond all reasonable doubt over the past of searching can include not just the use of the
decade.1,2 Indeed, Chapters 21–23 in this section many electronic databases (e.g. the Cochrane
all amply illustrate that we have moved well Library, Medline, CINAHL, Psychlit, Science
past exhortations for systematicity in reviewing, Citation Index), but also hand-searching the
to the scientific requirements of ensuring that relevant journals and identifying unpublished
systematicity is guaranteed. All three of these literature and conference proceedings.3 The
chapters explore methodological frontiers in principle underlying the systematicity of search-
systematic reviews. Distance 3 (closest to the ing is the avoidance of bias. The absence of a
end user) includes guidelines, syntheses of the systematic search can be a sign of a review
evidence on multiple interventions and struc- setting out to prove a point, rather than to
tured reviews aimed at end users. This group of appraise the evidence – a common feature of
‘products’ has partly been necessitated by the traditional reviews and editorials. Even when
increasing rigour used to synthesise evidence reviews have an unbiased intent, a systematic
within systematic reviews. The scientific syn- and comprehensive search is necessary because
thesis of evidence has the tendency both to of the bias that can arise when studies with
narrow the topics and to increase the ‘weight’ of positive or dramatic findings are more likely to
reviews, making them difficult to access in be published than those with negative or un-
day-to-day planning, either at the clinical or at exciting results. This ‘publication bias’ may be
the health authority/insurance planning level. It more of a problem with smaller studies. In
is, however, also recognised that the primary Chapter 21, Song and colleagues analyse these
368 The Advanced Handbook of Methods in Evidence Based Healthcare
issues and review a variety of methods for whether the search was sensitive and specific,
detecting and adjusting for publication bias in whether there was a quality assessment of the
systematic reviews, which they suggest should included studies, and whether the combination
be used critically in any systematic review or of the studies was undertaken reasonably. In
meta-analysis. They also stress that sensitivity Chapter 23, Moher and colleagues take such
analyses should cover both the effect of includ- considerations further, questioning the quality
ing or excluding unpublished studies and the assessments of RCTs included in meta-analyses.
choice of methods for assessing or adjusting for In other words, they probe at the sharp end of
publication bias. They further conclude that the reviews, and demonstrate that in any meta-
prospective registration of all research studies is analysis, exploration of the sensitivity of the
probably the only way in which the effect of overall results to the quality of included studies
publication bias may be eliminated. In the mean- is of paramount importance. However, they con-
time, further empirical research is required in clude that in the assessment of study quality,
order to assess the extent to which such bias and its integration within a meta-analysis, both
affects current healthcare decision making. further empirical and methodological research is
A second feature of systematic reviewing is required.
the grading of the evidence. The evidence for A fourth component of systematic reviewing
effectiveness or otherwise of healthcare has is that of integrating the evidence. Integrating
traditionally come from a variety of levels of evidence from different sources is essential, in
evidence. Systematic reviews explicitly judge reviews not just of trials and observational
the quality of the evidence, and draw on the studies but also of cost–effectiveness analyses
highest quality of evidence wherever practic- and qualitative research. The combining of evi-
able. A number of hierarchies of evidence are dence from multiple different studies is arguably
in common use for this purpose. All are very at its most straightforward when the studies are
similar, rating randomised controlled trials all high-quality RCTs. Here, the synthesis
(RCTs) highest, and evidence based on un- can use meta-analysis in which the results of
researched consensus lowest. A typical scale trials are quantitatively re-analysed collectively,
would run, therefore, from: (i) multiple RCTs, ideally using original patient data. In Chapter
preferably large ones, suitably meta-analysed; 22, Sutton and colleagues demonstrate that this
(ii) at least one properly designed RCT of appro- is far from straightforward in their exploration
priate size; (iii) well controlled trials without of the different statistical techniques which can
randomisation; (iv) well-designed cohort or be brought to bear on meta-analytic modelling.
case control studies; (v) multiple time series or They use a running theme of the integration of
dramatic results from uncontrolled experiments; studies on cholesterol reduction comparing a
(vi) opinions of respected authorities based on range of interventions, including drug, diet and
clinical evidence, descriptive studies or expert surgery, to improve outcomes in terms of all-
committee; to (vii) small uncontrolled case cause mortality. Whilst they review many of the
series and samples. Notwithstanding this well fundamental meta-analytic methods, they also
accepted hierarchy,4 it does not necessarily fol- consider the use of more recently developed
low that moderate studies at one level are better techniques such as meta-regression, Bayesian
than good studies at another level. methods and methods for the synthesis of evi-
Therefore a third element of systematic dence when it arises from a variety of sources.
reviews is the critical appraisal of the evidence They conclude that although the use of methods
– both of the studies collecting the primary such as meta-regression may be useful in exam-
data, and of the reviews. Critical appraisal is not ining potential sources of between-study hetero-
new: it has long been of concern to journal geneity, the sensitivity of the overall results and
editors (who receive, appraise and publish or conclusions to the choice of the specific meta-
reject) and to researchers (who find, appraise analytic methods used should also be routinely
and synthesise). But it has now become a core investigated.
skill for those planning to use evidence to sup- Finally, the key feature of the systematic
port healthcare decisions – that is all those review revolution, if not of systematic reviews
concerned with HTA. The underlying principles themselves, is the interpretation and delivery of
of critical appraisal concern the validity and the evidence. The scale and complexity of the
applicability of the research discernible in pub- evidence base of healthcare is at the same time
lished papers. Numerous check-lists have been vast and occasionally sparse. It needs to be
developed to inform critical appraisal for pri- interpreted and delivered to the user in a mean-
mary research and review articles.5 Typically, in ingful and valued form – irrespective of the
the case of a review, the questions concern fullness of the evidence for a particular element
whether the review addresses a clearly focused of healthcare. A number of initiatives has
issue, whether the search was appropriate, evolved to meet the need for interpretation and
Part V: Consensus, Reviews and Meta-Analysis 369
delivery, including tertiary reviews of disease- so. Indeed the over-arching conclusion of the
wide topics, guidelines, new information facili- chapters in this section is that researchers must
ties, and dissemination initiatives. Guidelines be aware that not only should the robustness of
are often at the heart of such initiatives. The results and conclusions they draw to the empirical
construction and implementation of guidelines is evidence used be assessed, but so too must their
a wide topic, beyond the scope of this book (but choice of methods.
see Nuffield Institute of Health, 1994).6 But as
the need for clear guidance will often run ahead
of systematic review evidence, a particularly
important frontier, greatly at risk of abuse, is the REFERENCES
construction of guidelines from consensus. In
Chapter 24, Black and colleagues discuss con- 1. Mulrow CD. The medical review article: state of
sensus development methods and their use in the science. Ann. Int. Med. 1987; 104: 485–8.
creating clinical guidelines. What they have to
2. Chalmers I, Altman D. Systematic Reviews.
say has a bearing on consensus methods for any
London: BMJ Publishing, 1995.
purpose. In theory, consensus could be one of
3. NHS Centre for Reviews and Dissemination.
the means of synthesis and systematic reviews.
Undertaking Systematic Reviews of Research On
As the authors note, consensus development is
intended to make the best use of available infor- Effectiveness. York: University of York, 1996.
mation, although unlike synthesis of primary 4. Woolf SH, Battista RN, Anderson GM, Logan
data, that can sometimes just be the collective AG, Wang E and the Canadian Task Force on
wisdom of the participants. For that reason con- the Periodic Health Examination. Assessing the
sensus development methods involve planning, clinical effectiveness of preventive manoeuvres:
individual judgement and group interaction. Not analytic principles and systematic methods in
only do Black and colleagues analyse the lit- reviewing evidence and developing clinical prac-
erature on consensus, but they also perform the tice recommendations. J. Clin. Epidemiol. 1990;
useful service of imposing scientific methods 43: 891–905.
and rigorous thinking, even where data paucity 5. Critical Appraisal Study Programme. Orientation
can turn observers towards nihilism. Given the Guide. Oxford: Institute of Health Sciences, 1996.
importance of sound judgement and precision in 6. NHS Centre for Reviews and Dissemination.
healthcare, and given the breadth of the technol- Implementing Clinical Practice Guidelines: can
ogy frontier with which evidence based health- guidelines be used to improve clinical practice?
care has to cope, it is important that they not do Effective Health Care 1992; 1: 1–12.
21
Publication and Related Biases
the potential sources of publication bias are EVIDENCE OF PUBLICATION AND RELATED
discussed. Then, the methods for dealing with BIASES
publication and related biases, classified accord-
ing to the stage of a literature review, are
described and their usefulness and weakness are The existence of publication bias was first sus-
discussed. The results of a survey of published pected from the observation that a large propor-
systematic reviews are also presented to provide tion of published studies had rejected the ‘null
further evidence of publication bias, and to hypothesis’. In 1959, Sterling found that the
illustrate the methods used for dealing with results of 97% of studies published in four major
publication bias. Finally, we summarise and psychology journals were statistically significant,
discuss the major findings of this review and concluding that studies with non-significant
make recommendations for future research. results might be under-represented.9 In 1995, the
same author concluded that practices leading to
publication bias had not changed over a period
of 30 years.10
METHODS An early example of the identification of
publication bias in medical research is that
The following databases were searched to iden- by Chalmers and colleagues11 who in 1965
tify relevant literature concerning empirical attempted to explain the variability in reported
evidence and methodological issues pertaining rates of deaths due to serum hepatitis. They
to publication and related biases: the Coch- concluded that there was a tendency for clinicians
rane Review Methodology Database, Medline, or editors to publish unusual findings. It was
EMBASE, BIDS, Library and Information suggested that small studies with unusually high
Science Abstracts, Psyclit, Sociofile, Eric, Dis- or low fatality rates might have a greater chance
sertation Abstracts, MathSci, British Education to be reported than those with ‘more average or
Index, SIGLE, and ASSIA. The reference lists mundane fatality rates’.
in identified articles were checked, and authors During the 1980s and 1990s, more evidence
of some articles contacted to identify further about the existence of publication bias has been
studies. provided by surveying authors, comparing pub-
It is difficult to define a clear and narrow lished with unpublished studies, and following-
criterion for including studies in this review up cohorts of registered studies.
because of the broad nature of related issues and
the great diversity of relevant studies. Therefore, Survey of Authors
all studies relevant to publication-related biases
were included, although studies may be excluded According to a survey of 48 authors of articles
if the issue of publication bias is only mentioned submitted to the journal of Personality and
but is not a major topic. The search results from Social Psychology in 1973, the probability of
the electronic databases were independently submitting for publication was 0.59 for studies
checked by two reviewers to identify all rele- with significant result and only 0.06 for studies
vant studies and any difference was solved by with non-significant result.12 A survey of 1000
discussion. members of the American Psychological Asso-
A survey of published systematic reviews was ciation found that studies with ‘neutral or
also undertaken, to identify further evidence of negative’ findings were less likely to be sub-
publication bias, and to illustrate what methods mitted or accepted for publication.13 Dickersin
are currently used for detecting and reducing et al.14 observed that the proportion of trials in
publication bias. A sample of systematic which the new treatment was superior to the
reviews was selected from the Database of control therapy was 55% in 767 published
Abstracts of Reviews of Effectiveness (DARE) studies compared to only 14% in 178 un-
produced by the NHS Centre for Reviews and published studies. In another survey, the rate of
Dissemination at the University of York. These publication was 73% for 30 studies with sig-
published reviews were independently assessed nificant results and 54% for 26 studies with non-
by two reviewers using a data-extraction sheet significant findings.15
to collect the following information: type of
review (narrative or meta-analysis); whether the
issue of publication bias was considered; Published versus Registered Trials in a
whether unpublished studies or those published Meta-Analysis
in non-English languages were searched for and
included; any evidence on the existence, extent The most direct and convincing evidence of the
and consequence of publication bias; and the existence of publication bias comes from the
methods used for dealing with publication bias. comparison between unpublished and published
Publication and Related Biases 373
results. In a meta-analysis that compared com- example, research results published only in
bination chemotherapy versus initial alkylating reports, working papers, theses or abstracts,
agent in advanced ovarian cancer, Simes found often have limited dissemination, and may be
that overall survival was significantly higher in termed as ‘grey literature’.24 Other biases that
the combination chemotherapy group than in are publication related include language bias,
the initial alkylating agent group when the pub- outcome reporting bias, duplicate bias and cita-
lished trials were combined.16 However, this tion bias. The available evidence about these
survival advantage for combination chemother- biases are discussed below.
apy was not found when all registered trials (both
published and unpublished) were combined.
Grey literature bias
Follow-up of Cohorts of Registered In 1964, Smart randomly selected 37 Ph.D.
Studies theses from Dissertation Abstracts in Psychol-
ogy and found that theses with positive results
The existence of publication bias has been were more likely to be published than those with
consistently confirmed by studies which have negative results.25 In the fields of psychological
followed-up cohorts of studies approved by and educational research, several authors ob-
research ethics committee or cohorts of trials served a tendency that the average effects
registered by the research sponsors (Figure estimated from journal articles were greater
1).17–20 By pooling results from these cohort than the corresponding effects from theses and
studies, the rate of publication of studies with dissertations.26–28
significant results was significantly higher than A large number of study results are initially
those with non-significant results (overall odds presented as abstracts at various scientific meet-
ratio 2.54, 95% confidence interval (CI): 1.44 to ing or journals. Two studies found that sub-
4.47).21 The publication bias in favour of studies mitted abstracts reporting positive results were
with significant results was confirmed in these more likely to be accepted for presentation than
studies by multivariate analysis adjusting for those reporting negative results.29,30 On average,
other factors such as study design, sample size about half of abstracts were published in full
and funding sources. after more than 12 months.31 The association
Studies with significant results were also between significant results and full publica-
more likely to generate multiple publications tion of abstracts have been observed in two
and more likely to be published in journals with studies31,32 but not in others.33–35
a high citation impact factor when compared to
studies with non-significant results.17 It may be Language and country bias
interesting to note that studies with a non-
significant trend were less likely to be published The effect of language bias was investigated by
than studies with no difference.17,20 including 19 studies that had been originally
Stern and Simes observed that clinical trials excluded for linguistic reasons from 13 meta-
with statistically significant results were pub- analyses.36 By doing this, one statistically non-
lished much earlier than those with non- significant result in a meta-analysis became
significant results (median 4.7 years versus 8.0 statistically significant. By comparing 40 pairs
years).20 This finding was confirmed using a of trials conducted by German-speaking invest-
cohort of randomised controlled trials on AIDS, igators, it was found that authors were more
in which it was found that the median time from likely to publish trials in an English-language
starting enrollment to publication was 4.3 years journal if the results were statistically signifi-
for positive trials and 6.5 years for negative cant.37 When 180 Scandinavian referees were
trials.22 In a study of publication bias among asked to review two fictitious manuscripts in
studies examining the effects of passive smok- English or in the national language, the English
ing, it was found that the median time to pub- version was on average given a higher quality
lication was 5 years (95% CI: 4 to 7) for score than the national-language version of the
statistically non-significant studies and 3 years same manuscript.38
(95% CI: 3 to 5) for statistically significant Ottenbacher observed that the estimated effect
studies (P 5 0.004).23 of spinal manipulation therapy was larger in
studies reported in English-language journals
Other Publication-Related Biases published outside the United States than similar
studies in journals published in the United
The accessibility of research results is dependent States.39 In a survey of Medline abstracts of
not only on whether and when a study is pub- trials evaluating acupuncture or other interven-
lished, but also on its format of publication. For tions, the proportion of studies with a positive
374 The Advanced Handbook of Methods in Evidence Based Healthcare
result was much higher for trials originating Outcome reporting bias
from China, Taiwan, Japan and Hong Kong than
trials conducted in western countries such as Outcome reporting bias happens when trials
USA, Sweden, UK, Denmark, Germany and with multiple outcome measures report only
Canada.40 It was suggested that publication bias those which are significant. In 45 clinical trials
is a possible explanation for the unusually high published in three general medical journals the
proportions of positive results observed for median number of end-points was six per trial.41
some countries. Tannock found that more than half of the
Publication and Related Biases 375
implied statistical comparisons had not been complicated twin pregnancies, but the findings
reported in 32 oncology trials.42 In addition, the remained unpublished for 7 years.52 Chalmers
methods of reporting and the completeness of pointed out that ‘at the very least, this delay
information may influence the interpretation of led to continued inappropriate deployment of
study findings.43–45 limited resources; at worst, it may have resulted
in the continued use of a harmful policy’.52
Duplicate (multiple) publication In 1980, a trial tested lorcainide in patients
with acute and recovery myocardial infarction,
In a meta-analysis of ondansetron on postoper- and observed more deaths in the treatment group
ative emesis, Tramer et al.46 found that efficacy than in the placebo group (9/48 versus 1/47).53
was over-estimated when duplicate publications The trial results were not published because the
were included as compared with when they were development of lorcainide was stopped for
excluded. This bias is often difficult to detect; ‘commercial reasons’. About a decade later,
for example, in a review of risperidone, it was an increased mortality was observed amongst
found that identifying duplicate publications of patients treated with encainide and flecainide in
the multicentre trials was far from obvious, two trials.54,55 Encainide, flecainide and lorcain-
because of the chronology of publications, ide all belong to a class of 1C anti-arrhythmic
changing authorship, lack of transparency in agents. If the results of the early trial in 1980
reporting, and the frequent citation of abstracts has been published, the mortality of patients
and unpublished reports.47 included in the later two trials might have been
avoided or reduced.
Citation (reference) bias Based on the results of a meta-analysis of
Citation bias may occur at the post-publication several small trials, intravenous magnesium was
stage.7 For example, supportive trials were found recommended as a treatment for acute myo-
to be cited almost six times more often than cardial infarction.56 However, a subsequent large
non-supportive trials of cholesterol-lowering trial (ISIS-4) showed that the rate of death was
interventions.48 By examining quotations in higher in patients receiving intravenous mag-
three influential reviews on diet–heart issues, it nesium than those receiving standard treatment
was found that only one of six relevant random- (7.64% versus 7.24%, P 5 0.07).57 Publication
ised trials with a negative result was cited, and bias has been identified as a possible explana-
this was only in one of the three reviews.49 tion for the discrepant results from the meta-
Conversely, the three reviews respectively cited analysis and the following large trial.58 The
two, four and six non-randomised trials with a number of hospital admissions due to acute
positive result. However, in reviews of the myocardial infarction was 116 635 in 1993–94
clinical effectiveness of pneumococcal vaccine, in England.59 If all these patients had received
non-supportive trials were more likely to be intravenous magnesium, there would have been
cited than supportive trials.50 This may be 466 more deaths than without using such
because the reviews tended to cite more recently therapy.
published trials in which outcomes were more
frequently found to be non-supportive.
Koren and Klein compared newspaper cov-
erage of one positive study and one negative
WHO IS RESPONSIBLE FOR PUBLICATION
study on radiation as a risk for cancer which BIAS?
were published in the same issue of JAMA in
1991.51 Nine of the 19 newspaper reports cov- Publication bias may be introduced intentionally
ered only the positive study. In 10 other reports or unintentionally, consciously or unconsciously
that covered both positive and negative studies, because of varying motivations or biased
the average number of words was 354 for the standards used to judge research evidence.60
positive study, and 192 for the negative study. Investigators, journal editors, journal peer
reviewers and research sponsors may all be
responsible for the existence of publication bias,
CONSEQUENCES OF PUBLICATION BIAS though the extent of such responsibility may be
different in different circumstances.
Although the existence of publication bias is
well demonstrated, there is limited empirical Investigators and Authors Investigators may
evidence about the consequences of publication get inappropriate recommendations not to pub-
bias. In the perinatal research, a study observed lish studies with non-significant results. For
that routine hospitalisation was associated with example, it has been suggested that non-
more unwanted outcomes in women with un- significant research ‘clutters up the journals and
376 The Advanced Handbook of Methods in Evidence Based Healthcare
does more harm than good to the author’s repu- of originality accounted for 14% of all reasons
tation in the minds of the discerning’.61 Avail- given for rejection of manuscripts in 1989 by the
able evidence indicates that investigators are the American Journal of Surgery.66 Confirmatory
main source of publication bias, for not writing trials, either positive or negative, have a low
up or submitting studies.18,19 The most frequent chance of being accepted.67 A journal on diabetes
reason given for failing to write up the study clearly stated that ‘mere confirmation of known
was the presence of a null result.17,20 Studies facts will be accepted only in exceptional cases;
with positive results were often submitted for the same applies to reports of experiments and
publication more rapidly after completion than observations having no positive outcome’.68
were negative studies.22 The New England Journal of Medicine would
Chalmers and colleagues7 observed that normally reject epidemiological studies with a
author’s speciality was significantly associated relative risk smaller than three.69 Not surpris-
with the enthusiasm for the procedure reviewed ingly, editors will publish negative studies that
in an article. For example, 21 of the 29 radio- may have a potential to change current practice
therapists were enthusiastic for radiotherapy by showing that a widely used intervention is
after radical mastectomy when stage is not dis- ineffective.70
tinguished, compared with only five of 34
authors with other specialities. In a systematic
review of the risks of stroke and death due to Peer Reviewing
endarterectomy for symptomatic carotid stenosis,
it was observed that the risk of stroke or death Journal peer review has been defined as ‘the
was highest in studies in which patients were assessment by experts (peers) of material sub-
assessed by a neurologist after surgery (7.7%) mitted for publication in scientific and technical
and lowest in studies with a single author affili- periodicals’.71 The degree to which peer review
ated with a department of surgery (2.3%).62 It is contributes to publication bias was investigated
possible that surgeons were less likely to report by Mahoney, who examined the recommenda-
the results if the operative risk of stroke and tions of 75 journal referees about a fictitious
death was high. manuscript with identical experimental pro-
cedures but different results.72 It was found that
referees were biased against the manuscript
Editorial Policy of Journals which reported results contrary to their own
perspectives. In another study, it was also found
Although editorial rejection was not a frequent that referees’ judgements on a fictitious research
reason given by investigators for studies remain-
report was associated with their own preconcep-
ing unpublished,17,18 it cannot be ruled out that
tions and experience.73 However, Abbot and
authors do not submit studies with negative
Ernst74 did not find peer reviewing bias against
results because of anticipated rejection accord-
ing to journals’ instructions to authors and their positive or negative outcome in complementary
own experience. In a survey of 80 authors of medicine.
articles published in psychology or education
journals in 1988, 61% of the 68 respondents Research Funding Bodies and
agreed that if the research result is not statis-
tically significant, there is little chance of the Commercial Interests
manuscript being published.63 Weber et al.34
found that anticipated rejection was given as a According to a telephone survey of 306 com-
reason for failure to submit a manuscript by panies in the United States in 1994, 82% of
20% of 179 authors. In another study, 17 of 45 companies interviewed asked investigators at
submitted trials were rejected by at least one universities ‘to keep the results of research
journal, and at least four negative trials with secret beyond the time needed to file a patent’.75
over 300 patients each were rejected two or Many clinical trials submitted by drug com-
three times, while no positive trial was multiply panies to licensing authorities have never been
rejected.22 published.76–78 Commercial interests and intel-
‘Originality’ is one of the most important lectual property are often the main reasons for
criteria upon which journals decide whether a the non-publication of clinical studies funded
submitted paper will be accepted.64 For ex- by drug companies.79–81 Abraham and Lewis80
ample, according to the Lancet’s instruction to argued that ‘the present European medicines
authors, articles published ‘are selected, from licensing system is biased in favour of commer-
among a huge number of submissions, if they cial interests at the expense of medical science,
are likely to contribute to a change in clinical public health, and the adequate provision of
practice or in thinking about a disease’.65 Lack information for doctors and patients’.
Publication and Related Biases 377
amnesty’,112,113 and by the end of 1998, this trial can be up-dated in order to reduce the con-
amnesty has registered 150 trials.98 Since the sequence of delayed publication of trials with
beginning of 1997, a general medical journal, unfavourable results. The methods that may be
The Lancet, has been assessing and registering used in systematic reviews to deal with the prob-
selected protocols of randomised trials, and lem of publication bias are discussed below.
providing a commitment to publish the main
clinical findings of the study.65 However, there
is little hope that conventional paper journals
Locating unpublished trials
can solve the problem of publication bias Only limited success has been reported in identi-
because of space limitation and the requirement fying unpublished studies by sending question-
of newsworthy articles for maintaining or naires to investigators, organisation or research
increasing the journals’ circulation level. sponsors.14,119 In a systematic review of near
Fox considered that medical journals have patient testing, no unpublished data were ob-
two basic functions: as a medical recorder, and tained by sending questionnaires to 194 academ-
as a medical newspaper.114 The function of ics and 152 commercial companies.120 Because
medical recorder is important for communica- the quality of unpublished trials may be differ-
tion between investigators for the advancement ent from that of published trials, there is dis-
of medical knowledge, while the function of agreement about whether unpublished studies
medical newspaper is to disseminate informa- should be included in meta-analyses.70,121,122
tion relevant to medical practice. Both of It has been suggested that the use of unpub-
these two basic functions may benefit from the lished data can not necessarily reduce the bias in
development of modern computer and Internet meta-analysis, particularly if the unpublished
technology.115 data are provided by interested sources such as
Publication bias may be reduced by intro- pharmaceutical companies.123
ducing new peer-reviewed electronic journals
without space limitation, in which papers
should be accepted based only on the research Estimating the number of unpublished
methodology or validity criteria, and not on the studies
results.52,116–118 Because the originality is no Rosenthal’s fail-safe N method (or the file–
longer a requirement, these types of electronic drawer method) is a statistical method to estimate
journals will encourage the submission of
the number of unpublished studies required,
trials with negative or non-striking results, and
with zero treatment effect on average, to over-
trials that simply replicate previous trials. Such
turn a significant result in a meta-analysis.124 If
journals would be mainly used as medical
recorders and therefore would be most useful for the number of unpublished studies required to
investigators and people who conduct system- overturn the statistically significant result is
atic reviews. large, and therefore unlikely to exist, the risk of
publication bias is considered to be low and thus
the results obtained from published studies are
Systematic Review deemed to be reliable.
The plausible number of unpublished studies
For many types of publication-related biases may be hundreds in some areas or only a few in
(such as grey literature bias, language bias, others. Therefore, the estimated fail-safe N
citation bias and duplication bias), a systematic should be considered in proportion to the num-
approach to searching, assessing and summaris- ber of published studies (K). Rosenthal sug-
ing published or unpublished studies is crucial. gested that the fail-safe N may be considered as
Conducting a thorough literature search when being unlikely if it is greater than a tolerance
carrying out systematic reviews may reduce the level of ‘5K 1 10’.124
risk of missing studies that are published in low- A meta-analysis of risperidone versus typical
circulation journals or in the grey literature. neuroleptics in the treatment of schizophrenia
The risk of potential publication bias may be shows that risperidone is associated with statis-
estimated according to its association with some tically significantly more patients who had clin-
study characteristics, such as observational ically improved (odds ratio 0.75; 95% CI: 0.61
studies or small sample size. In meta-analysis, to 0.92) (Figure 2).125 By applying Rosenthal’s
larger studies are normally given greater weight method, at least 23 unpublished studies with
than smaller studies. This procedure will have zero treatment effect on average are required in
an advantage in reducing the impact of publica- this meta-analysis to change the statistically
tion bias since less weight is given to smaller significant result into a statistically non-
studies that are associated with a greater risk of significant result. Although the fail-safe N is two
publication bias. In addition, systematic reviews times greater than the published studies, it is less
Publication and Related Biases 379
than 60 (that is, 5 3 10 1 10), a tolerance level bound) of unpublished studies.127 The models
suggested by Rosenthal. are based on the following three basic assump-
Two problems with the fail-safe N method tions: (i) all studies, published and unpublished,
have been identified.126 Firstly, the method over- are mutually statistically independent; (ii) the
emphasises the importance of statistical sig- P-value in each study is based on a continuous
nificance. Secondly, it may be misleading when test statistic; and (iii) the null hypothesis is
the unpublished studies have an average effect true. Because these methods are based on the
that is in the opposite direction to the observed assumption that the null hypothesis is true, they
meta-analysis. If the unpublished studies reported cannot be used to estimate the effects of publica-
contrary results compared to those in the pub- tion bias on effect size.
lished studies, the number of unpublished studies
required to overturn a significant result would be
smaller than that estimated assuming an average
Funnel plot
effect of zero in unpublished studies. Because of larger random error, the results from
The fail-safe N estimated by using Rosenthal’s smaller studies will be more widely spread
method is not necessarily related to the actual around the average effect. A plot of sample size
number of unpublished studies. Using the P- versus treatment effect from individual studies
values reported in the published studies, Gleser in a meta-analysis should be shaped like a
and Olkin proposed two general models for funnel if there is no publication bias.2 If the
estimating the number (and its lower confidence chance of publication is greater for trials with
380 The Advanced Handbook of Methods in Evidence Based Healthcare
statistically significant results, the shape of the statistically significant and more likely to be
funnel plot may become skewed. published. Consequently there will be a lack of
Light and Pillemer described two ways in small studies with small effect in the funnel plot,
which the shape of the funnel plot can be and the funnel plot will be skewed with a larger
modified when studies with statistically sig- effect among smaller studies and a smaller effect
nificant results are more likely to be published.2 among larger studies. This will result in an
Firstly, assume that the true treatment effect is over-estimation of the treatment effect in a
zero. Then the results of small studies can only meta-analysis.
be statistically significant when they are far In a funnel plot, the treatment effects from
away from zero, either positive or negative. If individual studies are often plotted against their
studies with significant results are published and standard errors (or the inverse of the standard
studies with results around zero are not pub- error) instead of the corresponding sample sizes
lished, there will be an empty area around zero (Figure 3). Use of standard errors may have
in the funnel plot. In this case, the funnel plot some advantages because the statistical signifi-
may not be obviously skewed but would have a cance is determined not only by the sample size
hollow area inside. These polarised results (sig- but also by the level of variation in the outcome
nificant negative or positive results) may cause measured, or the number of events in the case of
many debates; however, the overall estimate categorical data. However, the visual impression
obtained by combining all studies is unlikely to of a funnel plot may change by plotting treat-
be biased. ment effects against standard errors instead of
When the true treatment effect is small but against the inverse of standard errors.128
not zero, small studies reporting a small effect There are some limitations in the use of
size will not be statistically significant and funnel plot to detect publication bias. For a
therefore less likely to be published, while small funnel plot to be useful, there needs to be a
studies reporting a large effect size may be range of studies with varying sizes. The funnel
Figure 3 Funnel plot of log odds ratio against inverse of standard error.
Meta-analysis of risperidone versus typical neuroleptics for schizophrenia.
Publication and Related Biases 381
plot is an informal method for assessing the was larger in meta-analyses than in the corres-
potential publication bias subjectively, and dif- ponding large trials. Using Egger et al.’s
ferent people may interpret the same plot differ- method, significant asymmetry was found in
ently. It should be stressed that a skewed funnel three of these four meta-analyses; however,
plot may be caused by factors other than pub- when the rank correlation test was used, only
lication bias. Other possible sources of asym- one of the four meta-analyses showed signifi-
metry in funnel plots include different intensity cant asymmetry. Thus the linear regression
of intervention, differences in underlying risk, method129 appears to be more sensitive than the
poor methodological design of small studies, rank correlation test.130
inadequate analysis, fraud, choice of effect The results of applying this method to the
measure, and chance.129 meta-analysis of risperidone for schizophrenia
are shown in Figure 4. Because the intercept of
Rank correlation test the weighted regression is significantly less than
zero, it indicates that the small studies are asso-
Begg and Mazumdar suggested that a rank cor- ciated with a larger treatment effect.
relation test can be used to examine the asso- In the original Galbraith radial plot, the slope
ciation between effect estimates and their vari- of the line indicates the size and direction of
ances, as a complementary method to the funnel effect; a greater gradient of the slope indicates a
plot.130 The rank correlation test is a distribution- greater difference in the treatment effect.131
free method which involves no modelling However, for testing the asymmetry of a funnel
assumptions. However, it suffers from a lack of plot, the gradient of the slope will become closer
power and so the possibility of publication bias to zero or positive when the estimated effect is
cannot be ruled out when the test is non- greater in smaller studies. The operating charac-
significant. teristics of this method need to be evaluated
According to simulated results, the power of thoroughly by more analytic work, or computer
the rank correlation test is related to several
simulations.132
factors: the number of component studies in the
meta-analysis; the underlying effect size para-
meter; the range of variances across studies; Trim and fill method
the strength of the selection function; and the The trim and fill method is a simple rank-based
presence of one-sided or two-sided selection data augmentation technique to formalise the
pressures.130 The test is fairly powerful for large
use of the funnel plot.128 This recently developed
meta-analyses with 75 component studies, but
method can be used to estimate the number of
has only moderate power for meta-analyses with
missing studies and, more importantly, to pro-
25 component studies. In the meta-analysis of
risperidone for schizophrenia,125 the rank corre- vide an estimate of the treatment effect by
lation test did not find an association between adjusting for potential publication bias in a
the estimated treatment effects and their vari- meta-analysis. Briefly, the asymmetric outlying
ances (Spearman’s rho correlation coefficient part of the funnel is firstly ‘trimmed off’ after
0.018; P 5 0.96). estimating how many studies are in the asym-
metric part. The symmetric remainder is then
used to estimate the ‘true’ centre of the funnel.
Linear regression approach Finally, the ‘true’ mean and its variance are
Egger and colleagues129 suggested a method to estimated based on the ‘filled’ funnel plot in
test the asymmetry of a funnel plot, based on a which the trimmed studies and their missing
regression analysis of Galbraith’s radial plot.131 ‘counterparts’ symmetric about the centre are
The standard normal deviate, defined as the log replaced. In simulation studies, it was found that
odds ratio divided by its standard error, is the method estimated the point estimate of the
regressed against the estimate’s precision (i.e. overall effect size approximately correctly and
the inverse of the standard error). In this linear the coverage of the confidence interval is sub-
regression analysis, the intercept is used to stantially improved, as compared to ignoring
measure asymmetry; a negative intercept indi- publication bias.128
cates that smaller studies are associated with Applying the Trim and Fill method to meta-
bigger treatment effects. analysis of risperidone for schizophrenia (see
By applying this method, significant asym- Figure 3) suggests that the original funnel plot
metry was observed in 38% of 37 meta-analyses would need three trials to fill on the right side.
published in a selection of journals, and in 13% After filling with the assumed missing trials, the
of 38 Cochrane reviews.129 Egger et al. also adjusted odds ratio (0.82, 95% CI: 0.68 to 1.01)
identified four meta-analyses in which discord- become non-significant compared to that based
ant funnel plots showed that the treatment effect on the published trials (0.75, 95% CI: 0.61 to
382 The Advanced Handbook of Methods in Evidence Based Healthcare
The slope of line A corresponds to the pooled log odds ratio if there is no bias (i.e., the intercept 5 0).
The line B is an unweighted regression in which the intercept is negative but not statistically
significant (intercept 5 20.704, P 5 0.223). The line C is a weighted (by the inverse of variance)
regression in which the intercept is significantly less than zero (intercept 5 21.392, P 5 0.016),
therefore it indicates the existence of bias in favour of small trials.
0.92). The advantage of risperidone versus con- Unfortunately, the complexity of these meth-
ventional neuroleptics in clinical improvement ods means that they have largely been used only
for schizophrenia becomes smaller and no by statistical modelling experts. Hence, although
longer statistically significant after adjusting applications of their use do exist,139,142 they are
potential publication bias. limited in number and no comprehensive com-
parison of the methods has ever been carried
out. For this reason, many feel they are still in
More sophisticated modelling methods the experimental stage of development, despite
The impact of missing studies may also be being considered for well over a decade.143
assessed by using more sophisticated methods. Development of user-friendly software is
Several selection modelling methods have been required to bring the methods into more main-
developed to investigate/adjust the results of a stream use. They have the potential to be a
meta-analysis in the presence of publication valuable tool for assessing the robustness of a
bias. Many of these methods are related and meta-analyses results to publication bias; if not
based on weighted distribution theory derived to adjust the pooled estimate per se.143
from both classical133–137 and Bayesian138,139 per-
spectives, although other methods do exist.140 Confirmatory Large-Scale Trials
There are two aspects to the selection models
which use weighted distribution theory: an For the purpose of avoiding moderate biases and
effect size model which specifies what the dis- moderate random errors in assessing or refuting
tribution of the effect size estimate would be if moderate benefits, a large number of patients in
there were no selection, and the selection model randomised controlled trials are required.144
which specifies how this effect size distribution Large-scale trials are generally believed to be
is modified by the selection process.141 In some less vulnerable to publication bias; this is the
methods the nature of the selection process is fundamental assumption of many methods for
pre-defined by the researcher, while in others it detecting publication bias. Though the results of
is dictated by the available data. meta-analyses often agree with those of large
Publication and Related Biases 383
trials, important discrepancies have been ob- used methods were fail-safe N (in 14 reviews)
served.145–147 When the existence of publication and funnel plot-related methods (in 11 reviews).
bias is likely and the consequences of such bias Other methods used included large-scale trials,
are clinically important, a confirmatory, multi- sensitivity analysis, and comparing published
centre large-scale trial may be conducted to with unpublished studies.
provide more convincing evidence.
DISCUSSION
SURVEY OF PUBLISHED SYSTEMATIC REVIEWS
The existence of publication bias is demon-
At the end of August 1998, 193 systematic strated by showing an association between the
reviews published in 1996 were included in the chance or manner of publication and the strength
Database of Abstracts of Reviews of Effective- or direction of study findings. It is therefore
ness (DARE) at the NHS Centre for Reviews important to define what we mean by two basic
and Dissemination, University of York. These concepts: publication and study findings.
reviews have been selected to examine the The formats of publication include full pub-
issues and methods relevant to publication bias. lication in journals, presentation at scientific
Among the 193 systematic reviews, the majority conferences, reports, book chapters, discussion
(83%) evaluated the effectiveness of health papers, dissertations or theses. The studies
interventions; 8% focused on adverse-effects of presented at scientific meetings are often not
interventions; and 9% evaluated diagnostic tech- considered to be published, and there are dis-
nologies. In total, 131 of these reviews were agreements about how reliable they are. For
meta-analyses in which the results of primary example, 17 of the 133 meta-analyses surveyed
studies are combined quantitatively, while 62 in this review explicitly excluded abstracts
were narrative systematic reviews in which because they were not peer-reviewed.
quantitative pooling was not employed. Study findings are commonly classified as
Searching for unpublished studies was being statistically significant or statistically non-
reported more often in meta-analyses than in significant. Sometimes, study results are classi-
narrative systematic reviews (32% versus 16%). fied as being negative versus positive,22,40,148,149
The methods used were mainly to write to supportive versus unsupportive,50,90 or striking
investigators/authors, research organisations and versus unimportant.20 The classification of study
pharmaceutical companies. Meeting proceedings findings are often dependent upon subjective
were often searched to identify unpublished judgement, and may therefore be unreliable.
abstracts. By checking the reference lists of all In addition to the difficulties in defining pub-
193 reviews, it was found that conference lication and classifying outcomes, studies of
abstracts were included in some reviews in publication bias themselves may be as vulner-
which unpublished studies were not explicitly able as other studies to selective publication of
searched for. In total, 35% of meta-analyses and significant or striking findings.4 Therefore, the
31% of narrative systematic reviews explicitly available evidence about publication bias should
searched for, or included, unpublished studies. be interpreted with caution.
Seventeen meta-analyses explicitly excluded The empirical evidence demonstrates that
unpublished studies or abstracts because they studies with significant results or favourable
were not peer-reviewed and were therefore con- results are more likely to be published or cited
sidered to be unreliable. than those with non-significant or unfavourable
Non-English language literature were searched results. The extent and direction of such selec-
for or included in 30% of the systematic tive publication is still uncertain, and may vary
reviews. However, language restrictions may greatly depending on the circumstances. In addi-
still exist in some reviews that included non- tion, there is little empirical evidence about the
English language literature. For example, the impact of publication and related biases on
reviews were often limited to literature in major health policy, clinical decision-making and out-
European languages, or English plus the review- come of patient management.
ers’ native languages. Because of the lack of empirical evidence
The problem of potential publication bias was about the impact of publication bias, there is
discussed or mentioned more frequently in the disagreement on the actual importance of pub-
meta-analyses than in the narrative systematic lication bias in the evaluation of healthcare
reviews (46% versus 11%). Methods of testing interventions. Some have argued that the poten-
publication bias were used in 27 of the 131 tial consequences of publication bias are seri-
meta-analyses and in only one of the 62 narra- ous,4,52,150 whilst others argue that positive
tive systematic reviews. The most commonly studies are more important than negative studies
384 The Advanced Handbook of Methods in Evidence Based Healthcare
and selective publication of positive studies is mainly designed to assess healthcare interven-
not a clinical problem.151–153 At the very least, tions with small or moderate (but clinically
it is arguable that under-reporting research is important) effects.
scientific misconduct that may cause inappropri- The interest of investigators and funding
ate patient care, and also it is unethical to bodies may be associated with publication bias.
mistreat the trust of patients involved and waste Therefore, the risk of bias may be great if all
resources invested.6,52 trials are funded by a single body with explicit
Practitioners and researchers may have devel- or implicit reasons for favouring a particular
oped some immunity to publication bias, by finding. Conversely, when similar results are
waiting for confirmatory studies about a claimed obtained from trials funded by different bodies
advancement before changing their practice. with conflicting interests, the risk of bias due to
This skepticism about new developments may funding bodies may not be important.
reduce the impact of publication bias. However, The design quality of studies may be asso-
the corollary of this is that the introduction of a ciated with the risk of publication bias. Non-
new cost-effective technology may be unneces- randomised studies, single-centre studies, Phase
sarily delayed. I and II trials might be more susceptible to
It seems that the most common reason for publication bias than randomised studies, multi-
publication bias is that investigators fail to centre studies and Phase III trials.155,156 Risk
write-up or submit studies with non-significant factors for publication bias were assessed, but
results. Investigators may lose interest in non- not consistently identified, across several cohort
significant results, or be motivated not to publish studies of publication bias.17–20 Irwig and col-
results that are unfavourable to their own inter- leagues157 suggested that publication bias is
est. However, it is possible that investigators more of a problem for diagnostic tests than for
would not submit studies with non-significant randomised trials because ‘many studies of test
results mainly because of anticipated rejection accuracy may use data collected primarily as
by journals. Paper-based journals often select part of clinical care, there may be no clear
manuscripts that are original with an import- record of attempted evaluations’.
ant impact on practice. The potential role of
research sponsors needs to be emphasised,
because evidence has shown that studies with Implications for Decision-Makers and
favourable results to funding body’s interest are Researchers
more likely to be disseminated.
The potential problem of publication and other
selection biases should be taken into considera-
Risk Factors for Publication Bias tion by all who are involved in evidence based
decision making. For research funding bodies
Studies with small sample sizes tend to produce and research ethics committees, all funded or
results with great variation and therefore give approved studies should be prospectively regis-
people a great range of results to select for tered and such registrations should be accessible
publication. Simulations have demonstrated that to the public. The dissemination of research
small sample size is associated with a great results should be considered as an integral part
extent of publication bias when only studies of research.
with significant results are published.133,154 In In all systematic reviews, a thorough litera-
practice, a small study with a non-significant ture search is crucial to identify all relevant
result may be readily abandoned without trying studies, published or not. Whenever possible,
to publish because it is easy and cheap to carry registers of clinical trials should be used in a
out in terms of time, staff and other resources systematic review to identify relevant studies,
invested. In addition, small trials may often be and identified trials should be classified accord-
poorly designed and conducted. Therefore, the ing to whether they are prospectively or retro-
risk of publication bias will be great if many spectively registered. The search of electronic
small trials have been conducted.143 databases alone is seldom sufficient and should
The simulation results also indicated that the be supplemented by checking references of
extent of bias, by selecting only the significant relevant studies and contacting experts or
results to publish, is greater when the true effect organisations.120
is small or moderate than when the true effect is Because even a thorough literature search
zero or large.154 A small or moderate effect (or cannot eliminate publication bias, the risk of
weak association) may be considered as a risk publication bias should be assessed and, when
factor for publication bias. This risk factor may estimated, should be incorporated into the
exist in most cases because clinical trials are review’s conclusions and recommendations in
Publication and Related Biases 385
systematic reviews. The risk of publication bias Recommendations for Future Research
can be assessed, for example, according to the
sample sizes of studies, the potential number of Further research is needed to provide more reli-
studies that may have been conducted, research able empirical evidence about publication and
sponsors, time of publication of studies, and related biases. Especially, there is a lack of
heterogeneity across studies. Although it is still reliable evidence about the impact of publication
controversial, systematic reviews should not bias on health decision-making and outcomes of
systematically exclude unpublished studies or patient management.
abstracts presented at scientific meetings. The In the foreseeable future, many systematic
quality of unpublished studies or abstracts reviews may still have to depend upon studies
should be assessed using the same criteria for identified retrospectively from the published
assessing published studies. Sensitivity analysis literature. The available methods for detecting
could be used in systematic reviews to compare or adjusting for publication bias should be
the results with and without data from unpub- evaluated and measures taken to make the more
lished studies. complex methods easier to use. Further research
Some methods are available in systematic is also needed to develop new methods that are
reviews to deal with publication bias, such as sensitive and easy to use for detecting or adjust-
the fail-safe N method, funnel plot, rank correla- ing for publication bias in systematic reviews. In
tion method, linear regression method, trim- addition, there is a lack of methods that can be
and-fill method, and some complex modelling used to detect publication bias in narrative sys-
tematic reviews.
methods. Rosenthal’s fail-safe N and funnel
It is most important for future research to
plot-related statistical methods are the most identify the cost-effective and feasible methods
commonly used. These methods are mainly use- for preventing publication bias. Further research
ful to detect publication bias, although some is needed to answer questions about how to
methods (such as trim-and-fill method) could establish and maintain prospective registration
provide an estimate by adjusting for the detected of clinical trials and observational studies; how
bias. It should be stressed that all these methods to make all research findings accessible to the
are by nature indirect and exploratory, because public; and how the developments in computer
the true extent of publication bias is generally science and information technology can be used
unknown. to solve the problem of publication bias. The
There are some methodological difficulties in problem of publication bias is unlikely to be
using the available methods to test and adjust for solved by the conventional paper-based medical
publication bias in meta-analyses. In most cases, journals because of their intrinsic limitations.
it is impossible to separate the influence of Further research about publication bias should
factors other than publication bias on the ob- be an integral part of research that explores
served association between the estimated effects alternatives to the conventional methods for
and sample sizes across studies. The appro- generating, disseminating, preserving and utilis-
priateness of many methods is based on some ing scientific research findings.
strict assumptions that can be difficult to justify
in the real practice. For these reasons, it seems
reasonable to argue that these methods ‘are not ACKNOWLEDGEMENTS
very good remedies for publication bias’.6 The
attempt at identifying or adjusting for publica- This work was supported by the NHS R&D
tion bias in a systematic review should be Health Technology Assessment Programme.
mainly used for the purpose of sensitivity analy- The authors would like to thank Kathleen
ses, and the results should be interpreted with Wright for assisting with the search and location
great caution. of the literature; Sue Duval at University of
Large-scale confirmatory trials become neces- Colorado Health Sciences Center for help in
sary after a systematic review has reported a using the trim-and-fill method. We also thank
clinically significant finding, but publication Iain Chalmers, Trevor Sheldon, Jos Kleijnen,
bias cannot be safely excluded as an alternative Andrew Stevens, and Keith Abrams for com-
explanation. As compared with a universal menting on the early manuscript.
register of all trials, large confirmatory trials are
more selective about the research areas and
objectives, but more flexible at the same time to REFERENCES
minimise the impact of other biases, for ex-
ample, biases related to study design, selection 1. Mulrow CD. Rationale for systematic reviews.
of control, participants and setting. Br. Med. J. 1994; 309: 597–9.
386 The Advanced Handbook of Methods in Evidence Based Healthcare
2. Light RJ, Pillemer DB. Summing up: the science 19. Dickersin K, Min. YI. NIH clinical trials and
of reviewing research. Cambridge, Massachu- publication bias. Online J. Curr. Clin. Trials
setts, and London: Harvard University Press 1993; Doc. No. 50.
1984. 20. Stern JM, Simes RJ. Publication bias: evidence
3. Chalmers I, Altman DG. Systematic Reviews. of delayed publication in a cohort study of
London: BMJ Publishing Group 1995. clinical research projects. Br. Med. J. 1997; 315:
4. Begg CB, Berlin JA. Publication bias: a problem 640–5.
in interpreting medical data. J. Roy. Statist. Soc. 21. Dickersin K. How important is publication bias?
A. 1988; 151: 419–63. A synthesis of available data. Aids Education
5. Song F, Gilbody S. Increase in studies of pub- and Prevention 1997; 9: 15–21.
lication bias coincided with increasing use of 22. Ioannidis J. Effect of the statistical significance
meta-analysis [letter]. Br. Med. J. 1998; 316: of results on the time to completion and publica-
tion of randomized efficacy trials. JAMA 1998;
471.
279: 281–6.
6. Dickersin K. The existence of publication bias
23. Misakian AL, Bero LA. Publication bias and
and risk factors for its occurrence. JAMA 1990;
research on passive smoking. Comparison of
263: 1385–9.
published and unpublished studies. JAMA 1998;
7. Chalmers TC, Frank CS, Reitman D. Minimiz- 280: 250–3.
ing the three stages of publication bias. JAMA 24. Auger CP. Information sources in grey literature,
1990; 263: 1392–5. 4th edn. London, Melbourne, Munich, New
8. Song F, Eastwood A, Gilbody S, Duley L, Providence, N.J: Bowker-Saur, 1998.
Sutton A. Publication and related biases in 25. Smart RG. The importance of negative results in
health techology assessment. Health Technology psychological research. Can. Psychol. 1964; 5:
Assessment 2000; 4(10). 225–32.
9. Sterling TD. Publication decisions and their 26. Smith ML. Publication bias and meta-analysis.
possible effects on inferences drawn tests of Eval. Educ. 1980; 4: 22–4.
significance – or vice versa. Am. Statist. Assoc. 27. Glass GV, McGaw B, Smith ML. Meta-analysis
J. 1959; 54: 30–4. in social research. London: Sage Publications,
10. Sterling TD, Rosenbaum WL, Weinkam JJ. Pub- 1981.
lication decisions revisited – the effect of the 28. White KR. The relation between socioeconomic
outcome of statistical tests on the decision to status and academic achievement. Psychological
publish and vice-versa. American Statistician Bulletin 1982; 91: 461–81.
1995; 49: 108–12. 29. Koren G, Graham K, Shear H, Einarson T. Bias
11. Chalmers TC, Grady GF. A note on fatality in against the null hypothesis: the reproductive
serum hepatitis. Gastroenterology 1965; 49: hazards of cocaine. Lancet 1989; 2: 1440–2.
22–6. 30. Callaham ML, Wears RL, Weber EJ, Barton
12. Greenwald AG. Consequences of prejudice C, Young G. Positive-outcome bias and other
against the null hypothesis. Psychol. Bull. 1975; limitations in the outcome of research abstracts
82: 1–20. submitted to a scientific meeting. JAMA 1998;
13. Coursol A, Wagner EE. Effect of positive find- 280: 254–7.
ings on submission and acceptance rates: a note 31. Scherer RW, Dickersin K, Langenberg P. Full
on meta-analysis bias. Professional Psychology publication of results initially presented in
abstracts. A meta-analysis. JAMA 1994; 272:
1986; 17: 136–7.
158–62.
14. Dickersin K, Chan S, Chalmers TC, Sacks HS,
32. DeBellefeuille C, Morrison CA, Tannock IF.
Smith H, Jr. Publication bias and clinical trials.
The fate of abstracts submitted to a cancer
Controlled Clinical Trials 1987; 8: 343–53.
meeting: factors which influence presentation
15. Sommer B. The file drawer effect and publica- and subsequent publication. Ann. Oncol. 1992;
tion rates in menstrual cycle research. Psychol. 3: 187–91.
Women Q. 1987; 11: 233–42. 33. Chalmers I, Adams M, Dickersin K, et al. A
16. Simes RJ. Publication bias: the case for an cohort study of summary reports of controlled
international registry of clinical trials. J. Clin. trials. JAMA 1990; 263: 1401–5.
Oncol. 1986; 4: 1529–41. 34. Weber EJ, Callaham ML, Wears RL, Barton C,
17. Easterbrook PJ, Berlin JA, Gopalan R, Matthews Young G. Unpublished research from a medical
DR. Publication bias in clinical research. Lancet specialty meeting. Why investigators fail to pub-
1991; 337: 867–72. lish. JAMA 1998; 280: 257–9.
18. Dickersin K, Min YI, Meinert CL. Factors influ- 35. Cheng K, Preston C, Ashby D, OHea U, Smyth
encing publication of research results. Follow up R. Time to publication as full reports of ab-
of applications submitted to two institutional stracts of randomized controlled trials in cystic
review boards. JAMA 1992; 267: 374–8. fibrosis. Pediatr. Pulmonol. 1998; 26: 101–5.
Publication and Related Biases 387
36. Gregoire G, Derderian F, Lorier JL. Selecting 52. Chalmers I. Under-reporting research is scien-
the language of the publications included in a tific misconduct. JAMA 1990; 263: 1405–8.
meta-analysis: is there a tower of babel bias? J. 53. Cowley AJ, Skene A, Stainer K, Hampton JR.
Clin. Epidemiol. 1995; 48: 159–63. The effect of lorcainide on arrhythmias and
37. Egger M, Zellweger-Zahner T, Schneider M, survival in patients with acute myocardial
Junker C, Lengeler C, Antes G. Language bias in infarction – an example of publication bias. Int.
randomised controlled trials published in English J. Cardiol. 1993; 40: 161–6.
and German. Lancet 1997; 350: 326–9. 54. The Cardiac Arrhythmia Suppression Trial
38. Nylenna M, Riis P, Karlsson Y. Multiple (CAST) Investigators. Preliminary report: effect
blinded reviews of the same two manuscripts. of encainide and flecainide on mortality in a
Effects of referee characteristics and publication randomised trial of arrhythmia suppression after
language. JAMA 1994; 272: 149–51. myocardial infarction. N. Engl. J. Med. 1989;
39. Ottenbacher K, Difabio RP. Efficacy of spinal 321: 406–12.
manipulation/mobilization therapy. A meta- 55. The Cardiac Arrhythmia Suppression Trial II
analysis. Spine 1985; 10: 833–7. Investigators. Effect of the antiarrhythmic agent
40. Vickers A, Goyal N, Harland R, Rees R. Do moricisine on survival after myocardial infarc-
certain countries produce only positive results? tion. N. Engl. J. Med. 1992; 327: 227–33.
A systematic review of controlled trials. Con- 56. Yusuf S, Koon T, Woods K. Intravenous mag-
trolled Clinical Trials 1998; 19: 159–66. nesium in acute myocardial infarction: an effec-
41. Pocock SJ, Hughes MD, Lee RJ. Statistical tive, safe, simple, and inexpensive intervention.
problems in the reporting of clinical trials. A Circulation 1993; 87: 2043–6.
survey of three medical journals. N. Engl. J. 57. ISIS–4 Collaborative Group. ISIS–4: a random-
Med. 1987; 317: 426–32. ised factorial trial assessing early oral captopril,
42. Tannock IF. False-positive results in clinical oral mononitrate, and intravenous magnesium
trials: multiple significance tests and the prob- sulphate in 58050 patients with acute myocardial
lem of unreported comparisons. J. Natl Cancer infarction. Lancet 1995; 345: 669–85.
Inst. 1996; 88: 206–7.
58. Egger M, Davey-Smith G. Misleading meta-
43. Forrow L, Taylor WC, Arnold RM. Absolutely
analysis: lessons from ‘an effective, safe, simple’
relative: how research results are summarized
intervention that wasn’t. Br. Med. J. 1995; 310:
can affect treatment decisions. Am. J. Med.
752–4.
1992; 92: 121–4.
59. The Government Statistical Service. Hospital
44. Naylor CD, Chen E, Strauss B. Measured enthu-
episode statistics. Volume 1. Finished consultant
siasm: does the method of reporting trial results
episodes by diagnosis, operation and specialty.
alter perceptions of therapeutic effectiveness?
England: financial year 1993–94. London:
Ann. Intern. Med. 1992; 117: 916–21.
45. Bobbio M, Demichelis B, Giustetto G. Com- Crown Copyright, 1995.
pleteness of reporting trial results: effect on 60. MacCoun R. Biases in the interpretation and use
physicians’ willingness to prescribe. Lancet of research results. Annu. Rev. Psychol. 1998;
1994; 343: 1209–11. 49: 259–87.
46. Tramer MR, Reynolds DJM, Moore RA, 61. Beveridge WIB. The art of scientific investiga-
McQuay HJ. Impact of covert duplicate publica- tion. London: Mercury Books, 1961.
tion on meta-analysis: a case study. Br. Med. J. 62. Rothwell PM, Slattery J, Warlow CP. A system-
1997; 315: 635–40. atic review of the risks of stroke and death
47. Huston P, Moher D. Redundancy, disaggregation, due to endarterectomy for symptomatic carotid
and the integrity of medical research. Lancet stenosis. Stroke 1996; 27: 260–5.
1996; 347: 1024–6. 63. Kupfersmid J, Fiala M. A survey of attitudes and
48. Ravnskov U. Cholesterol lowering trials in behaviors of authors who publish in psychology
coronary heart disease: frequency of citation and and education journals [comments]. Am. Psychol.
outcome. Br. Med. J. 1992; 305: 15–19. 1991; 46: 249–50.
49. Ravnskov U. Quotation bias in reviews of the 64. Kassirer JP, Campion EW. Peer review: crude
diet-heart idea. J. Clin. Epidemiol. 1995; 48: and understudied, but indispensable. JAMA
713–19. 1994; 272: 96–7.
50. Hutchison BG, Oxman AD, Lloyd S. Compre- 65. The Lancet. http://www.thelancet.com.
hensiveness and bias in reporting clinical trials. 66. Abby M, Massey MD, Galandiuk S, Polk HC.
Study of reviews of pneumococcal vaccine Peer review is an effective screening process to
effectiveness. Can. Fam. Physician 1995; 41: evaluate medical manuscripts. JAMA 1994; 272:
1356–60. 105–7.
51. Koren G, Klein N. Bias against negative studies 67. Zelen M. Guidelines for publishing papers on
in newspaper reports of medical research. JAMA cancer clinical trials: responsibilities of editors
1991; 266: 1824–6. and authors. J. Clin. Oncol. 1983; 1: 164–9.
388 The Advanced Handbook of Methods in Evidence Based Healthcare
68. Editor. Manuscript guideline. Diabetologia 1984; 87. Bero LA, Galbraith A, Rennie D. Sponsored
25: 4A. symposia on environmental tobacco-smoke.
69. Taubes G. Epidemiology faces its limits. Science JAMA 1994; 271: 612–17.
1995; 269: 164–9. 88. Barnes DE, Bero LA. Industry-funded research
70. Angell M. Negative studies. N. Engl. J. Med. and conflict of interest: an analysis of research
1989; 321: 464–6. sponsored by the tobacco industry through the
71. Bailar JC, Patterson K. Journal peer review: the center for indoor air research. J. Health Polit.
need for a research agenda. N. Engl. J. Med. Policy Law 1996; 21: 515–42.
1985; 312: 654–7. 89. Cho MK, Bero LA. The quality of drug studies
72. Mahoney MJ. Publication prejudices: an experi- published in symposium proceeding. Ann.
mental study of confirmatory bias in the peer Intern. Med. 1996; 124: 485–9.
review system. Cognitive Ther. Res. 1977; 1: 90. Stelfox HT, Chua G, O’Rourke K, Detsky AS.
161–75. Conflict of interest in the debate over calcium-
73. Ernst E, Resch KL. Reviewer bias – a blinded channel antagonists. N. Engl. J. Med. 1998; 338:
experimental study. J. Lab. Clin. Med. 1994; 101–6.
124: 178–82. 91. CCOHTA. Canadian Coordinating Office For
74. Abbot NC, Ernst E. Publication bias: direction Health Technology Assessment: Annual Report
of outcome less important than scientific quality. 1997–1998. Ottawa, 1998.
Perfusion 1998; 11: 182–4. 92. Millstone E, Brunner E, White I. Plagiarism or
75. Blumenthal D, Causino N, Campbell E, Sea- protecting public health. Nature 1994; 371:
shore K. Relationships between academic insti- 647–8.
tutions and industry in the life sciences – an 93. Simes RJ. Confronting publication bias: a cohort
industry survey. N. Engl. J. Med. 1996; 334: design for meta analysis. Stat. Med. 1987; 6:
368–73. 11–29.
76. Hemminki E. Study of information submitted 94. Kleijnen J, Knipschild P. Review Articles and
by drug companies to licensing authorities. Br. Publication Bias. Arzneim. Forsch./Drug Res.
Med. J. 1980; 280: 833–6. 1992; 42: 587–91.
77. Bardy AH. Report Bias in Drug Research. 95. Easterbrook PJ. Directory of registries of clin-
Therapie 1996; 51: 382–3. ical trials. Stat. Med. 1992; 11: 345–423.
78. Bardy A. Bias in reporting clinical trials. Br. J. 96. Silagy CA. Developing a register of randomised
Clin. Pharmacol. 1998; 46: 147–50. controlled trials in primary care. Br. Med. J.
79. Dent THS, Hawke S. Too soon to market: doc- 1993; 306: 897–900.
tors and patients need more information before 97. Robertson SE, Mayans MV, Horsfall S, et al.
drugs enter routine use. Br. Med. J. 1997; 315: The WHO Global Programme for Vaccines and
1248–9. Immunization Vaccine Trial Registry. Bull.
80. Abraham J, Lewis G. Secrecy and transparency World Health Org. 1997; 75: 295–305.
of medicines licensing in the EU. Lancet 1998; 98. Cochrane Controlled Trials Register. In: The
352: 480–2. Cochrane Library (Issue 1). Oxford: Update
81. Roberts I, Li-Wan-Po A, Chalmers I. Intellectual Software, 1999.
property, drug licensing, freedom of informa- 99. The National Research Register. The NHS R&D
tion, and public health. Lancet 1998; 352: Information Systems Strategy. Leeds, 1998.
726–9. 100. The Current Science Group. Current Controlled
82. Lauritsen K, Kavelund T, Larsen LS, Rask- Trials. http://www.controlled-trials.com.
Madsen J. Withholding unfavourable results in 101. Downs JR, Gotto A, Clearfield M, Gordon D,
drug company sponsored clinical trials. Lancet Manolio T, Goldbourt U, et al. Protocol for a
1987; i: 1091. prospective collaborative overview of all current
83. Rennie D. Thyroid storm. JAMA 1997; 277: and planned randomized trials of cholesterol
1238–43. treatment regimens. Am. J. Cardiol. 1995; 75:
84. Kotelchuk D. Asbestos research: winning the 1130–4.
battle but losing the war. Health/PAC Bulletin 102. Simes RJ. Prospective metaanalysis of
1974; 61: 1–27. cholesterol-lowering studies – the prospective
85. Davidson RA. Source of funding and outcome pravastatin pooling (Ppp) project and the
of clinical trials. J. Gen. Intern. Med. 1986; 1: Cholesterol Treatment Trialists (Ctt) Collabora-
155–8. tion. Am. J. Cardiol. 1995; 76: C122–6.
86. Rochon PA, Gurwitz JH, Simms RW, Fortin PR, 103. Meinert CL. Toward prospective registration of
et al. A study of manufacturer-supported trials clinical trials. Controlled Clinical Trials 1988; 9:
of nonsteroidal anti-inflammatory drugs in the 1–5.
treatment of arthritis. Arch. Intern. Med. 1994; 104. Savulescu J, Chalmers I, Blunt J. Are Research
154: 157–63. Ethics Committees behaving unethically? Some
Publication and Related Biases 389
suggestions for improving performance and data be included in meta analyses? Current con-
accountability. Br. Med. J. 1996; 313: 1390–3. victions and controversies. JAMA 1993; 269:
105. Dickersin K, Garcia Lopez F. Regulatory pro- 2749–53.
cess effects clinical trial registration in Spain. 123. Davey-Smith G, Egger M. Meta-analysis: un-
Controlled Clinical Trials 1992; 13: 507–12. resolved issues and future developments. Br.
106. Chollar S. A registry for clinical trials [news]. Med. J. 1998; 316: 221–5.
Ann. Intern. Med. 1998; 128: 701–2. 124. Rosenthal R. The ‘file drawer problem’ and
107. Freedom of information, when it suits (edi- tolerance for null results. Psychol. Bull. 1979;
torial). Lancet 1998; 352: 665. 86: 638–41.
108. Bardy AH. Freedom of information. Lancet 125. Kennedy E, Song F, Hunter R, Gilbody S.
1998; 352: 229. Risperidone versus ‘conventional’ antipsychotic
109. Newcombe RG. Discussion of the paper by medication for schizophrenia. In: The Cochrane
Begg and Berlin: Publication bias: a problem in Library (Issue 3). Oxford: Update Software,
interpreting medical data. J. Roy. Statist. Soc. A 1998.
1988; 151: 448–9. 126. Evans S. Statistician’s comments on: ‘Fail safe
110. Julian D. Meta-analysis and the meta- N’ is a useful mathematical measure of the
epidemiology of clinical research. Registration stability of results [letter]. Br. Med. J. 1996; 312:
of trials should be required by editors and regis- 125.
tering agencies [letter]. Br. Med. J. 1998; 316: 127. Gleser LJ, Olkin I. Models for estimating the
311. number of unpublished studies. Stat. Med. 1996;
111. Smith R. Beyond conflict of interest: transpar- 15: 2493–507.
ency is the key. Br. Med. J. 1998; 317: 291–2. 128. Duval S, Tweedie R. Trim and fill: a simple
112. Smith R, Roberts I. An amnesty for unpublished funnel plot based method of adjusting for pub-
trials: send us details on any unreported trials. lication bias in meta-analysis. Am. Statist. Assoc.
Br. Med. J. 1997; 315: 622. J. (in press).
129. Egger M, Davey-Smith G, Schneider M, Minder
113. Horton R. Medical Editors Trial Amnesty. Lancet
C. Bias in meta-analysis detected by a simple,
1997; 350: 756.
graphical test. Br. Med. J. 1997; 315: 629–34.
114. Fox T. Crisis in communication: the functions
130. Begg CB, Mazumdar M. Operating character-
and future of medical journals. London: The
istics of a rank correlation test for publication
Athlone Press, 1965.
bias. Biometrics 1994; 50: 1088–101.
115. Huth EJ. Electronic publishing in the health
131. Galbraith RF. A note on graphical presentation
sciences. Bull. PAHO 1995; 29: 81–7.
of estimated odds ratios from several clinical
116. Berlin JA. Will publication bias vanish in the
trials. Stat. Med. 1988; 7: 889–94.
age of online journals? [editorial]. Online J. 132. Naylor CD. Meta-analysis and the meta-
Curr. Clin. Trials 1992; Jul 8; Doc. No. 12. epidemiology of clinical research: meta-analysis
117. Song F, Eastwood A, Gilbody S, Duley L. The is an important contribution to research and
role of electronic journals in reducing publica- practice but it’s not a panacea. Br. Med. J. 1997;
tion bias. Med. Inform. 1999; 24: 223–9. 315: 617–19.
118. Chalmers I, Altman DG. How can medical 133. Hedges LV. Estimation of effect size under
journals help prevent poor medical research? nonrandom sampling: the effects of censoring
Some opportunities presented by electronic pub- studies yielding statistically insignificant mean
lishing. Lancet 1998; 353: 490–3. differences. J. Education. Stat. 1984; 9: 61–85.
119. Hetherington J, Dickersin K, Chalmers I, 134. Iyengar S, Greenhouse JB. Selection models and
Meinert CL. Retrospective and prospective iden- the file drawer problem. Stat. Sci. 1988; 3:
tification of unpublished controlled trials: lessons 109–35.
from a survey of obstetricians and pediatricians. 135. Hedges LV. Modelling publication selection
Pediatrics 1989; 84: 374–80. effects in meta-analysis. Stat. Sci. 1992; 7:
120. McManus R, Wilson S, Delaney B, et al. 246–55.
Review of the usefulness of contacting other 136. Dear HBG, Begg CB. An approach for assessing
experts when conducting a literature search for publication bias prior to performing a meta-
systematic reviews. Br. Med. J. 1998; 317: analysis. Stat. Sci. 1992; 7: 237–45.
1562–3. 137. Vevea JL, Hedges LV. A general linear model
121. Chalmers TC, Levin H, Sacks HS, Reitman D, for estimating effect size in the presence of
Berrier J, Nagalingam R. Meta-analysis of clin- publication bias. Psychometrika 1995; 60:
ical trials as a scientific discipline. I: control of 419–35.
bias and comparison with large co-operative 138. Cleary RJ, Casella G. An application of Gibbs
trials. Stat. Med. 1987; 6: 315–25. sampling to estimation in meta-analysis:
122. Cook DJ, Guyatt GH, Ryan G, Clifton J, Buck- accounting for publication bias. J. Education.
ingham L, Willan A, et al. Should unpublished Behav. Stat. 1997; 22: 141–54.
390 The Advanced Handbook of Methods in Evidence Based Healthcare
139. Givens GH, Smith DD, Tweedie RL. Publica- 148. Moher D, Dulberg CS, Wells GA. Statistical
tion bias in meta-analysis: a Bayesian data- power, sample size, and their reporting in ran-
augmentation approach to account for issues domized controlled trials. JAMA 1994; 272:
exemplified in the passive smoking debate. Stat. 122–4.
Sci. 1997; 12: 221–50. 149. Moscati R, Jehle D, Ellis D, Fiorello A, Landi
140. Copas J. What works? selectivity models and M. Positive-outcome bias: comparison of emer-
meta-analysis. J. Roy. Statist. Soc. A 1998; 162: gency medicine and general medicine literatures.
95–109. Acad. Emerg. Med. 1994; 1: 267–71.
141. Hedges LV, Vevea JL. Estimating effect size 150. Moher D. Publication bias (letter). Lancet 1993;
under publication bias: Small sample properties 342: 1116.
and robustness of a random effects selection 151. Bailar JC. Discussion of the paper by Begg and
model. J. Education. Behav. Statist. 1996; 21: Berlin: Publication bias: a problem in inter-
299–332. preting medical data. J. Roy. Statist. Soc. A
142. Linde K, Clausius N, Ramirez G, et al. Are the 1988; 151: 451.
clinical effects of homoeopathy placebo effects? 152. Simon R. Discussion of the paper by Begg and
A meta-analysis of placebo-controlled trials.
Berlin: Publication bias: a problem in inter-
Lancet 1997; 350: 834–43.
preting medical data. J. Roy. Statist. Soc. A
143. Begg CB. Publication bias. In: Cooper H, Hedge
1988; 151: 459–60.
LV (eds). The Handbook of Research Synthesis.
153. de-Melker HE, Rosendaal FR, Vandenbroucke
New York: Russell Sage Foundation, 1994.
JP. Is publication bias a medical problem? [letter].
144. Peto R, Collins R, Gray R. Large-scale random-
ized evidence: large, simple trials and overviews Lancet 1993; 342: 621.
of trials. J. Clin. Epidemiol. 1995; 48: 23–40. 154. Lane DM, Dunlap WP. Estimating effect size:
145. Villar J, Carroli G, Belizan JM. Predictive abil- bias resulting from the significance criterion in
ity of meta-analyses of randomised controlled. editorial decisions. Br. J. Math. Statist. Psychol.
Lancet 1995; 345: 772–6. 1978; 31: 107–12.
146. Cappelleri JC, Ioannidis JPA, Schmid CH, 155. Berlin JA, Begg CB, Louis TA. An assessment
Ferranti SDd, Aubert M, Chalmers TC, et al. of publication bias using a sample of published
Large trials vs meta-analysis of smaller trials. clinical trials. J. Am. Stat. Assoc. 1989; 84:
How do their results compare? JAMA 1996; 276: 381–92.
1332–8. 156. Begg CB, Berlin JA. Publication bias and dis-
147. LeLorier J, Gregoire G, Benhaddad A, Lapierre semination of clinical research. J. Natl Cancer
J, Derderian F. Discrepancies between meta- Inst. 1989; 81: 107–15.
analyses and subsequent large randomized, con- 157. Irwig L, Macaskill P, Glasziou P, Fahey M. Meta-
trolled trials. N. Engl. J. Med. 1997; 337: analytic methods for diagnostic test accuracy. J.
536–42. Clin. Epidemiol. 1995; 48: 119–30.
22
Meta-Analysis in Health Technology
Assessment
stages of carrying out a systematic review. The these are mentioned briefly below, and further
whole process is however, outlined here in a set details given in Sutton et al.2 Here, a practical
of guidelines (Figure 1). For the most part these approach is adopted throughout, with algebra
follow standard and widely agreed guidelines and formulae kept to a minimum (though refer-
published previously.3–5 The reader is recom- ences are given to where technical details can
mended to consult these3,4 for detailed descrip- be obtained). Fortunately, software is becom-
tions of the procedural stages required in carry- ing available to implement all the methods
ing out a high-quality systematic review. After a described below, much of it freely available.
brief description of the project methods, it is A detailed review of this topic has been pub-
meta-analytic methods (or quantitative synthesis lished elsewhere,7 and software is available via
– described in point 6 of the recommendations the internet at http://www.prw.le.ac.uk/epidemio/
in Figure 1) on which we focus our attention for personal/kra1/srotos.html; the reader is referred
the remainder of this chapter. The review identi- to these for details of practical implementation.
fied a considerable body of work on this topic, All the methods described below focus on
with increasingly sophisticated methods being pooling treatment estimates from different
developed over time. This is an area where studies. This yields an overall pooled estimate of
previous guidelines have been less detailed. In treatment effect, together with a confidence
addition to development of general methods, interval. Methods which only tally the numbers
much research has been carried out which of studies showing beneficial and harmful
focuses on meta-analysis of specific outcome effects, known as vote-counting, and ones which
measures and data types, as well as other prob- pool the P-values of significance tests from each
lems researchers face at the meta-analytic stage study also exist.8 Neither of these methods pro-
of a systematic review. duces an overall pooled estimate and, due to this
and other drawbacks, are not recommended for
routine use and are not considered further here.
EXAMPLE: EFFECT ON MORTALITY OF
LOWERING SERUM CHOLESTEROL LEVELS Binary Outcomes
A meta-analysis of cholesterol-lowering trials is When outcome data are binary in form, several
used to illustrate the methods throughout the different measures of treatment effect may be
chapter. Since 1962 a series of randomised con- used (and are used in varying degrees of popu-
trolled trials (RCTs) have investigated the effect larity); these are described briefly below. Firstly,
of lowering cholesterol levels on the risk from it is convenient to arrange the data from each
all-cause mortality, as well as specific causes. trial into the form of tables, like the one illus-
Our example draws on a data set consisting of trated below in Table 1.
34 RCTs, originally analysed by Smith et al.6 For example, cholesterol trial 16 (see the first
(see this paper for a listing of these trials; line of Table 2) takes values: a 5 174; b 5 250;
numbering in this chapter corresponds to theirs, c 5 178; and d 5 244.
although knowledge of this is not necessary). It
should be noted that since this review was The odds ratio
completed, further relevant trials have been car-
The odds ratio can then be calculated by the
ried out. It therefore should be stressed that the
ad
analyses presented are to illustrate the various formula . This measure gives a relative meas-
methods discussed, and are in no way meant bc
to provide a definitive analysis. The binary ure of risk in the form of the ratio of the odds of
outcome used in the primary studies is total an event in the two groups.9 An odds ratio of
mortality. less than 1, when comparing a new treatment to
the control, would indicate an improvement on
the new treatment; while a ratio greater than 1
AN INTRODUCTION TO META-ANALYTIC would imply that the new treatment was less
effective than any treatment received by the
METHODS INCORPORATING FIXED EFFECTS control group.
MODELS
Table 1 Outcome data from a single study
This section considers methods for combining
binary outcome measures, which are probably Failure/death Success/alive
the most common type of outcome used when New treatment a b
assessing medical interventions. However, con- Control c d
tinuous and ordinal outcomes are also used;
394 The Advanced Handbook of Methods in Evidence Based Healthcare
Î Î
When deciding on the choice of measure to
summarise studies, two opposing considerations 1 1
are present: (i) whether it is statistically con- T 2 1.96 3 # u # T 1 1.96 3 , (4)
Ow Ow
k k
venient to work with; and (ii) whether it conveys
the necessary clinically useful information.13 i i
i51 i51
The number needed to treat is a measure that is
being used increasingly when reporting the where u is the true treatment effect. On the
results of clinical trials. The motivation for its mathematical level, this is essentially all that is
use is that it is more useful than the odds ratio and required to synthesise study treatment effects at
the relative risk for clinical decision-making.11 a basic level, though other ‘real world’ issues
However, its role as a measure of comparative need considering, outlined later and in the
treatment effects is limited.14 Detailed discus- guidelines. For a fuller coverage of this method
sions of the relative merits of using the other see Fleiss,12 or Shadish et al.18
scales have been reported.10,13 Historically, the
odds ratio is most commonly reported for RCTs, Pooling the cholesterol data using the
while relative risks are often used for reporting
observational studies. One should be aware that
inverse-variance weighted method
while analyses using different scales will often For this illustrative analysis, only a subset of the
come to the same conclusion, in some instances 34 RCTs will be used. The subset of trials
considerably different results will be obtained.15 chosen consists of those in which patients were
Meta-Analysis in Health Technology Assessment 395
largely without pre-existing cardiovascular dis- a 95% confidence interval for this pooled esti-
ease, so that cholesterol lowering was imple- mate using Equation (4), and converting back
mented predominantly as a primary intervention to the odds ratio scale gives: OR 5 1.09 (0.98
(identification (id.) numbers 16, 20, 24, 28, 29, to 1.21).
30 and 31). For clarity, data from these seven The results of this analysis are displayed
trials is reproduced in Table 2. graphically in Figure 2. This is a very common
The odds ratio (OR) for each study is calcu- way of displaying the results of a meta-analysis,
lated (e.g. study id. 16: OR 5 (174 3 244)/(250 sometimes known as a forest plot. Each study’s
3 178) 5 0.954). As previously mentioned, it is point estimate, together with its 95% CI, is
advisable to log transform the data before com- displayed. The size of the box representing each
bining (e.g., id. 16: LOR 5 20.051). The vari- point estimate is inversely proportional to the
ance of LOR is calculated (e.g., id. 16: Var(LOR) variance of the study, and hence proportional to
5 1/250 1 1/174 1 1/244 1 1/178 5 0.0195). the weight of that study in the analysis. Below
Confidence intervals for LOR can be constructed the estimates of each study is the pooled esti-
(e.g., id. 16: 20.051 6 1.96 Ï 0.0195 5 [20.31 mate together with its 95% confidence interval.
to 0.22]), and transforming back to the odds Results using two other methods are also dis-
ratio scale (e.g., id. 16: [e(20.31) to e(0.22)] 5 [0.73 played in this figure. These are described later.
to 1.25]). The combined odds ratio is slightly greater
The eighth column of Table 2 displays the than 1. However, because its corresponding con-
point estimates along with their 95% CI for all fidence interval includes one, a conclusion that
seven studies. It can be seen from Table 2 that no evidence of a treatment effect exists is drawn
three of the point estimates are below one, from the combined results of the seven studies.
indicating a reduced risk for people on the The confidence interval around this estimate is
treatment (cholesterol-reducing) and four are narrow due to the large numbers of patients/
above one indicating an increased risk for the events combined, and only just crosses unity.
treatment. However, with the exception of that From this illustrative analysis, the possibility
that cholesterol-lowering treatment may actually
for study 31, every confidence interval includes
be harmful as a primary intervention cannot be
one. From this, we conclude that no significant
ruled out.
treatment effect was detected when considering
each study separately. The 95% CI of study 31
spans 1.07 to 1.59, indicating evidence of a Other fixed effect methods for combining
statistically significant increased risk for the odds ratios
patients in the treatment arm. By combining Other fixed effect methods, specific to combin-
these studies it is hoped that an estimate which ing odds ratios, have been proposed. Under most
is more generalisable than any of the individual conditions the estimates obtained from each
study results (not least because studies using method should be very similar to one another,
different populations are being combined), and but there are circumstances where they differ
more precise (through reducing the effect of markedly.
sampling variations in estimates based on A common alternative to the above is the
smaller numbers) can be produced. Mantel–Haenszel method;20 the use of which for
The weighting given to each study (wi) needs meta-analysis has been described previously.12
to be calculated; using the inverse variance- A variance estimate of the treatment effect cal-
weighted method, this is simply equal to culated in this manner has been derived.21,22
1/Var(LOR) (e.g., for study 16: wi 5 1/0.0195 Using the Mantel–Haenszel estimate, a study
5 51.38). Column 11 of Table 2 shows the with zero total events is completely excluded
values for the rest of the studies. from the analysis if no continuity correction is
Combining the results using formula, [Equa- used. This is unappealing, as a trial with zero
tion (2)], gives a point estimate of the log odds events from 200 subjects would then be equally
ratio: LOR 5 non-informative as a trial with no events but
[(51.38 3 20.05) 1 ...... 1 (98.489 3 0.27)] only 20 subjects. To get round this problem it
(51.38 1 ...... 1 98.48) has been recommended23 that a continuity cor-
5 0.085. An estimate of its variance is given rection (adding 0.5 to each cell) should be used
by the reciprocal of the sum of the weights for sparse data. Another common method, first
described by Peto et al.,24 and more thoroughly
O w ); hence the standard error (square
k
(i.e. 1y i
by Yusuf et al.,25 can be regarded as a modifica-
i51 tion of the Mantel–Haenszel method. An advant-
root of the variance) is given by: SE(LOR) 5 age it has over the latter method is that it can be
Ï 1 / (51.38 1 ..... 1 98.48) 5 0.054. Calculating used, without any corrections, if there are zeros
396
The Advanced Handbook of Methods in Evidence Based Healthcare
Table 2 Results of the seven primary studies investigating the effect of cholesterol lowering on mortality
No of No of No of deaths No of deaths No still alive No still alive Log odds
Study subjects in subjects in in treatment in the control in treatment in the control Estimate of odds ratio ratio var(LOR)
id. treatment control arm arm arm arm (a 3 d)/(b 3 c) 5 (LOR) 5 5 1/a 1 1/b w5
number arm (nt) arm (nc) (dt) 5 a (dc) 5 c (nt 2 dt) 5 b (nc 2 dc) 5 d OR [95% CI] ln(OR) 1 1/c 1 1/d 1/var(LOR)
16 424 422 174 178 250 244 0.95 [0.73, 1.25] 20.051 0.0195 51.38
20 1149 1129 37 48 1112 1081 0.75 [0.48, 1.16] 20.288 0.0497 20.13
24 4541 4516 269 248 4272 4268 1.08 [0.91, 1.29] 0.077 0.0082 121.68
28 1906 1900 68 71 1838 1829 0.95 [0.68, 1.34] 20.051 0.0300 33.47
29 2051 2030 44 43 2007 1987 1.01 [0.66, 1.55] 0.010 0.0465 21.28
30 6582 1663 33 3 6549 1660 2.79 [0.85, 9.10] 1.026 0.3644 2.74
31 5331 5296 236 181 5095 5115 1.31 [1.07, 1.59] 0.270 0.0102 98.48
Meta-Analysis in Health Technology Assessment 397
present in cells for any of the studies. Combin- have been described previously27 for which
ing the cholesterol studies using these two meth- the inverse-weighted and the Mantel–Haenszel
ods produces virtually the same results as the methods are to be preferred respectively. If the
inverse variance-weighted method. number of studies to be combined is small, but
Having a number of different approaches to the within-study sample sizes per study are
combine odds ratios at the researchers disposal, large, the inverse-weighted method should be
it would be desirable to have guidelines indicat- used. If one has many studies to combine, but
ing when particular methods are most appro- the within-study sample size in each study is
priate. The Peto method has come under strong small, the Mantel–Haenszel method is preferred.
criticism as it has been demonstrated that it is Comparisons between the various methods have
capable of producing seriously biased odds been carried out.26 Other maximum likelihood
ratios and corresponding standard errors when and exact methods which are available28–30 are
there is severe imbalance in the numbers in the only worth considering if cells from several
two groups being compared,26 or when the esti- studies have less than five observations, due to
mated odds ratio is far from unity.12 Conditions the much increased computation involved.
398 The Advanced Handbook of Methods in Evidence Based Healthcare
Combining Outcomes Measured on a alone. Clearly, in many instances this may not
Continuous Scale seem realistic, possibly due to qualitative differ-
ences between the studies, e.g. differences in the
Many of the continuous scales used to measure way an intervention is administered, or differ-
outcomes in the medical literature such as lung ences between study populations. However, the
function, pulse rate, weight and blood pressure decision will not always be clear-cut, and for
have the commonality that they are measured on this reason a formal test of heterogeneity is used
a positive scale. For this reason it is possible – as a guide to when the use of a fixed effect
and common practice – to use a logarithmic model is appropriate.
transformation, after which the data are usually
assumed to be normally distributed (a condition Test for Heterogeneity
which is required for statistical inferences).30
Usually, the value of interest is the difference in The test devised by Cochran,36 which is widely
effect size between the treatment and control used, is given below (5). It tests the assumption
groups. If it can be assumed that all the studies that all the studies are estimating a single under-
estimate the same parameter, and that the esti- lying effect size, and hence that the variation
mates of continuous outcome measures are observed between study results is due to chance
approximately normal, then the inverse variance- alone. The test involves calculating
O w (T 2 T ) ,
weighted method can be used directly, combin- k
ing the data in their original metric. If different Q5 i i
2
(5)
studies measured their outcomes on different i51
scales then synthesis is still possible, but the
data must first be standardised. This can be where k is the number of studies being com-
achieved by dividing the difference of interest bined, Ti is the treatment effect estimate in the
by an estimate of its standard deviation; possible ith study, and T is the inverse variance-weighted
estimates of this standard deviation are dis- estimator of treatment effect (as calculated
cussed by Hedges and Olkin.8 However, it above), and wi is the weight of that study
should be noted that by doing this the resulting (usually the inverse of the ith sampling vari-
estimate may be difficult to interpret clinically. ance) in the meta-analysis.
If the data are censored in any way, as is often Q is approximately distributed x2 on k 2 1
the case for survival data, then special methods degrees of freedom under the null hypothesis,
are needed.2,31–33 which is to say, when heterogeneity is not
present. Hence, if the value for Q exceeds the
upper-tail critical value of chi-square distribu-
Combining Outcomes Measured on tion with k 2 1 degrees of freedom, the observed
Ordinal Scales variance in study effect sizes is significantly
greater than that expected by chance if all studies
If the outcome of interest is measured on a are estimating a common population effect size.
categorical scale and ordered in terms of desir- Thus, one concludes that an assumption of
ability, then one can consider the data as ordinal, homogeneity is not appropriate.
e.g. gastrointestinal damage is often measured Choosing an appropriate critical value for this
by the number of lesions.34 Two situations exist test is made difficult because of its low statis-
when combining ordinal data from different tical power.37 This means heterogeneity may
studies: (i) when the response variable (the out- indeed be present even if the Q statistic is not
come of interest) is the same in each study; and statistically significant. As a consequence, using
(ii) when different response variables are used in a cut-off significance level of 0.10, rather than
the studies. A unified framework has been de- the usual 0.05, has been recommended.38 This
scribed that can incorporate both possibilities.35 has become a customary practice in meta-
analysis. The potential presence of publication
biases (see later and Chapter 21) makes the inter-
pretation of heterogeneity tests more complex.39
ASSESSING HETEROGENEITY In addition to the formal test, several plots
have been advocated to assist the exploration of
Using a fixed effects model to combine treatment heterogeneity in meta-analysis; these include
estimates assumes no heterogeneity between the a radial plot,40 L’Abbé plot,41,42 and a plot of
results of the studies; that is to say, the studies normalized (z) scores43 as well as the forest
are all estimating a single ‘true’ value under- plot.
lying all the study results. Hence, all observed Due to the limitations of the test, and because
variation in the treatment effects between the it is unrealistic for there to be no between study
studies is considered due to sampling error heterogeneity (all studies vary in at least some
Meta-Analysis in Health Technology Assessment 399
respects), it has been suggested that rather than each study is regarded as coming from a popula-
testing for heterogeneity, quantifying estimates tion of similar studies, a concept some people
of the between-study variability and incorporat- feel is unrealistic, though random effects models
ing these into the model is a more sensible ap- have a long history in other fields of application.
proach.44 In this way, some of the heterogeneity Hence, the total variation of the esti-
between studies is accounted for when infer- mated effect size can be broken down into two
ences about the intervention effect are being components:
made. This is addressed below using random Variance of 5 Between-study 1 Within-study
effects models. estimated random effects estimation
effects variance variance
Cholesterol data example
If the random effects variance were zero, the
Applying the test to the seven primary choles- above model would reduce exactly to the fixed
terol studies produces a value of 10.1854 for Q. effects model. Details of estimation of the
This is compared to a chi-squared statistic on six underlying mean effects size and variance com-
(n 2 1 studies) degrees of freedom, giving a P- ponents parameters is omitted, but have been
value of 0.117. In this case P . 0.1, and is described clearly elsewhere.18,46 Computer code
therefore non-significant, but the result is is recommended for their practical implementa-
marginal. It is clearly possible that study results tion.7 Random effects models can also be
may vary by a greater amount than chance alone derived using Bayesian methods, an approach
would permit, and hence it is wise to consider which is described later in this chapter.
using a method that would take this extra vari-
ation into account. These are described below.
Pooling the Cholesterol Data Example
Using a Random Effects Model
RANDOM EFFECTS MODELS The same studies used previously are combined
below using a random effects model. Table 3
Random effects models can be used when the shows the weightings given to each study in the
assumption that all the studies are estimating the fixed and random effects models.
same underlying effect size is violated, and It is instructive to examine how these relative
heterogeneity is present. Many meta-analysts weightings differ between models. Using the
regard the random effects model as always more random effects model, the larger studies have
appropriate than the fixed effects model for been down-weighted while the relative weight-
combining studies, regardless of the result of the ing of the smaller studies is increased. This
test for heterogeneity.45 trend generally holds true as the within-study
This model takes into account an extra vari- weights are ‘diluted’ by the inclusion of the
ation component implied by assuming that the between study term.
studies are estimating different (underlying) Using the random effects model produces a
effect sizes. In addition, the variation caused by pooled point estimate odds ratio of 1.06, and
sampling error described in the fixed effects associated 95% confidence interval (0.91 to
model is still present. These underlying effects 1.24). The model estimates t2, the between-
are assumed to vary at random; to make model- study variance, at 0.014.
ling possible, they are assumed to vary according A plot of the pooled results is included in
to a given distribution (usually the Normal).46 Figure 2. This result can be compared with those
This assumption has caused much dispute, since obtained from fitting a fixed effects model.
There, the point estimate was 1.09, slightly provides the researcher with a tool for exploring,
higher than that of the random effects model and ultimately explaining, heterogeneity between
above (1.06). Comparing confidence intervals, studies by including study level covariates in the
using a fixed effects model gives 0.98 to 1.21, analysis. Investigating why study results vary
whilst using a random effects model gave 0.91 systematically can lead to discovery of associ-
to 1.24. The random effects derived interval is ations between study or patient characteristics
thus wider, incorporating both higher and lower and the outcome measure, not possible in the
values than that of the corresponding fixed analysis of single studies (not least due to lack
effects one. This will generally hold true; the of power). This in turn can lead to clinically
random effects confidence interval will be important findings and may eventually assist in
wider, reflecting the inclusion of between study individualising treatment regimes.53 Covariates
heterogeneity. Although the overall conclusion may characterise study design features, e.g. the
is similar to that given earlier by the fixed dose level of a treatment, or may represent the
effects model – the treatment effect is non- mean of characteristics of patients, e.g. patient
significant – the result is more conservative. age in the study.
These modelling techniques can be used for
the synthesis of randomised clinical trials or
Extensions to the Random Effects Model observational studies (for examples of its use,
see Dickersin and Berlin54). In both cases, dose–
One drawback of this model is that the sampling response modelling is possible.55,56 Modelling of
variance is assumed to be known, although in effects of the underlying risk of patients is also
reality it is estimated from the data,46 which feasible, but specialist methods are required,57–60
means that the uncertainty associated with this due to the underlying problem of regression to
estimation is not taken into account. Two recent the mean.61 If regression to the mean is not
advances have attempted to address this prob- accounted for properly, it will cause an exag-
lem,47,48 both of which produce confidence inter- gerated relationship between baseline risk and
vals which are wider than those by the tradi- the outcome to be observed.
tional model described above, when the number It should be stressed, however, that this type
of studies being combined is small. A further of analysis should be treated as exploratory, as
extension of the random effects model is a associations between characteristics and the out-
likelihood approach which avoids the use of come can occur purely by chance, or may be due
approximating Normal distributions, and can be to the presence of confounding factors. Regres-
used when the assumptions of normality are sion analysis of this type is also susceptible to
violated.49 ecological bias, which occurs if the relation
between study means and outcomes do not cor-
Debate over Choice Between Fixed respond to the relations between individuals
Versus Random Effects Models values and individuals outcomes.62 Regression-
type models are most useful when the number
The argument over which model is theoretically of studies is large, and cannot be sensibly
superior has been running for many years, with attempted when very small numbers of studies
many comments scattered throughout the litera- are being combined.51 A further restriction is
ture. Random effects models have been criticised that the data on covariates required may be
on grounds that unrealistic/unjustified distribu- limited in study reports.
tional assumptions have to be made.50 However, Two main types of regression model are pos-
it has also been argued that they are consistent sible: one an extension of the fixed effects model,
with the aims of generalisation of results of the commonly known as a meta-regression model;
meta-analysis.51 A further consideration is that and the other an extension of the random effects
random effects models are more sensitive to model, called a mixed model (because it in-
publication bias.52 Thus, neither fixed nor ran- cludes both fixed and random terms). The meta-
dom effect analyses can be considered ideal.45 regression model is most appropriate when all
As the next section discusses, it is better to variation (above that explainable by sampling
explain the underlying causes of heterogeneity, error) between study outcomes can be con-
if possible. sidered accountable by the covariates included.
A mixed model is appropriate when the pre-
dictive covariates only explain part of the
variation/heterogeneity, as the random effect
META-REGRESSION AND MIXED MODELS term can account for the remainder. However, in
parallel to earlier argument concerning fixed and
Random effects models account for hetero- random effects models, it has been suggested
geneity between studies, while meta-regression that one should always include a random effects
Meta-Analysis in Health Technology Assessment 401
term as there will always be some degree of Table 4 Data used in the regression and
unexplainable heterogeneity.63 mixed modelling examples
Several authors have put forward extensions/
alternatives to the basic model.64,65 Further Log odds Cholesterol
extensions of mixed models are discussed in the Study id. ratio variance reduction
Generalized Synthesis of Evidence (p. 403) and number (LOR) of LOR (mmol/l)
Bayesian (p. 402) methods sections below. 1 20.71 0.073 0.4
7 20.38 0.059 1.1
8 20.13 0.124 0.6
Fixed Effects Regression Model
9 20.003 0.023 0.3
16 20.23 0.049 0.9
A general meta-regression model, which can
17 20.03 0.090 1.0
include continuous and discrete predictor vari- 21 0.42 0.076 0.3
ables, is clearly explained by Hedges et al.,66 22 1.13 0.845 0.6
and is briefly summarised below. Taking the 24 0.12 0.035 0.7
fixed effects model as a starting point; suppose
now, that there are p known covariates X1, . . .,
Xp which are believed to be related to the effects
via a linear model. Let xi1, . . ., xip be the values cholesterol reduction coefficient is negative,
of the covariates X1, . . ., Xp for the ith study and suggesting that trials in which cholesterol was
b0, b1, . . ., bp be the unknown regression co- lowered by greater amounts showed larger treat-
efficients, to be estimated, indicating the relation- ment benefits. However, the improvement in the
ship between its associated covariate and the model by including such a term is far from
outcome. Then the b0 1 b1xi1 1 . . . 1 bpxip esti- statistically significant (P 5 0.59), and certainly
mates the mean effect in the ith study. further research would be needed to confirm
The coefficients in the model are calculated such a result. (Thompson,45 analysing a different
via weighted least squares algorithms; any set of cholesterol trials, and using a different
standard statistical package that performs outcome, did establish a significant relation-
weighted (multiple) regression can be used.7 ship.) The trials, together with the fitted line, are
When binary outcomes are being combined, a displayed in Figure 3, where the size of the
possible alternative to the above is to use logis- circular plotting symbol is proportional to the
tic regression. This allows the outcome variables size of the trial.
to be modelled directly (on the odds ratio scale),
and removes the need for the normality assump-
tion (see Sutton et al.7 and Thompson45 for Mixed Models (Random-Effects
details). Regression)
Cholesterol data example A full description of mixed models for meta-
Returning to the cholesterol-lowering trials, one analysis follows a logical extension of the
covariate that the dataset includes is absolute regression model given above, and has been
cholesterol reduction. This is a measure of the given elsewhere.51,63 Their implementation is
difference in cholesterol reduction, on average, slightly more complex, but appropriate software
between the two arms in the RCTs. It seems is available.7
reasonable to speculate that the trials in which a
greater degree of lowering occurred may have Cholesterol data example
achieved a more favourable treatment effect and
hence produced a lower odds ratio estimate. The analysis carried out in the previous section
This can be investigated using regression meth- using a fixed effects meta-regression model,
ods. The analysis is restricted to those nine trials may be repeated, this time using a mixed model.
in which diet was the main intervention (i.e. not The model fitted is: ln(OR) 5 0.081 2 0.239 3
the same ones as used before), and the relevant (absolute cholesterol reduction), which has co-
data are provided in Table 4. A fixed effects efficients that are only marginally different from
regression model, including absolute cholesterol those of the previous model. The standard error
reduction as a single covariate yielded the follow- terms for the constant and cholesterol reduction
ing model: ln(OR) 5 0.034 2 0.208 3 (absolute coefficients are 0.27 and 0.38 respectively,
cholesterol reduction). The standard error terms which are marginally larger than before due
for the constant and cholesterol reduction co- to the incorporation of a between-study vari-
efficients are 0.26 and 0.36 respectively. The ance term (estimate 5 0.052). Again, the co-
402 The Advanced Handbook of Methods in Evidence Based Healthcare
Figure 3 Regression plot of cholesterol-lowering trials using diet: log odds ratio
against absolute reduction in cholesterol (between arms), together with regression
line.
efficient for cholesterol reduction is not statistic- hierarchical models have been used for some
ally significant. considerable time.77,78 Whilst the underlying
structure of such models is similar to that
assumed for the random effects model described
BAYESIAN METHODS IN META-ANALYSIS previously, the Bayesian formulation requires
that prior distributions are specified for the mean
Over the past few years, Bayesian methods have effect size, and the between- and within-study
become more frequently used in a number of variances. In estimating the parameters of such
areas of healthcare research, including meta- models a distinction should be drawn between
analysis.67–69 Though much of this increase in empirical and fully Bayesian methods: a fully
their use has been directly as a result of advances Bayesian approach specifies the prior distribu-
in computational methods, it has also been tions completely, whereas empirical Bayesian
partly due to their more appealing nature, and methods estimate such prior distributions from
also specifically the fact that they overcome the data.79 In a meta-analysis setting, empirical
some of the difficulties encountered by other Bayes approaches have received much attention
methods traditionally used. Philosophical and in the literature, as until recently the use of a
practical issues regarding their application in fully Bayesian approach has been hampered by
a number of areas in HTA are discussed in computational difficulties.80–86
Chapter 16.
Many of the authors who have considered a
Bayesian approach to meta-analysis have imple- Choice of Prior Distributions
mented a hierarchical model, frequently also
making an assumption of Normality at one or When adopting a fully Bayesian approach con-
more levels of the hierarchy.70–75 For an up-to- siderable flexibility exists as to the specification
date review and exposition of the Bayesian of the prior distributions for the mean effect and
approach to meta-analysis, see Hedges et al.76 In the between- and within-study variances, though
other areas of statistical science such Bayesian this is a non-trivial task, and the specific choice
Meta-Analysis in Health Technology Assessment 403
of prior distributions has recently received con- more general populations, may indicate better its
siderable attention.87,88 likely effectiveness in practice.
Having specified the prior distributions in In 1992, the Program Evaluation and Method-
terms of the relevant model parameters, estima- ology Division of the U.S. General Accounting
tion can then proceed using one of a number Office (GAO)93 reported an approach for syn-
of computational approaches. However, the thesising evidence from different sources. They
assumption of Normality that is frequently made called this new method cross-design synthesis,
– combined with the fact that there are often a though they cite the confidence profile method
reasonable number of studies in any specific of Eddy et al.94 as an influence. Their idea was
meta-analysis – make such models particularly to create methodology for a new form of meta-
suited to Markov Chain Monte Carlo (MCMC) analysis that aims to capture the strengths of
methods.89 In particular, Gibbs sampling has multiple-study designs (and the evidence they
been increasingly used in applied Bayesian analy- include) whilst minimising their weaknesses.
ses within a healthcare research setting68,89,90 and Operational methodology, in the spirit of that
a specific package BUGS91 has been developed laid out by the GAO, has been developed by
to facilitate its implementation.
Smith et al.87 This combines studies with dis-
parate designs into a single synthesis, using a
Cholesterol data example Bayesian hierarchical model approach. The hier-
A fully Bayesian analysis is carried out on the archical nature of the model specifically allows
cholesterol data in Table 2. Gibbs sampling, for the quantitative within- and between-sources
implemented via BUGS, was used for the analy- heterogeneity, whilst the Bayesian approach can
sis. Details of how this is done have been given accommodate a priori beliefs regarding qual-
elsewhere.74 The posterior mean for the odds itative differences between the various sources
ratio was estimated to be 1.07 with an associated of evidence. Prior distributions may represent
95% Credibility Interval (a Bayesian measure of subjective beliefs elicited from experts, or they
uncertainty similar to a Classical confidence might represent other data based evidence,
interval) from 0.89 to 1.25. This is plotted at the which though pertinent to the issue in question
bottom of Figure 2. Similarly, the posterior is not of a form that can be directly incorpor-
mean for the between-study variance was esti- ated, such as data from animal experiments.95
mated to be 0.0241 with an associated 95% This model can be viewed as an extension of
Credibility Interval from 0.0007 to 0.1352. The the standard random effects model46 described
overall estimate of the odds ratio can be seen above, but with an extra level of variation to
to be very similar to that obtained using the
allow for variability in effect sizes, and different
Classical random effects model previously, but
weightings, between different sources.
the 95% Credibility Interval is slightly wider,
reflecting the fact that uncertainty regarding the It has yet to be established when such analyses
between-study variance has now been accounted are appropriate, as there is concern that includ-
for. ing studies with poorer designs will weaken the
analysis, though this issue is partially addressed
by conducting sensitivity analyses under various
GENERALISED SYNTHESIS OF EVIDENCE assumptions. However, the issue of quality
assessment across studies of different designs
needs addressing.
Generalised synthesis of evidence is the term
given to combination of evidence on efficacy or
effectiveness of a particular technology derived
from a number of sources, not necessarily shar- OTHER ISSUES IN META-ANALYSIS
ing common biases and designs. Although RCTs
may be considered the ‘gold standard’ approach Meta-Analysis of Observational and/or
to assess the efficacy of an intervention, in some
circumstances in health technology assessment, Epidemiological Studies
randomised evidence will be less than adequate
due to economic, organisational or ethical con- Many of the methods described above can be
siderations.92 In these situations, considering all used to combine evidence from observational
the available evidence, including that from non- studies, but there may be problems additional to
randomised studies (such as case-control, cohort those encountered with RCTs.43,96 A set of
and database or audit data) may be beneficial. In guidelines for the meta-analysis of environ-
addition, while RCTs may provide a robust mental health data has been compiled,97 much of
estimate of the efficacy of an intervention, which is applicable to observational studies in
assessment of other types of data, derived from other areas.
404 The Advanced Handbook of Methods in Evidence Based Healthcare
DISCUSSION AND CONCLUSIONS 10. Fleiss JL. Cooper H, Hedges LV (eds). The
handbook of research synthesis. Chapter 17,
Measures of effect size for categorical data. New
Systematic review methods including quantita- York: Russell Sage Foundation, 1994, pp. 245–
tive combination of study results are already 60.
common in health technology assessments 11. Cook RJ, Sackett DL. The number needed to
(HTA) contexts, and are likely to become more treat: a clinically useful measure of treatment
common – perhaps almost universal – as moves effect. Br. Med. J. 1995; 310: 452–4.
towards evidence-based medicine and healthcare 12. Fleiss JL. The statistical basis of meta-analysis.
strengthen. This chapter describes the more Stat. Methods. Med. Res. 1993; 2: 121–45.
basic and common methods of meta-analysis 13. Sinclair JC, Bracken MB. Clinically useful
that may be employed, considers problems measures of effect in binary analyses of random-
likely to be encountered in applications of meta- ized trials. J. Clin. Epidemiol. 1994; 47: 881–9.
analysis in HTA, and briefly introduces other 14. Dowie J. The number needed to treat and the
more complex or powerful approaches already ‘adjusted NNT’ in health care decision making.
established or under development. Some of Presented at the conference ‘From Research
these may prove especially valuable in HTA Evidence to Recommendations Exploring the
contexts. Methods’; Harrogate 17–18 February 1997
(unpublished).
15. Deeks JJ, Altman DG, Dooley G, Sackett DLS.
Choosing an appropriate dichotomous effect
REFERENCES measure for meta-analysis: empirical evidence
of the appropriateness of the odds ratio and
1. Sacks HS, Berrier J, Reitman D, Ancona-Berk relative risk. Controlled Clinical Trials 1997;
VA, Chalmers TC. Meta-analysis of randomized 18: 84s–85s.
controlled trials. N. Engl. J. Med. 1987; 316: 16. Birge RT. The calculation of errors by the
450–5. method of least squares. Physiol. Rev. 1932; 16:
2. Sutton AJ, Abrams KR, Jones DR, Sheldon TA, 1–32.
Song F. Systematic reviews of trials and other 17. Cochran WG. Problems arising in the analysis of
studies. Health Technology Assessment 1998; a series of similar experiments. J. Roy. Statist.
2(19). Soc. 1937; 4 (Suppl.): 102–18.
3. Deeks J, Glanville J, Sheldon T. Undertaking 18. Shadish WR, Haddock CK. Cooper H, Hedges
systematic reviews of research on effectiveness: LV (eds). The handbook of research synthesis.
CRD guidelines for those carrying out or com- Chapter 18, Combining estimates of effect size.
missioning reviews. York: CRD 1996, Report New York: Russell Sage Foundation, 1994,
#4. pp. 261–84.
4. Oxman AD. The Cochrane Collaboration hand- 19. Rosenthal R. Cooper H, Hedges LV (eds). The
book: preparing and maintaining systematic handbook of research synthesis. Chapter 16,
reviews. Oxford: Cochrane Collaboration, 1996. Parametric measures of effect size. New York:
5. Cook DJ, Sackett DL, Spitzer WO. Methodo- Russell Sage Foundation, 1994, pp. 231–44.
logic guidelines for systematic reviews of ran- 20. Mantel N, Haenszel W. Statistical aspects of the
domized control trials in health care from the analysis of data from retrospective studies of
Potsdam Consultation on Meta-Analysis. J. Clin. disease. J. Natl Cancer Inst. 1959; 22: 719–48.
Epidemiol. 1995; 48: 167–71. 21. Robins J, Breslow N, Greenland S. Estimators of
6. Smith GD, Song F, Sheldon TA, Song FJ. the Mantel-Haenszel variance consistent in both
Cholesterol lowering and mortality: The import- sparse data and large-strata limiting models.
ance of considering initial level of risk. Br. Med. Biometrics 1986; 42: 311–23.
J. 1993; 306: 1367–73. 22. Robins J, Greenland S, Breslow NE. A general
7. Sutton AJ, Lambert PC, Hellmich M, Abrams estimator for the variance of the Mantel-
KR, Jones DR. Meta-analysis in practice: a Haenszel odds ratio. Am. J. Epidemiol. 1986;
critical review of available software. In: Berry 124: 719–23.
DA, Strangl DK (eds). Meta-analysis in medicine 23. Sankey SS, Weissfeld LA, Fine MJ, Kapoor W.
and health policy. New York: Marcel Dekker, An assessment of the use of the continuity
2000. correction for sparse data in meta-analysis.
8. Hedges LV, Olkin I. Statistical Methods for Communications in Statistics-Simulation and
meta-analysis. London: Academic Press, 1985. Computation 1996; 25: 1031–56.
9. Meinert CL. Clinical trials dictionary: Terminol- 24. Peto R, Pike MC, Armitage P, Breslow NE, Cox
ogy and usage recommendations. Baltimore, DR, Howard SV, Mantel N, McPherson K, Peto
Maryland: The Johns Hopkins Center for Clin- J, Smith PG. Design and analysis of randomized
ical Trials, 1996. clinical trials requiring prolonged observation of
406 The Advanced Handbook of Methods in Evidence Based Healthcare
each patient. II: Analysis and examples. Br. J. 42. Sharp SJ, Thompson SG, Altman DG. The
Cancer 1977; 35: 1–39. relation between treatment benefit and under-
25. Yusuf S, Peto R, Lewis J, et al. Beta blockade lying risk in meta-analysis. Br. Med. J. 1996;
during and after myocardial infarction: an over- 313: 735–8.
view of the randomised trials. Prog. Cardiovasc. 43. Greenland S. Quantitative methods in the review
Dis. 1985; 27: 335–71. of epidemiological literature. Epidemiol. Rev.
26. Greenland S, Salvan A. Bias in the one-step 1987; 9: 1–30.
method for pooling study results. Stat. Med. 44. National Research Council. Combining Infor-
1990; 9: 247–52. mation: Statistical Issues and Opportunities for
27. Fleiss JL. Statistical methods for rates and pro- Research. Washington, D.C. National Academy
portions. 2nd edn. New York: Wiley; 1981. Press, 1992.
28. Emerson JD. Combining estimates of the odds 45. Thompson SG. Controversies in meta-analysis:
ratio: the state of the art. Stat. Methods Med. the case of the trials of serum cholesterol reduc-
Res. 1994; 3: 157–78. tion. Stat. Methods Med. Res. 1993; 2: 173–92.
29. Hauck WW. A comparative study of conditional 46. DerSimonian R, Laird N. Meta-analysis in clin-
maximum likelihood estimation of a common ical trials. Controlled Clinical Trials 1986; 7:
odds ratio. Biometrics 1984; 40: 1117–23. 177–88.
30. Hasselblad VIC, McCrory DC. Meta-analytic 47. Hardy RJ, Thompson SG. A likelihood approach
tools for medical decision making: A practical to meta-analysis with random effects. Stat. Med.
guide. Medical Decision Making 1995; 15: 1996; 15: 619–29.
81–96. 48. Biggerstaff BJ, Tweedie RL. Incorporating vari-
31. Dear KBG. Iterative generalized least squares ability in estimates of heterogeneity in the ran-
for meta-analysis of survival data at multiple dom effects model in meta-analysis. Stat. Med.
times. Biometrics 1994; 50: 989–1002. 1997; 16: 753–68.
32. Hunink MGM, Wong JB. Meta-analysis of 49. Van Houwelingen HC, Zwinderman KH, Stijnen
failure-time data with adjustment for covariates. T. A bivariate approach to meta-analysis. Stat.
Medical Decision Making 1994; 14: 59–70.
Med. 1993; 12: 2273–84.
33. Messori A, Rampazzo R. Metaanalysis of
50. Peto R. Why do we need systematic overviews
clinical-trials based on censored end-points –
of randomised trials? Stat. Med. 1987; 6:
simplified theory and implementation of the
233–40.
statistical algorithms on a microcomputer. Com-
51. Raudenbush SW. Cooper H, Hedges LV (eds).
puter Methods and Programs in Biomedicine
The handbook of research synthesis. Chapter 20,
1993; 40: 261–7.
Random effects models. New York: Russell
34. Whitehead A, Jones NMB. A meta-analysis of
Sage Foundation, 1994, pp. 301–22.
clinical trials involving different classifications
of response into ordered categories. Stat. Med. 52. Greenland S. Invited commentary: a critical look
1994; 13: 2503–15. at some popular meta-analytic methods. Am. J.
35. Whitehead A, Whitehead J. A general para- Epidemiol. 1994; 140: 290–6.
metric approach to the meta-analysis of ran- 53. Gelber RD, Goldhirsch A. The evaluation of
domised clinical trials. Stat. Med. 1991; 10: subsets in meta-analysis. Stat. Med. 1987; 6:
1665–77. 371–88.
36. Cochran WG. The combination of estimates 54. Dickersin K, Berlin JA. Meta-analysis: state-of-
from different experiments. Biometrics 1954; the-science. Epidemiol. Rev. 1992; 14: 154–76.
10: 101–29. 55. Tweedie RL, Mengersen KL. Meta-analytic
37. Thompson SG, Pocock SJ. Can meta-analyses approaches to dose-response relationships, with
be trusted? Lancet 1991; 338: 1127–30. application in studies of lung cancer and expos-
38. Fleiss JL. Analysis of data from multiclinic ure to environmental tobacco smoke. Stat. Med.
trials. Controlled Clinical Trials 1986; 7: 1995; 14: 545–69.
267–75. 56. Berlin JA, Longnecker MP, Greenland S. Meta-
39. Matt GE, Cook TD. Cooper H, Hedges LV analysis of epidemiologic dose-response data.
(eds). The handbook of research synthesis. Epidemiology 1993; 4: 218–28.
Chapter 31, Threats to the validity of research 57. Thompson SG, Smith TC, Sharp SJ. Investiga-
synthesis. New York: Russell Sage Foundation, tion underlying risk as a source of heterogeneity
1994, pp. 503–20. in meta-analysis. Stat. Med. 1997; 16: 2741–58.
40. Galbraith RF. A note on graphical presentation 58. McIntosh MW. The population risk as an
of estimated odds ratios from several clinical explanatory variable in research synthesis of
trials. Stat. Med. 1988; 7: 889–94. clinical trials. Stat. Med. 1996; 15: 1713–28.
41. L’Abbé KA, Detsky AS, O’Rourke K. Meta- 59. Cook RJ, Walter SD. A logistic model for trend
analysis in clinical research. Ann. Intern. Med. in 2 3 2 3 kappa tables with applications to
1987; 107: 224–33. meta-analyses. Biometrics 1997; 53: 352–7.
Meta-Analysis in Health Technology Assessment 407
60. Walter SD. Variation in baseline risk as an 78. Raiffa H, Schlaifer R. Applied statistical decision
explanation of heterogeneity in meta-analysis. theory. Boston: Harvard Business School;
Stat. Med. 1997; 16: 2883–900. 1961.
61. Senn S. Importance of trends in the interpreta- 79. Louis TA. Estimating a population of parameter
tion of an overall odds ratio in the meta-analysis values using Bayes and empirical Bayes meth-
of clinical trials. Stat. Med. 1994; 13: 293–6. ods. J. Am. Statist. Assoc. 1984; 79: 393–8.
62. Lau J, Ioannidis JP, Schmid CH. Summing up 80. Raudenbush SW, Bryk AS. Empirical Bayes
evidence: one answer is not always enough. meta-analysis. J. Educ. Statist. 1985; 10:
Lancet 1998; 351: 123–7. 75–98.
63. Thompson SG, Sharp SJ. Explaining hetero- 81. Zhou XH. Empirical Bayes combination of esti-
geneity in meta-analysis: a comparison of meth- mated areas under ROC curves using estimating
ods. Stat. Med. 2000; 18: 2693–708. equations. Medical Decision Making 1996; 16:
64. Berkey CS, Hoaglin DC, Mosteller F, Colditz 24–8.
GA. A random-effects regression model for 82. van Houwelingen HC, Stijnen T. Monotone
meta-analysis. Stat. Med. 1995; 14: 395–411. empirical Bayes estimators based on more
65. Stram DO. Meta-analysis of published data informative samples. J. Am. Statist. Assoc. 1993;
using a linear mixed-effects model. Biometrics 88: 1438–43.
1996; 52: 536–44. 83. Waclawiw MA, Liang KY. Empirical Bayes
66. Hedges LV. Cooper H, Hedges LV (eds). The estimation and inference for the random effects
handbook of research synthesis. Chapter 19, model with binary response. Stat. Med. 1994;
Fixed effects models. New York: Russell Sage 13: 541–51.
Foundation; 1994, pp. 285–300. 84. Morris CN. Hierarchical models for combining
67. Breslow NE. Biostatisticians and Bayes (with information and for meta-analysis. Bayesian
Statistics 1992; 4: 321–44.
discussion). Statist. Sci. 1990; 5: 269–98.
85. Stijnen T, Van Houwelingen JC. Empirical
68. Gilks WR, Clayton DG, Spiegelhalter D, Best
Bayes methods in clinical trials meta-analysis.
NG, McNeil AJ, Sharples LD, Kirby AJ. Model-
Biometrical J. 1990; 32: 335–46.
ling complexity: applications of Gibbs sampling
86. Carlin JB. Meta-analysis for 2 3 2 tables:
in medicine. J. Roy. Statist. Soc. B 1993; 55:
a Bayesian approach. Stat. Med. 1992; 11:
39–52.
141–58.
69. Berry DA, Stangl DK. Bayesian Biostatistics.
87. Smith TC, Abrams KR, Jones DR. Using hier-
New York: Marcel Dekker, 1996.
archical models in generalised synthesis of evi-
70. DuMouchel W, Berry DA (eds). Statistical
dence: an example based on studies of breast
Methodology in the Pharmaceutical Sciences.
cancer screening. Department of Epidemiology
New York: Marcel Dekker, 1989, pp. 509–29. and Public Health Technical Report 95-02. Uni-
71. DuMouchel WH, Harris JE. Bayes methods for versity of Leicester, 1995.
combining the results of cancer studies in 88. Smith TC, Abrams KR, Jones DR. Assessment
humans and other species (with comment). J. of prior distributions and model parameterisa-
Am. Statist. Assoc. 1983; 78: 293–308. tion in hierarchical models for the generalised
72. DuMouchel, W. Hierarchical Bayes linear synthesis of evidence. Department of Epidemiol-
models for meta-analysis. Research Triangle ogy and Public Health Technical Report 96-01.
Park, NC: National Institute of Statistical University of Leicester, 1996.
Sciences, 1994; 27. 89. Gilks WR, Richardson S, Spiegelhalter DJ.
73. Abrams KR, Sanso B. Approximate Bayesian Markov Chain Monte Carlo in practice. London:
inference in random effects meta-analysis. Stat. Chapman & Hall, 1996.
Med. 1998; 17: 201–18. 90. Best NG, Spiegelhalter DJ, Thomas A, Brayne
74. Smith TC, Spiegelhalter DJ, Thomas A. CEG. Bayesian analysis of realistically com-
Bayesian approaches to random-effects meta- plex models. J. Roy. Statist. Soc. A 1996; 159:
analysis: a comparative study. Stat. Med. 1995; 323–42.
14: 2685–99. 91. Gilks WR, Thomas A, Spiegelhalter DJ. A
75. Skene AM, Wakefield JC. Hierarchical models language and program for complex Bayesian
for multicentre binary response studies. Stat. models. The Statistician 1994; 43: 169–78.
Med. 1990; 9: 919–29. 92. Black N. Why we need observational studies to
76. Hedges LV. Everitt BS, Dunn G (eds). Statis- evaluate the effectiveness of health care. Br.
tical analysis of medical data. New develop- Med. J. 1996; 312: 1215–18.
ments. London: Arnold, 1998. 93. General Accounting Office. Cross design syn-
77. Box GEP; Tiao GC. Bayesian inference in statis- thesis: a new strategy for medical effectiveness
tical analysis. Massachusetts: Addison-Wesley; research. Washington, DC. General Accounting
1973. Office, 1992.
408 The Advanced Handbook of Methods in Evidence Based Healthcare
94. Eddy DM, Hasselblad V, Shachter R. Meta- Completeness of reporting of trials published in
analysis by the Confidence Profile Method. San languages other than English: implications for
Diego: Academic Press, 1992. conduct and reporting of systematic reviews.
95. Abrams KR, Hellmich M, Jones DR. Bayesian Lancet 1996; 347: 363–6.
approach to health care evidence. Department of 106. Begg CB. Cooper H, Hedges LV (eds). The
Epidemiology and Public Health Technical handbook of research synthesis. Chapter 25,
Report 97–01. University of Leicester, 1997. Publication bias. New York: Russell Sage Foun-
96. Spitzer WO. Meta-meta-analysis: unanswered dation, 1994, pp. 399–409.
questions about aggregating data. J. Clin. Epi- 107. Begg CB, Mazumdar M. Operating character-
demiol. 1991; 44: 103–7. istics of a rank correlation test for publication
97. Blair A, Burg J, Foran J, et al. Guidelines for bias. Biometrics 1994; 50: 1088–101.
application of meta-analysis in environmental 108. Egger M, Smith GD, Schneider M, Minder C.
epidemiology. ISLI Risk Science Institute. Bias in meta-analysis detected by a simple,
Regul. Toxicol. Pharmacol. 1995: 189–97. graphical test. Br. Med. J. 1997; 315: 629–34.
98. Boissel JP, Blanchard J, Panak E, Peyrieux JC, 109. Iyengar S, Greenhouse JB. Selection models and
Sacks H. Considerations for the meta-analysis of the file drawer problem. Statist. Sci. 1988; 3:
randomized clinical trials: summary of a panel 109–35.
discussion. Controlled Clinical Trials 1989; 10: 110. Dear KBG, Begg CB. An approach for assessing
254–81. publication bias prior to performing a meta-
99. Stewart LA, Clarke MJ. Practical methodology
analysis. Statist. Sci. 1992; 7: 237–45.
of meta-analyses (overviews) using updated
111. Vevea JL, Hedges LV. A general linear model for
individual patient data. Cochrane Working
estimating effect size in the presence of publica-
Group. Stat. Med. 1995; 14: 2057–79.
tion bias. Psychometrika 1995; 60: 419–35.
100. Whiting GW, Lau J, Kupelnick B, Chalmers TC.
112. Little RJA, Rubin DB. Statistical analysis with
Trends in inflammatory bowel disease therapy: a
missing data. New York: Wiley, 1987.
meta-analytic approach. Can. J. Gastroenterol.
113. Irwig L, Tosteson AN, Gatsonis C, Lau J,
1995; 9: 405–11.
101. Antman EM, Lau J, Kupelnick B, Mosteller F, Colditz G, Chalmers TC, Mosteller F. Guide-
Chalmers TC. A comparison of results of meta- lines for meta-analyses evaluating diagnostic
analyses of randomized control trials and recom- tests. Ann. Intern. Med. 1994; 120: 667–76.
mendations of clinical experts: Treatments for 114. Berkey CS, Anderson JJ, Hoaglin DC. Multiple-
myocardial infarction. JAMA 1992; 268: 240–8. outcome meta-analysis of clinical trials. Stat.
102. Lau J, Schmid CH, Chalmers TC. Cumulative Med. 1996; 15: 537–57.
meta-analysis of clinical trials: builds evidence 115. Daniels MJ, Hughes MD. Meta-analysis for the
for exemplary medical care. J. Clin. Epidemiol. evaluation of potential surrogate markers. Stat.
1995; 48: 45–57. Med. 1997; 16: 1965–82.
103. Jefferson T, DeMicheli V, Mugford M. Ele- 116. Tori V, Simon R, Russek-Cohen E, Midthune D,
mentary economic evaluation in health care. 8. Friedman M. Statistical model to determine the
Current issues. London: BMJ Publishing Group, relationship of response and survival in patients
1996. with advanced ovarian cancer treated with
104. Easterbrook PJ, Berlin JA, Gopalan R, Matthews chemotherapy. J. Natl Cancer Inst. 1992; 84:
DR. Publication bias in clinical research. Lancet 407–14.
1991; 337: 867–72. 117. Margitic SE, Morgan TM, Sager MA, Furberg
105. Moher D, Fortin P, Jadad AR, Juni P, Klassen T, CD. Lessons learned from a prospective meta-
Le Lorier J, Liberati A, Linde K, Penna A. analysis. J. Am. Geriatr. Soc. 1995; 43: 435–9.
23
Assessing the Quality of Reports of
Randomised Trials Included in
Meta-Analyses: Attitudes, Practice,
Evidence and Guides
and component approach appears to be com- increasing evidence6 regarding the effects that
plementary. Quality assessments should be incor- low-quality trials have on the results of meta-
porated into meta-analyses through sensitivity analyses, few data exists to suggest whether any
analyses. Alternatively, the quality weighting one method of quality assessment provides a
method is showing promise as a valid option for more biased estimate than any other. More
incorporation of quality into overall treatment broadly, little attention has been given to syn-
effect. thesising the available evidence to help develop
practical guides to facilitate issues regarding
quality assessment in the conduct and reporting
INTRODUCTION of meta-analyses.
The importance of assessing trial quality and
its use in analysis is something that many meta-
Meta-analyses pool individual studies – either analysts have not consistently taken into account
observational or randomised controlled trials until recently. There is an on-going movement
(RCTs) – to provide an overall estimate of the towards developing methodologies to help
effect of the treatment under consideration, and reduce or avoid bias in the conduct and report-
are a key component of evidence based health- ing of meta-analysis. Three approaches to
care. Today, the RCT is considered the most assessing the quality of reports of RCT have
reliable method of assessing the efficacy of been developed: (i) component assessment (items
healthcare interventions.1 However, poorly con- such as randomisation and blinding); (ii) check-
ducted RCTs may yield misleading results, and lists; and (iii) scales. However, there are still no
hence it is important for those involved in set rules or regulations regarding the quality
healthcare to be able to assess the reliability of assessment of trials. Hence, the objectives of
the research evidence that is available. this chapter are to examine this issue and to
Quality is a construct (a concept) that has provide empirically based recommendations on
been defined in a variety of ways. In this chap- how to conduct meta-analyses with respect to
ter, our focus is on internal validity and quality quality assessment.
is defined2 as ‘the confidence that the trials
design, conduct, analysis, and presentation has
minimised or avoided biases in its intervention
comparisons’. As a meta-analysis is conducted METHODS
retrospectively, it is undoubtedly susceptible to
certain sources of bias. The need to assess
quality stems mainly from a desire to estimate To complete the research reported here, we first
the effects of such bias on the results of a created a database of 491 meta-analyses of
randomised trial. Meta-analyses of RCTs have RCTs using the refined strategy reported in
always included studies of variable methodo- Table 1. Meta-analyses were included in the
logical quality, and differences in quality study if they included only randomised trials,
between randomised trials may indicate that and if they were published in the Cochrane
some are more biased than others. It seems only Library or in paper-based journals indexed by
sensible that meta-analysts and others will want Medline. For the latter, they had to be coded by
to account for such differences to ensure that indexers as meta-analyses, or described by the
their findings are valid. authors as systematic reviews, meta-analyses,
The assessment of the quality of individual integrative research reviews, overviews, quanti-
randomised trials included in meta-analyses has tative syntheses, pooling or combining studies.
been viewed with some controversy,3–5 yet only In a review of 36 Cochrane reviews, com-
limited data exist to inform these views. At a pared to 39 paper based ones, published in 1995,
fundamental level, we know little about the we found that Cochrane reviews included a
views of a broad spectrum of health service description of the inclusion and exclusion cri-
professionals involved in the conduct of meta- teria, assessed trial quality, were more fre-
analyses, regarding quality assessment. Albeit quently updated, and did not have language
several aspects of trial quality have been shown restrictions.7 Such evidence suggests that the
empirically to be important – including the quality of reports of Cochrane reviews are
adequacy of reporting the randomisation strat- superior (in some aspects) to paper ones, and
egy, allocation concealment, and double-blinding since we are interested in producing systematic
procedures as well as the handling of drop-outs reviews with the least amount of bias, our
and withdrawals – the extent to which meta- results section includes a comparison of these
analysts assess trial quality and how they incor- two categories of meta-analyses.
porate a measure of its effect into the quantita- In order to develop guidelines regarding qual-
tive data synthesis is limited. Although there is ity assessment in the conduct and reporting of
Assessing the Quality of Reports of RCTs 411
001 meta-analysis.pt,sh.
002 (meta-anal: or metaanal:).tw.
003 (quantitativ: review: or quantitativ: overview:).tw.
004 (systematic: review: or systematic: overview:).tw.
005 (methodologic: review: or methodologic: overview:).tw.
006 (integrative research review: or research integration:).tw.
007 review.pt,sh. or review:.tw. or overview:.tw.
008 quantitativ: synthes:.tw.
009 1 or 2 or 3 or 4 or 5 or 6 or 8
010 (medline or medlars).tw,sh. or embase.tw.
011 (scisearch or psychinfo or psycinfo).tw.
012 (psychlit or psyclit).tw.
013 (hand search: or manual search:).tw.
014 (electronic database: or bibliographic database:).tw.
015 (pooling or pooled analys: or mantel haenszel).tw.
016 (peto or der simonian or dersimonian or fixed effect:).tw.
017 10 or 11 or 12 or 13 or 14 or 15 or 16
018 7 and 17
019 9 or 18
Table 2 Analysis procedures for each of the three studies in the methods section
Study Analysis
2 Current practice of ● Descriptive statistics for eight questions by source (MAPJ 1 CDSR) and for all
quality assessment 240 meta-analysis
● Differences between MAPJ 1 CDSR compared using chi-squared and Fisher’s
exact test
3 Effects of quality ● Paired t-test for mean differences in quality scores (masked versus unmasked)
assessment on ● Chi-squared 1 logistic regression for differences between masked 1 unmasked in
the results of proportion with adequately reported components
meta-analysis ● Reliability checked via replication of point estimate and 95% CI
● Logistic regression models to examine impact of quality assessment
● Quality scores incorporated into analysis as threshold a quality weight or
individual component (e.g. double-blinding)
● Sensitivity analysis for relationship between a component assessment of quality
compared to a scale
● Results reported as ROR and OR14
● Mean residual deviance of fitted models reflects degree of heterogeneity between
trials (adjusted for independent factors)
● F-test for effects of heterogeneity (P , 0.05)
meta-analysis, three studies were conducted. meta-analysis, and the third dealt with the
One was a survey of reviewers, methodologists effects of quality assessment on the results of
and editors designed to appraise attitudes meta-analysis. Each study is described here,
regarding quality assessment. The second looked and the details of the analysis can be found in
at the current practice of quality assessment of Table 2.
412 The Advanced Handbook of Methods in Evidence Based Healthcare
reporting. The individual components assess the impossible and may not be a reasonable quality
adequacy of reporting of allocation concealment assessment item; however, evaluation of re-
and are described in detail elsewhere.6 bleeding events using explicit, a priori criteria
In addition to the quality assessment of each by an adjudication committee blinded to treat-
randomised trial the following data were also ment may minimise the chance of a biased
extracted: the number of events and patients in outcome assessment, and could be a more dis-
the control group, and the number of events and criminating quality assessment item.
patients in the experimental group. The data Respondents also reported that quality assess-
were extracted independently by two people ment should ideally be based on empirical evi-
(A.L.J. and D.M.) and consensus was achieved dence of bias, i.e. that the essential features of
for any discrepancies before data entry. study design that minimise bias should be rep-
resented. If quality assessment is a vague part
of the analysis process this is likely to lead to a
RESULTS: WHAT THE OPINION LEADERS FEEL variety of problems.2 For example, the conduct
and reporting of trial quality assessment in
SHOULD BE THE ‘RULES’ meta-analyses may be conflated in instruments
such as the 100-point scale9 for ‘quality rating of
Assessing Attitudes Regarding Quality randomised trials’. This tool combines items
Assessment about reporting style (e.g. inclusion of trial com-
mencement and cessation dates), sample size,
The response rates were 78% for meta-analysts, unrelated to study quality or systematic error
74% for methodologists, and 60% for editors. (e.g. whether or not investigators included a
The overwhelming majority of meta-analysts, power analysis), and features designed to min-
methodologists and editors reported that assess- imise systematic error (e.g. treatment allocation,
ment of the quality of randomised trials included blinding, and losses to follow-up).
in a meta-analysis was very or somewhat im- The majority of respondents believed that the
portant (97%, 94% and 100%, respectively). We methods used to develop a quality check-list or
asked all respondents to consider how, in their scale were somewhat or very important (92%,
role (or potential role) as editors of biomedical 94% and 95%, respectively). Several methods2
journals, they would deal with an otherwise by which the quality assessments of randomised
rigorous meta-analysis in which the quality of trials could be incorporated into meta-analyses
randomised trials had not been assessed. Over- were endorsed by respondents.
all, 66% said they would be unenthusiastic about The majority of reviewers, methodologists
publishing it unless the quality of the trials was and editors believed that guidelines for assessing
assessed, 5% were indifferent, and 28% said the quality of randomised trials included in a
they would be willing to publish an otherwise meta-analysis would be likely or very likely to
rigorous meta-analysis if the trial quality was increase the rigour and reporting of published
not assessed. Reviewers, methodologists and meta-analyses (Table 3). Most respondents also
editors had different views (P 5 0.04), editors believed that such guidelines would likely or
being less enthusiastic about publishing a meta- very likely make it easier for clinicians to inter-
analysis in which quality had not been assessed. pret meta-analyses (Table 4).
In considering ways in which the quality of
randomised trials included in a meta-analysis
should be assessed, use of a series of items as in Current Practice of Quality Assessment
a check-list was recommended by 45% of by Meta-Analysts
reviewers, 57% of methodologists and 62% of
editors. Assessment of a series of items that Trial quality was assessed in 114 (48%) of the
would generate an overall summary score, i.e. a 240 meta-analyses. The quality of the primary
scale, was recommended by 28%, 30% and 38% trials was assessed more frequently in the CDSR
of the respondents, respectively. Qualitative meta-analyses than in the MAPJ (100% versus
analysis yielded recommendations that a modest 38%, P , 0.001]. Fifty-seven (50%) of the 114
number of criteria be used by meta-analysts to meta-analyses in which trial quality was
assess and report trial quality. assessed provided data on the reproducibility of
Reliance solely on universal criteria was con- the assessments. CDSR meta-analyses evaluated
sidered inappropriate, and specific items tailored the reproducibility of the assessments more fre-
to the question under investigation were thought quently than MAPJ (56% versus 36% respect-
to be important. For example, when evaluating ively; P 5 0.04). Individual components and
trials comparing drug treatment versus sclero- scales were the methods most frequently used
therapy for bleeding oesophageal varices, tradi- (46% each) to assess trial quality. Most of the
tional blinding of patients and care-givers is CDSR used individual components to assess
414 The Advanced Handbook of Methods in Evidence Based Healthcare
The potential impact of guidelines for assessing the quality of randomised trials included in meta-analyses on the
rigour and reporting of published meta-analyses is presented here according to views of meta-analysts,
methodologists and editors. Data are presented as number of respondents and proportion of the total respondents
in each category.
Values in parentheses are percentages.
The potential impact of guidelines for assessing the quality of randomised trials included in meta-analyses on the
way in which meta-analyses may be interpreted by clinicians is presented here according to views of meta-
analysts, methodologists and editors. Data are presented as number of respondents and proportion of the total
respondents in each category.
Values in parentheses are percentages.
trial quality, whilst most MAPJ meta-analyses When only the meta-analyses published in
used scales (Table 5). A total of 21 quality 1995 were analysed, most of the patterns out-
assessment instruments were identified. None of lined above persisted. Twelve of the 39 MAPJ
these instruments appeared to have undergone (32%) included assessments of trial quality, and
validation following established methodological most of them used scales but did not incorporate
procedures,10 and 43% were described for the the assessments into the analyses.
first time in 1994 and 1995. Eleven of these
instruments were not included in a recently Effects of Quality Assessment on the
published systematic review.10
Of the 114 meta-analyses that included Results of Meta-Analyses
assessments of trial quality, only 29 (25%) took
The overall quality of reporting of randomised
such assessments into account during data trials using a scale assessment was 2.74 (out of
analyses (Table 5). MAPJ incorporated the five; SD 5 1.1), corresponding to 54.8% of the
quality assessment in the analyses more fre- maximum possible value. There were statis-
quently than CDSR meta-analyses (34% vs. tically significant differences in the evaluation
6% respectively; P , 0.001). The two CDSR of the quality of reporting of randomised trials
meta-analyses that incorporated the quality under masked and unmasked conditions. Masked
assessments into data analysis used the quality assessment resulted in statistically higher quality
assessments as thresholds. Of the MAPJ meta- scores compared to unmasked assessments; this
analyses, one-third incorporated the quality difference corresponds to 3.8%. All further ana-
assessments as thresholds for inclusion or exclu- lyses presented below are based on masked
sion from the analyses, and one-third incor- assessments only.
porated the quality scores in the formulae as Using a component approach to quality
weights. assessment, few randomised trials reported on
Assessing the Quality of Reports of RCTs 415
MAPJ MAPJ
CDSR (1995) (1977–1995) Total
either the methods used to generate the random- We conducted a threshold analysis to deter-
isation schedule (15.0%) or the methods used to mine whether the exaggerated intervention
conceal the randomisation sequence until the effects reported above, in relation to the quality
point of randomisation occurred (14.3%). When scores, could be explained by those randomised
assessed under masked conditions, compared to trials in which allocation concealment was
unmasked ones, allocation concealment was inadequately reported and inadequately carried
assessed more frequently (14.3% versus 10.7%) out, as has been previously suggested.6 Our
as adequate. analyses (see Table 6) did not result in any
Evaluating the influence that quality assess- meaningful differences in terms of magnitude
ments of the primary trials have on the results of and direction of bias or statistical significance
the meta-analyses is presented in Table 6. A than those already reported here.
quality score of 2 was utilised, based on empirical Incorporating estimates of quality based on
evidence, as the threshold value discriminating individual components also detected exagger-
between low- and high-quality trials.8,11 Trials ated estimates of treatment effect (see Table 6).
with a low quality score (#2), compared to a Clinical trials reporting allocation concealment
high quality one (.2), resulted in a significant inadequately, compared to those trials reporting
increase in the benefits of interventions by 34% it adequately, produced statistically exaggerated
(ratio of odds ratios; ROR 5 0.66; 95% CI: 0.52, estimates of treatment effects of 37% (ROR 5
0.83). The effects of quality assessment on the 0.63; 95% CI: 0.45, 0.88).
results of an individual meta-analysis can be
seen in Box 2. DEVELOPMENT OF GUIDES FOR ASSESSING
The use of quality scores as a quality weight
THE QUALITY OF RANDOMISED TRIALS
also identified exaggerated effect estimates.
Including trials with low quality scores (#2), INCLUDED IN META-ANALYSIS
compared to a quality weight, resulted in sig-
nificantly exaggerated treatment effect of 55% Once the three studies described above were
(ROR 5 0.45; 95% CI: 0.22, 0.91). However, conducted and results were available, a confer-
using a sensitivity analysis whereby high-quality ence was held of all the investigators involved
trials (score .2) were incorporated into the with the studies, in Ottawa, Ontario, Canada.
analysis compared to a quality weight did not Following this conference, monthly teleconfer-
produce different treatment effects (see Table 6). ences of investigators were conducted to further
Using a quality weight to incorporate estimates develop the evidence into guides. Based on
of quality also produced less statistical hetero- the evidence that resulted from these studies,
geneity (see Table 6). supplemented with systematically assembled
416 The Advanced Handbook of Methods in Evidence Based Healthcare
Scale
Low (#2) versus High (.2)* 0.66 (0.52, 0.83) 1.00l (0.51)
Low (#2) versus Weight† 0.45 (0.22, 0.91) 1.77l (0.02)
High (.2) versus Weight† 0.97 (0.70, 1.33) 1.76l (0.006)
Sensitivity analysis
[excluding trials in which allocation
concealment is reported adequately]
Low [#2] versus High [.2]* 0.73 (0.56, 0.94) 0.82l (0.72)
Low [#2] versus Weight† 0.69 (0.55, 0.86) 1.70l (0.03)
High [.2] versus Weight† 1.16 (0.93, 1.44) 2.08l (0.003)
The analysis used the convention that treatment was more effective to prevent an adverse outcome (i.e. a summary
odds ratio in each trial of less than 1). Hence a ratio of odds ratios (ROR) of less than 1 indicates an exaggeration
of treatment effect. A base model consisted of treatment, trials and the different odds ratios (OR) in each meta-
analysis (i.e. interaction term).
* Allowing for summary ORs to vary according to quality (i.e. quality by treatment interaction) in the base
model.
†
Partial ROR of average treatment effects derived from fitting separate base models.
‡
Allowing for summary ORs to vary simultaneously according to the components (i.e. component by treatment
interactions).
§
The residual deviance derived from fitting the base model reflects the degree of heterogeneity between
trials.
** An approximate F-distribution was assumed for the ratio of residual deviances to compare the heterogeneity
between different ways of incorporating quality. l A larger degree of heterogeneity between trials results in a
ratio greater than 1.
††
Sensitivity analysis including trials with allocation concealment reported inadequately.
evidence from the literature,12,13 guides were respondents would be unenthusiastic about pub-
formulated and underwent a rigorous review lishing a meta-analysis, unless the quality of the
process, and were then revised accordingly by primary studies had been assessed.
this process. Strong support for quality assessments comes
from studies indicating that several dimensions
Should the Quality of Randomised of study quality can be used to detect bias in
Trials be Assessed? treatment estimates (see Table 7).6,14–17 We
recently confirmed Schulz’s finding and demon-
We have demonstrated that the quality of reports strated that low-quality randomised trials, com-
of randomised trials included in meta-analyses pared to high-quality ones, exaggerated the
was assessed in only 38% of meta-analyses effectiveness of the intervention by 34%, on
published in peer-reviewed journals. However, average (see Table 6).14 Meta-analysis based on
our survey of methodologists, journal editors biased randomised trials will also have a similar
and meta-analysts indicated that 95% or more of tendency toward bias (the ‘garbage in, garbage
respondents believe that quality assessment of out’ phenomenon). Therefore, we recommend
randomised trials is very, or somewhat, import- that the quality of all randomised trials included
ant. The survey also revealed that 66% of in a meta-analysis should be assessed.
Assessing the Quality of Reports of RCTs 417
How should the quality of reports of trials (Table 7).14 However, there was no sig-
randomised trials be assessed? nificant difference detected in studies that were
not double-blinded, or in which there was not
adequate description of how the sequence of
Items for which there is empirical evidence random numbers were generated.
If quality assessment is performed, the next Khan and colleagues assessed the probability
question becomes how should it be evaluated? of bias in crossover trials as compared to paral-
Quality assessments should be based on those lel trials in infertility research that utilised preg-
dimensions of randomised trials that are related nancy as the outcome measure. They found that
to bias in treatment estimates. There is increas- crossover trials over-estimated the odds ratio by
ing evidence for which dimensions of trial 74% (Table 7). The underlying reason for this
design and conduct affect the estimate of treat- bias is that crossover trials will over-estimate
ment effectiveness. By using 145 trials exam- the effectiveness of interventions when preg-
ining the treatment of myocardial infarction, nancy is the outcome because once patients
Chalmers and colleagues demonstrated that become pregnant they cannot be crossed over to
trials which were not randomised had a 58.1% the comparison intervention.15
We recommend using primarily evidence
difference in case-fatality rates in treatment
based components when assessing the quality of
in comparison to control groups; this com-
reports of randomised trials. The selection and
pared with a difference of 24.4% in trials that
use of other items cannot be guaranteed to guard
were randomised but not blinded, and an 8.8% against providing erroneous and/or biased infor-
difference in trials randomised and blinded mation to meta-analysts and the results of meta-
(Table 7).18 analyses. The items for which there are varying
Schulz examined 250 controlled trials from degrees of empirical evidence and items that are
33 meta-analyses published by the Pregnancy commonly used, but without such evidence, are
and Childbirth Group of the Cochrane Collab- summarised in Table 8.
oration. Compared to trials that had adequate
allocation concealment, he found that the odds
ratio was exaggerated by 41% for inadequately The Use of Scales for Measuring
concealed trials, and by 30% for unclearly con- Quality
cealed trials (Table 7). Trials that were not
double-blind, compared to those that were, One attractive feature of using a scale for meas-
yielded odds ratios exaggerated by 17%.6 The uring the quality of randomised trials is that in a
impact of not using double-blind randomised scale each item is scored numerically and used
trials was confirmed in studies by Miller and to generate an overall quantitative estimate for
Colditz.16,17 Our study showed that inadequately quality. However, most scales have been devel-
concealed trials generated a 37% increased treat- oped in an arbitrary fashion with minimal atten-
ment effect compared to adequately concealed tion to accepted methodological standards of
418
The Advanced Handbook of Methods in Evidence Based Healthcare
Table 7 Summary of empirical evidence relating to quality assessment of randomised trials included in meta-analyses
Year of No. of Disease(s) of
Reference publication Study design studies interest Methodological item(s) Results
Berlin28 1997 RCT 5 meta-analyses Various Masking of primary Masked summary OR (95% CI) 5 0.63
studies to reviewers (0.57–0.70) and unmasked summary OR (95%
CI) 5 0.64 (0.57–0.72). Mean quality score for
masked reviewers 7.4 and 8.1 for unmasked
reviewers (P 5 0.036).
Chalmers et al.18 1983 Observational 145 Acute myocardial Blinding/randomisation Differences in case-fatality rates 8.8% in
infarction blinded-randomisation studies, 24.4% in
unblinded randomisation studies, and 58.1% in
non-randomised studies.
Cho and Bero26 1996 Observational 152 Drug studies Pharmaceutical 98% of drug company-sponsored trials were
sponsorship favourable to the drug of interest compared to
79% of those with no drug company
sponsorship.
Colditz et al.16 1989 Observational 113 Studies in medical Randomisation/double- Non-randomised trials with sequential
journals blinding assignment had significantly better outcomes
for new therapies (P 5 0.004). Randomised
trials that were not double-blinded trials
favoured the new therapy significantly more
often (P 5 0.02).
Detsky et al.4 1992 Observational 8 trials TPN in Incorporating quality Four methods of incorporating quality into
(1 meta-analysis) chemotherapy into meta-analysis meta-analyses were demonstrated: 1) Inclusion/
exclusion criteria; 2) Quality scores as weights;
3) Plot effect size versus quality score; and 4)
Sequential combination of trial results based on
quality scores.
Jadad et al.8 1996 RCT 36 Pain Masking of primary Mean quality score was 2.7 in unmasked group
studies for quality compared to 2.3 in masked group (P , 0.01).
assessment
Khan et al.15 1996 Observational 34 Infertility Crossover versus Cross-over trials overestimated odds ratio by
parallel study design 74% compared to parallel design (95% CI, 2%
to 197%).
Khan et al.23 1996 Observational 1 meta-analysis Infertility Impact of quality The summary odds ratio from all studies was
(9 trials) assessment on treatment 1.6 (95% CI 0.9–2.6); for low-quality studies
estimate 2.6 (95% CI 1.2–5.2) and for high-quality
studies 0.5 (0.2–1.5).
Miller et al.17 1989 Observational 221 Surgical trials Randomisation Significantly greater benefit in non-randomised
419
420 The Advanced Handbook of Methods in Evidence Based Healthcare
Table 8 Empirical evidence in clinical trial that this was not present in trials assessed as
assessment high quality.23
In our recent assessment of published meta-
Items analyses, nine new scales were identified24 that
had not been previously identified in 1995.10
Empirical evidence ● Allocation concealment
None of these newly identified scales had been
● Double-blinding
● Type of RCT (parallel
developed using established methodological
or crossover trial)
standards.19,20 Moher and colleagues have pre-
viously shown that the results of quality assess-
No empirical evidence, ● Sample size ments depend on how the scales have been
but commonly used developed (see Table 7).2 We recommend using
appropriately developed scales when assessing
the quality of reports of randomised trials. There
validation and reliability testing.19,20 In addition, is strong evidence for one scale8 (see Box 3).
many scales are not truly measuring quality as The selection and use of less rigorously devel-
defined earlier, but rather focusing on extraneous oped scales for randomised trial quality may
factors more related to generalisability.2,10 lead to erroneous and/or biased information.
In fact, through a systematic search of the
literature we could find only one scale, initially
used for evaluating the pain literature,8 devel- Scales Versus Components
oped according to accepted methodological
principles.19,20 This scale has been used subse- Our survey of an international group of method-
quently to compare trials in different languages ologists, editors and meta-analysts indicated that
and specialty areas.21,22 It is an interval scale between 45% and 62% recommended perform-
ranging from 0 to 5 (0 5 lowest quality; 5 5 ing quality assessment through the use of a
highest quality) that assesses method of random- series of items.24 There was less support for the
isation, double-blinding and handling of with- use of a scale ranging from 28% of meta-
drawals and drop-outs. Using this scale we have analysts to 38% of editors. We have also found
recently shown that low-quality studies exag- that 92% of systematic reviews published in the
gerated the odds ratio by 34% compared to CDSR utilised components (e.g. allocation con-
high-quality ones (see Table 7).14 Kahn and cealment) compared to only 26% of meta-
colleagues, using this scale in a meta-analysis analyses published in paper based peer-reviewed
based on trials conducted in the infertility journals.7 In contrast, none of the reviews in the
domain, have recently demonstrated that an esti- CDSR used scales for quality assessments com-
mate based on low-quality trials produced a pared to 67% of meta-analyses published in
statistically significant result with treatment, and paper based peer-reviewed journals.
This scale, which was originally used to evaluate the likelihood of bias in pain research reports, is the only one
found that was developed according to accepted methodological principles.
Assessing the Quality of Reports of RCTs 421
assessed quality, but only 25% (29/114) incorpor- disagreements between reviewers at the various
ated quality into the analyses. phases of the review (study selection, quality
There is no strong evidence to support one assessment and data extraction), there was no
method over the other. It may be that at present significant impact on the summary odds ratios
most meta-analysts will find threshold analyses for the meta-analyses included in their study.
intuitively simpler to perform and that it will be They did not, however, explore the impact of
more transparent to the readers. The quality quality assessments on the overall summary
weighting approach is based on the assumption estimate.
that the scale used for quality weighting is linear There is also indirect evidence from the peer-
and, hence, it is logical to give a study with a reviewed literature29–31 that can be used to help
score of 5 out of 5 a full quality weight of 1. inform any recommendation. McNutt and col-
There appear to be certain conceptual advantages leagues29 randomised 127 manuscripts submit-
to use of a quality weight rather than a threshold ted for publication to be assessed under masked
approach. One of these may be that with the use or open conditions. The authors reported that
of quality weight all trials can be included rather
masked assessments, compared to open ones,
than a selected sample, as would be common
with a threshold approach. In our study the produced statistically higher assessments of
use of quality weighting produced results that quality. Evidence of an effect in a different
were less heterogeneous and maintained greater direction has recently been reported.30,31 Seventy-
precision than results from threshold analysis four pairs of peer reviewers were randomised to
because all studies contributed to the overall receive a masked or open version of a manu-
estimate (see Table 7)14 depending on the con- script, and the quality of peer review was
tribution of their quality scores. Regardless of assessed using a validated instrument. The
the method selected, we recommend that all authors reported no statistical differences
meta-analysts should incorporate an estimate of between both groups in terms of the quality of
quality assessment into the quantitative analysis peer review.30 Similar results have been reported
as a ‘first-line’ sensitivity analysis. elsewhere.31 A prudent next move may be to
conduct a systematic review; this is likely to
provide insight into the apparent inconsistent
SHOULD THE REPORTS OF RANDOMISED results across the trials.
TRIALS BE MASKED WHEN QUALITY IS Given the current state of evidence and the
BEING ASSESSED? effort required to conduct masked quality
assessment, strong recommendations are not
suggested by available research evidence. How-
One aspect of quality assessment that may ever, we recommend that meta-analysts should
increase the quantity of work in performing a at least consider this issue and explicitly justify
meta-analysis is the masking of individual trials. their decision. Further research evidence is
There is direct evidence that masking does needed before making a more definitive
impact on the assessment of quality (see Table recommendation.
7).8,14,28 What has not been consistent is the
direction of such impact. Both Jadad and col-
leagues8 and Berlin and colleagues28 have shown Number and Backgrounds of Assessors
that masked quality assessments scored lower
than those performed under open conditions At least two individuals should assess the qual-
(P , 0.01 and P 5 0.04, respectively; see Table
ity of each study included in the meta-analysis,
7). We on the other hand, found that masked
quality assessment resulted in significantly higher as one reviewer or both may make random or
scores compared to open assessments, but systematic errors. We do not know much about
showed a more normal distribution and greater the ideal background and number of quality
consistency (see Table 7).14 In all studies, while assessors. When such assessments are per-
the differences were statistically significant, it is formed, we recommend using a measure of
not clear whether these are methodologically agreement, such as kappa (or weighted kappa
meaningful differences and whether the magni- as appropriate)32 or intra-class correlation,33 as
tude of the difference would be sufficient to appropriate. After completing the quality assess-
make an important difference when quality is ments, these individuals should meet and reach
being incorporated into the quantitative analy- consensus on areas of disagreement; if disagree-
ses. Berlin and colleagues randomised reviewers ment persists, then a third party adjudication
to utilise masked trials versus unmasked trials may be used. These recommendations are not
throughout the systematic review process (see based on empirical evidence, but rather reflect
Table 7).28 They found that, although there were our opinions.
Assessing the Quality of Reports of RCTs 423
studies of methodological issues in meta- 2. Moher D, Jadad AR, Tugwell P. Assessing the
analyses will help to increase their rigour and quality of randomized controlled trials: current
thus their usefulness in practice.34 A useful issues and future direction. Int. J. Technol. Assess.
checklist for meta-analysts, editors, peer review- Health Care 1996; 12: 195–208.
ers, and readers to assess a meta-analysis for its 3. Greenland S. Quality scores are useless and
handling of quality assessment of randomised potentially misleading. Am. J. Epidemiol. 1994;
trials is provided in Table 9. 140: 300–1.
4. Detsky AS, Naylor CD, O’Rourke K, McGeer
AJ, L’Abbe KA. Incorporating variations in the
quality of individual randomized trials into meta-
COMMENT analyses. J. Clin. Epidemiol. 1992; 45: 225–65.
5. Chalmers TC, Lau J. Meta-analysis stimulus for
Most often, assessing the quality of randomised changes in clinical trials. Statist. Methods Med.
trials means assessing the quality as stated in the Res. 1993; 2: 161–72.
report, and not of the actual events that occurred 6. Schulz KF, Chalmers I, Haynes RJ, Altman DG.
during execution of the trial. With initiatives Empirical evidence of bias: dimensions of meth-
that attempt to encourage systematic and com- odological quality associated with estimates of
prehensive reporting of trials, such as the treatment effects in controlled trials. JAMA 1995;
CONSORT statement,35 it is expected that the 273: 408–12.
reporting of randomised trials will become more 7. Jadad AR, Cook DJ, Jones A, Klassen T, Moher
transparent and comprehensive over time. As M, Tugwell P, Moher D. The Cochrane Col-
evidence accumulates in favour of important laboration: its impact on the methodology and
reports of systematic reviews and meta-analyses.
design features of randomised trials that reduce
JAMA 1998; 280: 278–80.
the probability of bias, the quality of how ran-
8. Jadad AR, Moore RA, Carroll D, Jenkinson C,
domised trials are conducted should improve,
Reynolds DJM, Gavaghan DJ, McQuay HJ.
particularly if journal editors and funding
Assessing the quality of reports of randomized
agencies encourage investigators to conduct and
clinical trials: is blinding necessary? Controlled
report their work with this evidence in mind.
Clinical Trials 1996; 17: 1–12.
Whilst strong evidence exists that certain
9. Chalmers TC, Smith H, Blackburn B, Silverman
characteristics of the design and execution of
B, Schroeder B, Reitmen D, Ambroz A. Method
randomised trials do impact on the probability
for assessing the quality of randomized controlled
of bias, further research is needed to identify trials. Controlled Clinical Trials 1981; 2: 31–49.
other potential aspects influencing the results 10. Moher D, Jadad AR, Nichol G, Penman M,
of randomised trials. Further studies are also Tugwell P, Walsh S., Assessing the quality
needed to clarify the role of masking studies of randomized controlled trials: an annotated bib-
before performing quality assessments, and liography of scales and checklists. Controlled
additionally there is a need for more than one Clinical Trials 1995; 16: 62–73.
reviewer and different backgrounds and levels 11. Pham B, Moher D, Klassen TP. Incorporating
of expertise to perform such assessments. Whilst quality assessments of randomized controlled
several methods have been described in this trials into meta-analyses: are effect sizes adjusting
chapter for the incorporation of study quality for quality scores appropriate? Technical Report.
into the results of a meta-analysis, further Thomas C. Chalmers Centre for Systematic
research in this area is required to determine Reviews, 1999.
whether there is any significant advantage of 12. Kunz RA, Oxman AD. The unpredictability para-
one approach over the other.4 Additionally, if dox: review of empirical comparisons of ran-
weighting of studies based on quality assess- domised and non-randomised clinical trials. Br.
ment proves to be an appropriate method, then Med. J. 1998; 317: 1185–90.
software will need to be developed to assist 13. Kleijnen J, Gotzsche P, Kunz RA, Oxman AD,
reviewers in this task. Chalmers I. So what’s so special about random-
ization? In: Maynard A, Chalmers I (eds). Non-
random reflections on health services research.
REFERENCES London: BMJ publishing group, 1997, pp. 93–
106.
14. Moher D, Pham B, Jones A, Cook DJ, Jadad AR,
1. Cook DJ, Guyatt GH, Laupacis A, Sackett DL. Moher M, Tugwell P, Klassen TP. Does the poor
Rules of evidence and clinical recommendations quality of reports of randomized trials exaggerate
of the use of antithrombotic agents. Chest 1992; estimates of intervention effectiveness reported in
4: 102. a meta-analysis? Lancet 1998; 352: 609–13.
Assessing the Quality of Reports of RCTs 425
15. Khan KS, Daya S, Collins JA, Walter S. Empiri- 25. Aaronson NK. Quality of life assessment in
cal evidence of bias in infertility research: over- clinical trials: methodologic issues. Controlled
estimation of treatment effect in cross-over trials Clinical Trials 1989; 10: 195S (Supplement).
using pregnancy as the outcome measure. Fertil. 26. Cho MK, Bero LA. The quality of drug studies
Steril. 1996; 65: 939–45. published in symposium proceedings. Ann. Intern.
16. Colditz GA, Miller JN, Mosteller F. How study Med. 1996; 124: 485–9.
design affects outcomes in comparison of therapy. 27. Lau J, Ioannidis JPA, Schmid CH. Summing
I: Medical. Stat. Med. 1989; 8: 441–5. up evidence: one answer is not always enough.
17. Miller JN, Colditz GA, Mosteller F. How study Lancet 1998; 351: 123–7.
design affects outcomes in comparisons of therapy. 28. Berlin JA, on behalf of the University of Pennsyl-
II: Surgical. Stat. Med. 1989; 8: 455–66. vania meta-analysis blinding study group. Does
18. Chalmers TC, Celano P, Sacks HS, Smith H. Bias blinding of readers affect the results of meta-
in treatment assignment in controlled clinical analyses? Lancet 1997; 350: 185–6.
trials. N. Engl. J. Med. 1983; 309: 1358–61. 29. McNutt RA, Evans AT, Fletcher RH, Fletcher
19. McDowell I, Newell C. Measuring health: a SW. The effects of blinding on the quality of peer
guide to rating scales and questionnaires, 1st edn. review. JAMA 1990; 263: 1371–6.
Oxford: Oxford University Press, 1987. 30. Justice AC, Cho MK, Winker MA, Berlin JA,
Rennie D, Berkwits M, Callaham M, Fontanarosa
20. Streiner DL, Norman GR. Health measurement
P, Frank E, Goldman D, Goodman S, Pitkin R,
scales, 1st edn. Oxford: Oxford University Press,
Varma R, Waeckerle J. Does masking author
1989.
identity improve peer review quality? A random-
21. Moher D, Forin P, Jadad AR, Juni P, Klassen T,
ized controlled trial. JAMA 1998; 280: 240–2.
Lelorier J, Liberati A, Linde K, Penna A. Com-
31. VanRooyen S, Godlee F, Evans S, Smith R,
pleteness of reporting of trials published in lan- Black N. The effect of blinding and unmasking
guages other than English: implications for con- on the quality of peer review: a randomized
duct and reporting of systematic reviews. Lancet controlled trial. JAMA 1998; 280: 234–7.
1996; 347: 363–6. 32. Kramer MS, Feinstein AR. The biostatistics of
22. Egger M, Zellweger-Zahner T, Schneider M, concordance. Clin. Pharm. Ther. 1982; 29:
Junker C, Lengeler C, Antes G. Language bias in 111–23.
randomised controlled trials published in English 33. Shrout PE, Fleiss JL. Intraclass correlations: uses
and German. Lancet 1997; 350: 326–9. in assessing rater reliability. Psychol. Bull. 1979;
23. Khan KS, Daya S, Jadad AR. The importance of 86: 420–8.
quality of primary studies in producing unbiased 34. Naylor CD. Meta-analysis and the meta-
systematic reviews. Arch. Intern. Med. 1996; 156: epidemiology of clinical research. Br. Med. J.
661–6. 1997; 315: 617–19.
24. Moher D, Cook DJ, Jadad AR, Tugwell P, Moher 35. Begg C, Cho M, Eastwood S, Horton R, Moher
M, Jones A, Pham B, and Klassen TP. Assessing D, Olkin I, Pitkin R, Rennie D, Schultz KF, Simel
the quality of randomized controlled trials: impli- D, Stroup D. Improving the quality of reporting
cations for the conduct of meta-analyses. Health of randomized controlled trials: The CONSORT
Technology Assessment 1999; 3: 1–98. statement. JAMA 1996; 276: 637–9.
24
Consensus Development Methods, and
their Use in Creating Clinical Guidelines
based evidence exists or may ever exist.1 In such How Do Individuals and Groups Make
situations, the development of guidelines will Decisions?
inevitably have to be based partly or largely on
the opinions and experience of clinicians and
Consensus decision-making is a complex pro-
others with knowledge of the subject at issue.2,3
There are two main ways that judgement cess which involves decisions at both individual
based guidelines could be devised: have the and group levels. Judgements of appropriateness
‘best person’ make the judgement, or have a require people to pay attention to new informa-
group do it. In theory, the advantages of a group tion, from literature or other group members for
decision are: a wider range of direct knowledge example, and they require them to draw on
and experience is brought to bear; the inter- information and experience from memory. They
action between members stimulates considera- must make judgements about the value of infor-
tion of a wide range of options and debate that mation; they must integrate the various pieces of
challenges received ideas and stimulates new information; and they must determine how the
ones; idiosyncrasies are filtered out (sometimes information relates to the judgement task. They
wrongly!); and, in terms of influencing the must also provide judgements of the appro-
behaviour of others, the group as a whole may priateness of an intervention. This may involve
carry more weight than any one individual. estimating the probabilities of outcomes and
Given the likely diversity of opinion that any assessing values of outcomes. These are largely
group of people may display when considering a cognitive aspects of decision-making. As well as
topic, methods are needed for organising sub- these cognitive aspects, there are also motiva-
jective judgements. Although a variety of meth- tional aspects. People may be more or less
ods exist (which are described below), they motivated to put effort into the process for
share the common objective of synthesising example. When the decision-making involves a
judgements when a state of uncertainty (differ- group of people there are added complexities
ences of opinion) exists. resulting from social and inter-personal pro-
Our concern is with the use of these methods cesses. People will try to influence one another
for developing clinical guidelines, but they have for example, and may respond to each other not
been used for several other purposes in the simply on the basis of the relevant task informa-
health sector including forecasting, conflict tion provided, but also in terms of the social and
resolution and prioritisation. Despite their wide- personal characteristics of the persons involved.
spread use, consensus development methods Research in psychology has been conducted
have been the subject of relatively little meth-
at four levels of analysis: intra-individual, inter-
odological research within the health field.4
individual (or intra-group), inter-group, and
They have, however, been the subject of a
considerable amount of investigation elsewhere, societal. Research on intra-individual processes,
in particular in the behavioural sciences, techno- generally carried out within cognitive and social-
logical and social forecasting literature,5,6 but cognitive psychology, examines cognitive pro-
this research has had little impact on their appli- cesses within individuals. Social psychology
cation in healthcare. focuses on inter-individual and inter-group
It is essential to be clear what consensus levels, looking at how individuals and groups
development is and is not. It is a process for interact. The societal level of analysis is gen-
making policy decisions, not a scientific method erally found in economics, history, and sociol-
for creating new knowledge. At its best, con- ogy, but also in social psychology. This level
sensus development merely makes the best use focuses on institutional and societal influences
of available information, be that scientific data on behaviour. Models of behaviour which have
or the collective wisdom of the participants. been generated from empirical research in social
Thus, although it may capture collective knowl- and cognitive psychology are directly relevant to
edge, it is inevitably vulnerable to the possibility consensus decision-making as follows:
of capturing collective ignorance.
The objectives of this chapter are to: (i) 1 Members of a consensus development group
identify the factors that shape and influence the are required to draw on information both
clinical guidelines that emerge from consensus from their own experience and from new
development methods; (ii) make recommenda- information presented within the group in
tions about best practice in the use of consensus order to make decisions.
development methods for producing clinical 2 How people attend to, organise and remem-
guidelines; and (iii) recommend further meth- ber information helps to understand what
odological research for improving the use of information is likely to be used, the impact
consensus development methods as a basis for that it will have, and the possible biases that
guideline production. may operate in decision-making.
428 The Advanced Handbook of Methods in Evidence Based Healthcare
3 Attempts to reach consensus will involve the (a) safety in numbers – several people are less
need for some people to change their posi- likely to arrive at a wrong decision than a
tions – a process involving persuasion and single individual;
social influence. (b) authority – a selected group of individuals
4 Behaviour within and between group mem- is more likely to lend some authority to the
bers may be influenced by the perceptions of decision produced;
the groups involved: do members of con- (c) rationality – decisions are improved by
sensus groups see themselves as members of reasoned argument in which assumptions
a common group with a common goal, or are are challenged and members forced to jus-
subgroup identities more salient, perhaps tify their views;
leading to a conflict of interest between (d) controlled process – by providing a struc-
consensus subgroup members? tured process formal methods can elim-
inate negative aspects of group decision
Why We Need ‘Formal’ Methods of making; and
(e) scientific credibility – formal consensus
Consensus Development methods meet the requirements of scien-
tific methods.
It is only since the 1950s that formal consensus
development methods have been used in the
health sector. This does not mean that collective Types of Formal Consensus
decisions were not made before then, simply Development Methods
that such decisions emerged through informal
methods. Indeed, over the past 40 years the vast Three main approaches have been used in the
majority of collective decisions in healthcare health field. In the 1950s the Delphi method was
have continued to be based on group meetings, introduced;12,13 this was followed by the use of
such as committees, which have been largely un- the nominal group technique (NGT) in the
structured with few formal rules or procedures. 1960s,14 and in 1977 the National Institute of
Such group discussions, sometimes termed Health in the USA introduced the consensus
‘free discussion’, ‘freely interacting’ or simply development conference.15 The major differ-
‘consensus’ groups involve bringing together a ences between these methods are described in
group of people to discuss a problem with the Table 1 and relate to: (i) whether a mailed
aim of reaching agreement. They are usually not questionnaire is used; (ii) whether individuals
instructed in how to reach a consensus, though make separate decisions ‘in private’ or not, and
they may be given simple instructions such as if so, the degree of anonymity; (iii) whether
not to criticise other members’ contributions. information on the group’s deliberations or
There may or may not be someone chairing interim decisions is fed back to the participants
the group. A jury is an example of this type of for reconsideration during the process; (iv)
group. whether there is face-to-face contact between
The case for using formal methods is based group members, and if so, whether it is struc-
on a number of assumptions about decision- tured or not; and (v) the method used to aggre-
making in groups:7–11 gate participants’ views.
* Although there may be no prearranged structure to the group interaction, groups may adopt formal rules, on
their own initiative.
Consensus Development Methods for Clinical Guidelines 429
One feature common to all the methods, when group feedback. This process may be repeated a
used as a basis for creating clinical guidelines, is number of times. The judgements of participants
the use of cues. Cues are the dimensions or are then statistically aggregated, sometimes after
indications that group members are asked to weighting for expertise.
take into account when making their decisions.
For example, if participants were deciding on
the appropriate use of a treatment, one of the
Nominal group technique (NGT)
cues they would need to consider would be the The aim of NGTs is to structure interaction
severity of the condition being treated. Others within a group. First, each participant records
might include age, gender and co-morbidity. his or her ideas independently and privately.
Some methods present cues to participants as One idea is collected from each individual in
part of a scenario or vignette – a description of a turn and listed in front of the group by the
situation. Participants are presented with a set of facilitator, continuing until all ideas have been
scenarios, each describing a different clinical listed. The NGT attempts to structure the inter-
situation, and are asked to decide on the appro- action that follows by means of a facilitator.
priateness of a particular intervention (investiga- Each idea is discussed in turn. Thus, all ideas
tion or treatment) in each. will be discussed, rather than focusing discus-
sion on only one or two ideas. Individuals then
Delphi method privately record their judgements or vote for
Participants never meet or interact directly. options. Further discussion and voting may take
Instead, they are sent questionnaires and asked place. The individual judgements are aggregated
to record their views. Commonly, participants statistically to derive the group judgement.
are initially asked to suggest the factors or cues The most commonly used method for clinical
that should be considered by the group. Having guideline production is a ‘modified NGT’ devel-
contributed to drawing up the agenda, the next oped by the RAND Corporation during the
stage is a questionnaire which seeks their indi- 1970s and 1980s, although the developers
vidual views about the items (usually by indicat- referred to it as a ‘modified Delphi’. Initially
ing on a Likert scale) that they and their co- individuals express their views privately via
participants have suggested. The responses are mailed questionnaires. These are fed back to the
collated by the organisers and sent back to group. They are then brought together as a
participants in summary form, usually indicating group to discuss their views, after which they
the group judgement and the individual’s initial again privately record their views on a ques-
judgement. Participants are given the opportu- tionnaire (Box 1). An example of the use of this
nity to revise their judgement in the light of the approach is described in Box 2.
Based on Bernstein SJ, Laouri M, Hilborne LH et al. 1992. Coronary angiography: a literature review and
ratings of appropriateness and necessity. RAND JRA-03.
430 The Advanced Handbook of Methods in Evidence Based Healthcare
Consensus development conference open meeting, possibly over the course of a few
days. Evidence is presented by various interest
A selected group (of around 10 people) is groups or experts who are not members of the
brought together to reach consensus about an decision-making group. The latter then retire to
issue. The format involves the participants in an consider the questions in the light of the evi-
Consensus Development Methods for Clinical Guidelines 431
dence presented and attempt to reach consensus. The decision on which methodological issues to
Both the open part of the conference and the focus was based on three criteria:
private group discussion are chaired.
1 the importance of the particular aspect to
consensus decision-making in the health
sector;
METHODS USED FOR THE REVIEW 2 the amount and quality of the literature
available on the particular aspect; and
Development of a Conceptual Model 3 the potential for offering practical guidance
for those conducting consensus development
To develop a conceptual model to guide and groups.
structure this review, we explored relevant Following group discussion, members of the
material in the health sector, the experiences of review group indicated which six cells they
members of the review group who had used would give priority (shaded in Figure 1).
consensus methods, and drew on psychological
theory and research.
Consensus development methods may be Selection of Material to be Reviewed
considered to involve three types of activity:
planning, individual judgement, and group inter- The amount and type of methodological research
action. These activities are not necessarily on consensus methods used within the health
clearly separated or ordered. In addition, five sector is very limited. For example, when con-
components can be identified: three inputs sensus groups are compared only a small num-
(questions, participants, information), the pro- ber of groups (often only two) are used. In
cess (consensus development method) and the addition, many basic issues concerning group
output. As with the three activities, the compo- decision-making have not been addressed at
nents are also interrelated. all.
By combining these five components and the In contrast, vast amounts of research have
three activities, a matrix was formed which is been conducted in social and cognitive psychol-
shown in Figure 1. Each cell in the matrix ogy which address underlying issues involved
describes a component in a particular activity. in consensus development methods, whilst not
Brief descriptions of some of the important necessarily directly examining these methods.
features of each cell are included in the matrix. However, many aspects of these studies (the
approach, subjects, tasks) mean that while the is an attempt to describe the level of agreement.
research is relevant, it can only throw an indirect Whilst the first of these is the principal goal
light on consensus development within the when drawing up clinical guidelines, the latter is
health sector. It was necessary, therefore, to also of interest. There is value in differentiating
glean general ideas and findings from this wider between areas of clinical practice in which there
literature rather than pursue specific findings of is close, moderate or little agreement.
particular pieces of research which may be of
little relevance in the health sector. How can ‘consensus’ be defined?
The majority of the literature reviewed came
from published sources. Most was identified The answer depends on which of the two objec-
through searches of electronic bibliographic tives is being addressed. When prioritisation is
databases, though the reference lists of retrieved the objective, the output of the consensus method
articles were also used. Each of the six cells will typically take the form of a rank ordering of a
reviewed drew on the results of the principal set of alternatives. Each participant’s scores or
general search. For some cells, further focused ranks are pooled to arrive at a group ranking.
searches were used. The production of clinical guidelines gen-
The general criteria for including a paper in erally involves weighing the balance of benefits
the review were that it dealt with a methodo- and risks in order to estimate the parameter,
logical aspect of consensus decision-making and such as the effectiveness of a treatment for a
it was relevant to consensus decision-making series of different categories of patient.
in the health sector. Five electronic databases Technical questions (judgement needed be-
were searched: Medline (1966–1996), PsychLIT cause of insufficient data) need to be distin-
(1974–1996), Social Science Citation Index guished from value questions (judgement needed
(1990–1996), ABI inform, and Sociofile. From about competing social goals), because in gen-
the searches and reference lists of articles a total eral they should involve different types of
of 177 empirical and review articles were selected participants. This is an important distinction as
for review. there can be no ‘correct’ answer with value
questions, but for technical questions there is a
correct, if undiscovered, answer. The distinction
Some Preliminary Considerations between these two types of question is, how-
ever, often ignored.
Before starting the review, three fundamental
questions had to be considered: what are the
objectives of consensus development methods?; How can the validity of consensus
how can ‘consensus’ be defined?; and how judgements be determined?
can the validity of consensus judgements be How do we ensure that in general we make – or
determined? recognise – ‘good’ judgements? The nature of
a ‘good’ judgement depends critically on the
What are the objectives of consensus question being asked and, by definition, we can
development methods? rarely know whether any particular judgement is
One objective is to arrive at a single statement a good one at the time. The best we can do is try
or set of statements that all participants accept to identify a method of arriving at judgements
(or at least no one disagrees strongly enough to that will, on average: (i) produce more good
veto the agreement). Clearly, if participants per- judgements than other methods; or (ii) produce
sist in disagreeing, the consensus statement(s) fewer bad judgements than other methods.
will have little or no content. Where the process Although we might not be able to say at the
goes through a number of stages, participants time whether a particular decision is good or
have the opportunity to revise their views in the bad, we might – if we can identify a method
light of discussion and new information. This with these properties – be able to support the use
allows them to identify which aspects of their of that method over alternatives. This still leaves
position are relatively unimportant to them and us with the problem of how to evaluate whether
so can be abandoned. a method produces good or bad results, and in
The other type of objective is to identify any turn, some way of assessing what a good or bad
‘central tendency’ among the group and the result is.
degree of spread of opinion around it. Again the There are five possible ways of assessing
consensus development process may give par- validity:
ticipants the opportunity to revise their views in 1 Comparison with ‘gold standard’. Test the
the light of discussion and new information. method on questions which have correct
Thus, on the one hand there is an attempt to answers which the participants do not know
facilitate consensus, and on the other hand there with any precision. For example, with
Consensus Development Methods for Clinical Guidelines 433
almanac questions of the type ‘What is the questions (such as, in what circumstances would
diameter of Jupiter?’ you recommend treatment X?) or a more
2 Predictive validity. In forecasting it is pos- detailed set of subquestions in which hundreds
sible to look at whether the forecast that was of possible clinical scenarios may be considered.
made ‘came true’. Consensus development conferences usually ask
3 Concurrent validity. If a decision conflicts only a few questions. For the Delphi method and
with research-based evidence, without good the NGT, the task usually involves judgements
reason, we can say that it is invalid.16 over a wide range of scenarios.
4 Internal logic. An alternative concurrent
approach is to look at the internal logical Do the particular cues included in the
order of a consensus group’s output. In this
way, the consistency of the decisions can be question influence judgement?
determined.17 The particular cues included in a question
5 Usefulness in terms of commitment and addressed by consensus development groups
implementation. A good decision-making influence individual and group judgements.18
process might be seen to be one that pro- The use of explicit cues in the design may lead
duces results that are not only correct but participants to ignore or undervalue other cues
also influential. However, this is a difficult that they may otherwise have included. This
definition to defend because whether guide- applies both to specific clinical cues and to
lines are used or not may have little to do general contextual cues, such as the level of
with the quality of the decision. resources available.19–22 Table 2 illustrates how
Thus, for clinical guidelines there is no absolute the higher overall levels of funding for healthcare
means for judging at the time whether a decision in the USA is associated with US physicians’
is valid, and thus whether a particular method judgement that more clinical situations are
for producing consensus is valid. This leaves us appropriate indications for treatment than was
with the problem of how to evaluate different true for British physicians. Individuals may
consensus development methods. Our focus has not be aware of the importance they attach to
been to look at the factors which are likely to specific cues and, when they are aware, the
influence the process of consensus development, importance they actually attach to a particular
and where possible how they might influence cue may differ from the importance they believe
the outcome. they attach to it.17,23–25 This may lead to difficul-
ties in resolving differences during interaction.
Appropriate 78 86 41 85 69 33
Equivocal 3 63 91 4 25 42
Inappropriate 0 5 113 0 2 40
The proportion of scenarios deemed appropriate is higher in the US where health care expenditure is much greater
than in the UK.19
Note: larger numbers in top right than bottom left of each table indicates lower thresholds for intervening in the
US.
434 The Advanced Handbook of Methods in Evidence Based Healthcare
groups considering the appropriateness of health- questions about the type of participant and the
care interventions, it suggests that the way a composition of the group. Most writers suggest
question is posed may influence judgements that consensus development groups should be
about appropriateness. composed of people who are expert in the
The level of detail specified in a task affects appropriate area and who have credibility with
not only the options participants may consider the target audience.15,33,34 This raises the ques-
but also their estimation of probabilities.26–28 tion of what constitutes an expert. Clinicians
Although research has not been conducted on have clinical expertise, researchers have scien-
this issue within consensus judgements of tific expertise, and lay people or patients have
appropriateness, it is possible that judgements expertise from having experienced the impact of
derived from general questions may differ from the condition or intervention. All of these may
those derived from specific questions. Differ- be required. Once this has been decided, ques-
ences may also occur between appropriateness tions about the procedures for selecting, or
ratings derived when the starting point is a type sampling, individuals need to be considered.
of patient and those derived when the starting
point is an intervention. To what extent is a group decision affected
Does the way judgements are elicited by the particular individuals who
influence those judgements? participate?
Little is known about the representativeness of
The answers obtained from a group may be
participants in consensus development groups.35
influenced by the way in which the judgement is
Knowledge of the impact of individuals is also
elicited. Global views on appropriateness do not
limited by a lack of research. Studies that have
necessarily reflect participants’ probability esti-
been performed, listed in Table 3, suggest that
mates of different outcomes.29–31 This is because
the selection of individuals has some, though not
global views also take values or utilities of differ-
ent outcomes into account. There is little work on a great deal of influence on outcome.36 Some
this within consensus decision-making. studies show similarity between similarly com-
posed panels,1,37 others show differences.38
Table 4 illustrates this – appropriateness ratings
Does the level of comprehensiveness or of 82 out of 96 scenarios were the same for two
selectivity of the scenarios affect groups, each of which was composed of people
judgement? with the same disciplinary background. In most
Given that groups may experience greater agree- studies the number of groups compared is small,
ment and less disagreement when they are and thus the findings are weak. The particular
restricted to considering scenarios with which tasks and procedures used also have some
shortcomings.
they are familiar, the temptation to be compre-
Although the Delphi method has received
hensive by including every theoretically pos-
some attention, very few studies have compared
sible scenario may be counter-productive.32
Also, being selective allows larger numbers of similarly composed groups using a NGT. Thus,
cues to be considered explicitly, which should any definitive statement about whether similar
be an advantage in terms of reliability. groups will produce similar results is not pos-
sible. The particular application of the method
is likely to be an important factor. We do not
Participants expect that, as a general class, ‘questionnaires’
will be equally reliable for all samples of
There are essentially two stages to decisions respondents. Rather, we assess the reliability of a
about who to include as participants in a con- particular questionnaire. So too, with consen-
sensus development group. The first involves sus instruments; their reliability across samples
Kastein et al. (1993)36 Mixed (GPs and specialists): abdominal pain 1 constipation
Duffield (1993)37 Nurse managers: competencies
Chassin (1989)1 Physicians: appropriate indications for coronary angiography
Pearson et al. (1995)38 GPs: sinusitis, dyspepsia
Brown and Redman (1995)39 Mixed (doctors): women’s health issues
Penna et al. (1997) (pers. comm.) Mixed (doctors): breast cancer
Consensus Development Methods for Clinical Guidelines 435
Appropriate 59 6 0
Equivocal 2 4 4
Inappropriate 0 2 19
Ratings of 82 out of 96 scenarios were the same for the group convened
in 1984 as it was for the 1988 group (Chassin, 1989).1
seems highly dependent on the particular appli- likely to determine the final decision.46 If there
cation of the method. is an initial consensus, a shift may occur in
which the final decision is more extreme.47 If
What effect does heterogeneity in group there is a split view initially, members will tend
to move towards one another,48,49 but this
composition have on group judgement? depends on the degree to which those with
The weight of evidence suggests that hetero- differing views form cohesive subgroups.50 The
geneity in a decision-making group can lead to a more cohesive the subgroups, the less chance of
better performance than homogeneity.40 There achieving consensus and the more chance that
is, however, some evidence that heterogeneity there may be polarisation.
may have an adverse effect because of conflict
that may arise between diverse participants.41,42 Do different categories of participants
The effect of heterogeneity depends to some
extent on the task being undertaken. produce different results?
Studies, although few in number, show that
Which personal characteristics are differences in group composition may provide
different judgements. This issue has been
important influences on group decisions? addressed in two ways (Table 5). There have
The status of participants affects their degree of been four studies in which groups of people with
influence on the group.43 In groups of diverse different disciplinary composition have been
status, those with higher status exert more influ- compared, and five studies in which different
ence.44 In more homogeneous groups, group disciplinary subgroups within a group have been
decisions tend to reflect the majority view.45 compared. The overall finding is that members
Initial opinions of participants affect the of a specialty are more likely to advocate treat-
group process. If there is a majority view, this is ments that involve their specialty.20,36,51–56 Table
Table 5 Studies that compared (a) groups of different composition, and (b)
subgroups of mixed groups
(a) Groups of different composition
Lomas and Pickard (1987)57 Physicians versus patients (communication)
Scott and Black (1991)51 Mixed versus surgeons (cholecystectomy)
Leape et al. (1992)52 Mixed versus surgeons (carotid endarterectomy)
Coulter et al. (1995)53 Mixed versus chiropractics (spinal manipulation)
* Those who perform the procedures tend to rate more indications or scenarios
as appropriate than do colleagues in the same specialty or primary care
physicians (Kahan et al. 1996).55
CABG 5 coronary artery bypass grafting.
Perform 5 physicians who perform the procedure; Related 5 physicians in same
specialty but do not perform procedure; Primary 5 primary care physicians.
Consensus Development Methods for Clinical Guidelines 437
The extraction and synthesis of pertinent How does information influence individual
information for use by consensus development decision-making?
groups is a major challenge.15 Without such
reviews participants are more likely to rely on The information presented to individuals is an
their own particular experiences. Anecdotal evi- important element in decision-making. Informa-
dence suggests that when literature reviews have tion can influence judgement in a variety of ways.
been provided, the evidence was used both in The way information is presented can influence
deliberation and in the decisions, making the the likelihood of it being used and how it is
consensus process easier.67 The provision of used.68 The particular way in which information
information to participants varies from expert is framed can also influence judgement.69–73 This
groups which rely on the expertise and knowl- can be seen in Table 8, which reports on the
edge of group members to those which provide a level of support by an English health authority
comprehensive synthesis of relevant research.34 (purchaser) for two policy proposals – the provi-
Sometimes it may be decided not to use formal sion of mammography and of cardiac rehabilita-
information because the available evidence is so tion. People tend to be risk averse for losses and
weak.15 risk seeking for gains.74–76 Individuals’ own
Table 8 Mean score (%) (95% CI) on two health policy issues made by health
authority members when same results presented by four different methods
Method of data presentation Mammography Cardiac rehabilitation
* The number of scenarios changed from the first to the second round making comparisons difficult.
CABG 5 coronary artery bypass grafting.
Van de Ven and Delbecq (1974)95 Informal 5 t Idea generation NGT and Delphi
NGT 5 e-f-t-e better than
Delphi 5 e-f-e informal
White et al. (1980)96 Informal 5 t Idea generation
NGT 5 e-f-t-e NGT best
Structured 5 t
Jarboe (1988)97 NGT 5 e-f-t-e Idea generation
Structured 5 t NGT best
this question.104–106 However, of the many fac- Should the judgements of participants be
tors which can influence decision-making, the weighted when using an explicit method?
environment is likely to have only a marginal
impact except in extreme circumstances.107 Although weighting by expertise may seem
attractive in theory, the benefits are uneven and
Do the characteristics of a group difficult to predict.117 In practice, it is unclear
how weights should be assigned to different
facilitator affect the consensus decision? participants.118 Inappropriate weights may be
Although work on leadership suggests that worse than no weights.
aspects of leader behaviour are important for
group decision-making,108,109 the models of lead- How should individuals’ judgements be
ership used are often not directly transferable to
facilitation.10 There is little work which exam-
aggregated for any one scenario in an
ines what a good facilitator is and very little explicit method?
work which looks at the effects of facilitation on The appropriate method of aggregating individ-
group decision-making. However, it is likely ual judgements will depend on whether the
that this key role will influence group decision- objective is to identify consensus (rather than a
making.44,110,111 lack of consensus) or the nature and extent of
consensus. Use of a frequency distribution avoids
arbitrary value judgements. Criteria for choosing
Outputs: Methods of Synthesising a method include robustness, accuracy and fair-
Individual Judgements ness. In other contexts, the median and inter-
quartile range have been shown to be robust.119
The nature of the outputs of a consensus devel-
opment process will be determined during its
planning, along with decisions as to whom the How should consensus be defined in an
outputs are aimed. The process of arriving at explicit, quantitative method?
these decisions will reflect and often help to Outcome is partly dependent on the way con-
refine the objectives of the exercise. There may sensus is defined.58 Table 11 shows how relax-
be further revisions when individual judgements ing the definition of agreement makes little
are being made in the light of participants’ difference to the proportion of scenarios or state-
opinions about what they are being asked to do. ments a group agrees about. In contrast, Table
In formal consensus development methods 12 shows how relaxing the definition of dis-
individual judgements may remain confidential, agreement has a marked effect on the proportion
aggregation is by explicit methods including of scenarios about which a group disagrees.
some indication of the individuals’ degree of Excluding the judgements of outliers also has
agreement, and the group’s deliberations are a dramatic effect on both levels of agreement
controlled by externally imposed rules. The pro- and disagreement.120 Although a number of
cedures involved in aggregation fall into two analytical methods exist, there is no agreed
distinct stages. In the first, differential weights standard.121,122 Analyses should recognise the
may be attached to the contributions of the difference between reliability and agreement.
various participants. (Weights may be fixed for
all questions, or vary from one question to
another.) In the second, group scores and indices
of spread are calculated from each set of partici- IMPLICATIONS AND RECOMMENDATIONS
pants’ assessments, weighted or otherwise.
Although a considerable amount of research has
Is an implicit approach to aggregation of been carried out on consensus development
individuals’ judgements sufficient? methods,123 many aspects have not been invest-
igated sufficiently. For the time being at least,
Voting is only suitable for choosing, ranking advice on those aspects has, therefore, to be
or prioritising options rather than assigning based on the user’s own commonsense and the
specific values to each scenario or statement. experience of those who have used or partici-
Research in this field suggests that the more pated in these methods. To avoid confusion, the
demanding the rules, the less likely a consensus extent to which research support for any guid-
will be achieved.112,113 The very act of voting ance on good practice is indicated as:
and the timing of a vote can affect the group
decision.114 Voting systems are subject to sev- A 5 clear research evidence;
eral paradoxes that might undermine the validity B 5 limited supporting research evidence;
of the outcome.115,116 C 5 experienced commonsense judgement.
Consensus Development Methods for Clinical Guidelines 441
Strict 5 all ratings in one of three predefined ranges 1–3, 4–6, 7–9; Relaxed 5 all ratings within any 3-point range.
NA 5 no analysis reported.
B and C should not be regarded as necessarily given the opportunity to say which cues they
unsatisfactory, as some aspects of the conduct consider important. Doing so may help main-
of consensus development methods are not tain their participation and help them justify
amenable to scientific study but can be adequately their judgements. Participants’ views of the rele-
justified on the basis of experience. vant cues can usefully be obtained during a
preliminary round of individual open-ended
consultations. [C]
Setting the Task(s) or Question(s) to be Contextual cues (such as whether judgements
Addressed should assume unlimited healthcare resources or
the reality of restricted resources) are as import-
Considerable care must be given to the selection ant as those specific to the topic. Participants are
of the cues – the characteristics that influence likely to make differing assumptions about these
clinical management, such as the severity of cues if they are not specified in the task, so it is
the patient’s condition. Participants should be important to make them explicit. [B]
442 The Advanced Handbook of Methods in Evidence Based Healthcare
The focus of the task may either be ways of of agreement, then groups should be homo-
managing a particular condition or the indica- geneous in composition. If, in contrast, the aim
tions for using an intervention. If the latter, care is to identify and explore areas of uncertainty, a
needs to be taken as to how other relevant heterogeneous group is appropriate. [B]
interventions are dealt with because views of In judgements of clinical appropriateness, the
the appropriateness of any intervention will most influential characteristic of participants is
depend on whether or not the possibility of their medical speciality. Specialists tend to
using alternative interventions is taken into favour the interventions they are most familiar
account. Ambiguity as to whether or not altern- with. Consensus based guidelines should there-
atives should be considered can affect group fore be interpreted in the context of the specialty
consensus. For example, participants may be composition of the group. [A]
instructed either to make their judgements in In general, having more group members will
terms of the intervention under consideration increase the reliability of group judgement.
being more or less appropriate than other inter- However, large groups may cause co-ordination
ventions in general or than some specified inter- problems. Although it is theoretically likely that
vention(s). [C] group size will affect decision-making, the
Participants may be asked to provide an over- effects are subtle and difficult to detect. Below
all judgement (for example, treatment is or is about six participants, reliability will decline
not appropriate in a particular scenario) or an quite rapidly, while above about 12, improve-
attempt can be made to break the judgement ments in reliability will be subject to diminish-
down into probability and utility estimates. ing returns. [B]
There are theoretical advantages with the latter,
but it is likely to be a more difficult task for Choosing and Preparing the Scientific
some participants (probability estimates are a
matter for technical experts, though utility
Evidence
estimates may appropriately involve others, Such research-based information as is available
including patients) and it is unclear whether it should be provided to all participants at an early
enhances judgements. [C] stage because:
Although including all theoretically possible
scenarios will increase the comprehensiveness (a) if all members of the group have access to
of the exercise, if many of the scenarios rarely such information it is more likely to be
occur in practice, the increased burden on the discussed within the group;
participants may not be justified by the limited (b) providing a literature review to group
value of the information provided. It is also members before discussion may enhance
likely that judgements of scenarios which never the perception that the task is research
or rarely occur are less reliable than judgements based, which will in turn encourage mem-
of scenarios which more commonly occur in bers to be more reliant on information;
practice [B]. Further, requiring participants to and
judge what may be seen as numerous irrelevant (c) providing a common starting point may
scenarios may alienate them from the task. [C] foster group cohesion. [C]
Participants should be encouraged to bring the
review and any personal notes to the group
Selecting the Participants sessions as a memory aid. Information presented
in the form of articles or abstracts is less easily
Within defined specialist or professional cat- assimilated than information presented in a syn-
egories, the selection of the particular individ- thesised form, such as tables. Presenting infor-
uals is likely to have little impact on the group mation in an easy-to-read and understandable
decision as long as the group is of sufficient format may mean that participants are more
size. To enhance the credibility and widespread likely to use it. Tabulating the information in a
acceptance of the guidelines, it is advisable way which increases the salience of the cues to
that the participants reflect the full range of be used for making judgements means that the
key characteristics of the population they are information is more likely to be processed in
intended to influence. The basis for selection this manner. [C]
should also be seen to be unbiased. [C] Those with expertise in judging the quality of
Whether a homogeneous or heterogeneous scientific research should be involved in con-
group (defined in terms of specialist or pro- ducting any literature review. Organisation of
fessional characteristics) is best will in part materials by methodologists can highlight those
depend on the purpose of the exercise. If the aim factors which are most relevant to making a
is to define common ground and maximise areas judgement, thus making it more likely that
Consensus Development Methods for Clinical Guidelines 443
group is desirable. Do minority views get prop- 8. Asch SE. Studies of independence and con-
erly aired? If so, are they eliminated during formity: A minority of one against a unanimous
discussion or by the process of aggregation? majority. Psychological Monographs 1956; 70
(9, Whole No. 416).
4 What effect does research based information
9. Zajonc RB. Social facilitation. Science 1965;
have on individual and on group judgements?
149: 269–74.
Does the effect depend on the amount of
10. Janis I. Groupthink, 2nd edn. Boston: Houghton-
information or how it is presented?
Mifflin, 1982.
This could be examined by studying group inter- 11. Diehl M, Stroebe W. Productivity loss in brain-
actions or by examining the effect of prompts storming groups: Toward the solution of a
from the facilitator regarding the information. riddle. J. Personality Soc. Psychol. 1987; 53:
Studies could compare groups with and with- 497–509.
out literature reviews, or with the evidence 12. Dalkey NC, Helmer O. An experimental ap-
presented in different ways (written, oral plication of the Delphi method to the use of
presentation, video) to determine if the informa- experts. Management Science 1963; 9: 458–67.
tion is exchanged within groups and whether it 13. Pill J. The Delphi method: substance, context, a
affects the outcome. What is the optimal amount critique and the annotated bibliography. Socio-
of information that should be presented? What economic Planning Science 1971; 5: 57–71.
techniques can be used to make the information 14. Delbecq A, Van de Ven A. A group process
presented more accessible? model for problem identification and program
5 What effect does the method of feedback planning. J. Appl. Behav. Sci. 1971; 7: 467–92.
of participants’ views have on group 15. Fink A, Kosecoff J, Chassin M, Brook RH.
judgement? Consensus methods: Characteristics and guide-
lines for use. Am. J. Public Health 1984; 74:
Studies of different ways of providing feedback 979–83.
on group views to individuals are needed. These 16. Merrick NJ, Fink A, Park RE, Brook RH,
could examine the effect on outcome, inter- Kosecoff J, Chassin MR, Solomon DH. Deriva-
individual understanding and discussion within tion of clinical indications for carotid endarter-
the group. Using a Delphi method, what differ- ectomy by an expert panel. Am. J. Public Health
ence does it make if the feedback is attributed to 1987; 77: 187–190.
a named individual or not? 17. Hunter DJW, McKee CM, Sanderson CFB,
Black NA. Appropriate indications for prosta-
tectomy in the UK – results of a consensus
REFERENCES panel. J. Epidemiol. Community Health 1994;
48: 58–64.
1. Chassin M. How do we decide whether an 18. Fischoff B, Slovic P, Lichtenstein S. Fault trees:
investigation or procedure is appropriate? In: sensitivity to estimated failure probabilities to
Hopkins A. (ed.). Appropriate investigation and problem representation. J. Exp. Psychol. Hum.
treatment in clinical practice. London: Royal Percept. Performance 1978; 4: 330–44.
College of Physicians, 1989. 19. Brook RH, Kosecoff JB, Park E, Chassin MR,
2. AHCPR. AHCPR Clinical Practice Guideline Winslow CM, Hampton JR. Diagnosis and treat-
Program. Report to Congress. US Department of ment of coronary heart disease: comparison of
Health and Human Services, AHCPR, 1995. doctors’ attitudes in the USA and the UK.
3. Mann T. Clinical guidelines. Using clinical Lancet 1988; i: 750–3.
guidelines to improve patient care in the NHS. 20. Fraser GM, Pilpel D, Hollis S, Kosecoff J,
London: Department of Health, 1996. Brook RH. Indications for cholecystectomy: the
4. Black NA. Appropriateness of medical care in results of a consensus panel approach. Qual.
Europe: a commentary. Int. J. Qual. Health Care Ass. Health Care 1993; 5: 75–80.
1994; 6: 231–2. 21. Imamura K, Gair R, McKee M, Black N. Appro-
5. Parente FJ, Anderson-Parente JK. Delphi priateness of total hip replacement in the United
inquiry systems. In: Wright G, Ayton P (eds). Kingdom. World Hospitals & Health Services
Judgmental Forecasting. Chichester: Wiley, 1997; 32: 10–14.
1987. 22. Bernstein SJ, Hofer TP, Meijler AP, Rigter H.
6. Rowe G, Wright G, Bolger F. Delphi: a re- Setting standards for effectiveness: A com-
evaluation of research and theory. Technological parison of expert panels and decision analysis.
Forecasting and Social Change 1991; 39: Int. J. Qual. Health Care 1977; 9: 255–64.
235–51. 23. Brehmer B, Joyce CRB (eds). Human Judge-
7. Sherif M. An experimental approach to the study ment: the SJT view. Amsterdam: Elsevier,
of attitudes. Sociometry 1937; 1: 90–8. 1988.
Consensus Development Methods for Clinical Guidelines 445
24. Brown RL, Brown RL, Edwards JA, Nutz JF. 38. Pearson SD, Margolis CZ, Davis C, Schreier
Variation in a medical faculty’s decisions to LK, Sokol HN, Gottlieb LK. Is consensus repro-
transfuse. Implications for modifying blood ducible? A study of an algorhythmic guidelines
product utilization. Medical Care 1992; 30: development process. Medical Care 1995; 33:
1083–93. 643–60.
25. Evans JStBT, Harries C, Dennis I, Dean J. 39. Brown WJ, Redman S. Setting targets: a three-
General practitioners’ tacit and stated policies in stage model for determining priorities for health
the prescription of lipid lowering agents. Br. J. promotion. Aust. J. Publ. Health 1995; 19:
Gen. Pract. 1995; 45: 15–18. 263–9.
26. Tversky A, Koeler DJ. Support theory: a non- 40. Jackson SE. Team composition in organizational
extensional representation of subjective prob- settings: issues in managing an increasingly
ability. Psychol. Rev. 1994; 101: 547–67. diverse work force. In: Worchel S et al. (eds).
27. Redelmeier DA, Koehler DJ, Liberman V, Group Processes and Productivity. London:
Tversky A. Probability judgment in medicine: Sage, 1992.
discounting unspecified possibilities. Medical 41. Maznevski ML. Understanding our differences:
Decision Making 1995; 15: 227–30. Performance in decision-making groups with
28. Redelmeier DA, Shafir E. Medical decision diverse members. Human Relations 1994; 47:
making in situations that offer multiple altern- 531–52.
atives. JAMA 1995; 273: 302–5. 42. Guzzo RA, Dickson MW. Teams in organiza-
29. Oddone EZ, Samsa G, Matchar DB. Global tions: recent research on performance and effec-
judgments versus decision-model-facilitated tiveness. Annu. Rev. Psychol. 1996; 47: 307–38.
judgments: are experts internally consistent? 43. Levine JM, Moreland RL. Progress in small
Medical Decision Making 1994; 14: 19–26. group research. Annu. Rev. Psychol. 1990; 41:
585–634.
30. McClellan M, Brook RH. Appropriateness of
44. Vinokur A, Burnstein E, Sechrest L, Wortman
care – a comparison of global and outcome
PM. Group decision making by experts: Field
methods to set standards. Medical Care 1992;
study of panels evaluating medical technologies.
30: 565–86.
J. Personality Soc. Psychol. 1985; 49: 70–84.
31. Silverstein MD, Ballard DJ. Expert panel assess-
45. Kirchler E, Davis JH. The influence of member
ment of appropriateness of abdominal aortic
status differences and task type on group con-
aneurysm surgery: global judgment versus prob-
sensus and member position change. J. Person-
ability estimates. J. Health Serv. Res. Policy
ality Soc. Psychol. 1986; 51: 83–91.
1998; 3: 134–40.
46. Davis JH, Kerr NL, Atkin RS, Holt R, Meek D.
32. Park RE, Fink A, Brook RH, Chassin MR, Kahn
The decision processes of 6- and 12-person
KL, Merrick NJ, Kosecoff J, Soloman DH. mock juries assigned unanimous and two thirds
Physician ratings of appropriate indications for majority rules. J. Personality Soc. Psychol.
three procedures: theoretical indications vs indi- 1975; 32: 1–14.
cations used in practice. Am. J. Public Health 47. Williams S, Taormina RJ. Unanimous versus
1989; 79: 445–7. majority influences on group polarization in
33. Jones, J. & Hunter, D. Consensus methods for business decision making. J. Soc. Psychol. 1993;
medical and health services research. Br. Med. J. 133: 199–205.
1995; 311: 377–80. 48. Vinokur A, Burnstein E. Novel argumentation
34. Lomas J. Words without action? The production, and attitude change: The case of polarization
dissemination, and impact of consensus recom- following group discussion. Eur. J. Soc. Psychol.
mendations. Annu. Rev. Public Health 1991; 12: 1978; 11: 127–48.
41–65. 49. Vinokur A, Burnstein E. The depolarization of
35. McKee M, Priest P Ginzler M, Black N. How attitudes in groups. J. Personality Soc. Psychol.
representative are members of expert panels? 1978; 36: 872–85.
Qual. Ass. Health Care 1991; 3: 89–94. 50. Whitney JC, Smith RA. Effects of group co-
36. Kastein MR, Jacobs M, Van der Hell RH, Luttik hesiveness on attitude polarization and the
K, Touw-Otten FWMM. Delphi, the issue of acquisition of knowledge in a strategic planning
reliability: A qualitative Delphi study in primary context. J. Market. Res. 1983; 20: 167–76.
health care in the Netherlands. Technological 51. Scott EA, Black N. When does consensus exist
Forecasting and Social Change 1993; 44: in expert panels? J. Public Health Med. 1991;
315–23. 13: 35–9.
37. Duffield C. The Delphi technique: a comparison 52. Leape LL, Freshour MA, Yntema D, Hsiao W.
of results obtained using two expert panels. Small group judgment methods for determining
International Journal of Nursing Studies 1993; resource based relative values. Medical Care
30: 227–37. 1992; 30: 11 (suppl.) NS28–NS39.
446 The Advanced Handbook of Methods in Evidence Based Healthcare
53. Coulter I, Adams A, Shekelle P. Impact of acquisition strategies. J. Consumer Res. 1977; 3:
varying panel membership on ratings of appro- 233–40.
priateness in consensus panels: a comparison of 70. Jarvenpaa SL. Graphic displays in decision
a multi- and single-disciplinary panel. Health making – the visual salience effect. J. Behav.
Services Res. 1995; 30: 577–91. Decision Making 1990; 3: 247–62.
54. Park RE, Fink A Brook RH, Chassin MR, Kahn 71. Fahey T, Griffiths S, Peters TJ. Evidence based
KL, Merrick NJ, Kosecoff J, Solomon DH. purchasing: understanding results of clinical
Physician ratings of appropriate indications trials and systematic review. Br. Med. J. 1995;
for six medical and surgical procedures. Santa 311: 1056–60.
Monica, CA: RAND, R-3280-CWF/HF/PMT/ 72. Laupacis A, Sackett DL, Roberts R. An assess-
RJW, 1986. ment of clinically useful measures of the con-
55. Kahan JP, Park RE, Leape LL, et al. Variations sequences of treatment. N. Engl. J. Med. 1988;
by specialty in physician ratings of the appro- 318: 1728–33.
priateness and necessity of indications for pro- 73. Bucher HC, Weinbacher M, Gyr K. Influence of
cedures, Medical Care 1996; 34: 512–23. method of reporting study results on decision of
56. Zadinsky JK, Boettcher JH. Preventability of physicians to prescribe drugs to lower choles-
infant mortality in a rural community. Nursing terol concentration. Br. Med. J. 1994; 309:
Res. 1992; 41: 223–7. 761–4.
57. Lomas J, Pickard L. Patient versus clinician item 74. Tversky A, Kahneman D. The framing of
generation for quality-of-life measures. The case decisions and the psychology of choice. Science
of language-disabled adults. Med. Care 1987; 1981; 211: 453–8.
25: 764–9. 75. McNeil BJ, Pauker SG, Sox HC, Tversky A. On
58. Scott EA, Black N. Appropriateness of chol- the elicitation of preferences for alternative ther-
ecystectomy in the United Kingdom – a con- apies. N. Engl. J. Med. 1982; 306: 1259–62.
sensus panel approach. Gut 1991; 32: 1066–70. 76. McNeil BJ, Pauker SG, Tversky A. On the
59. Leape LL, Park RE, Kahan JP, Brook RH. framing of medical decisions. In: Bell D, Raiffa
Group judgements of appropriateness: the effect
H, Tversky A (eds). Decision making: Descrip-
of panel composition. Qual. Assur. Health Care
tive, normative and prescriptive interactions.
1992; 4: 151–9.
Cambridge: Cambridge University Press, 1988.
60. Lomas J, Anderson G, Enkin M, Vayda E,
77. Koehler JJ. The influence of prior beliefs
Roberts R, Mackinnon B. The role of evidence in
on scientific judgements of evidence quality.
the consensus process: Results from a Canadian
Organizational Behavior and Human Decision
consensus exercise. JAMA 1988; 259: 3001–5.
Processses 1993; 56: 28–55.
61. Richardson FMacD. Peer review of medical
78. Shanteau J. How much information does an
care. Medical Care 1972; 10: 29–39.
62. Shaw ME. Group Dynamics. The Psychology of expert use? Is it relevant? Acta Psychol. 1992;
Small Group Behavior. 3rd edn. New York: 81: 75–86.
McGraw-Hill, 1981. 79. Woudenberg F. An evaluation of Delphi. Tech-
63. McGrath JE. Small group research. Am. Behav. nological Forecasting and Social Change 1991;
Scientist 1978; 21: 651–74. 40: 131–50.
64. Nagao DH, Davis JH. Some implications of 80. Gowan JA, McNichols CW. The effects of
temporal drift in social parameters. J. Exp. Soc. alternative forms of knowledge representation
Psychol. 1980; 16: 479–96. on decision-making consensus. Int. J. Man
65. Davis JH. Some compelling intuitions about Machine Studies 1993; 38: 489–507.
group consensus decisions, theoretical and 81. Kaplan MF. The influencing process in group
empirical research, and interpersonal aggrega- decision making. In: Hendrick C (ed). Review of
tion phenomena: Selected examples, 1950–1990. Personality and Social Psychology, 8, Group
Organizational Behavior and Human Decision Processes. London: Sage, 1987.
Processes 1992; 52: 3–38. 82. Kaplan MF, Miller CE. Group decision making
66. McGrath JE. Groups: Interaction and perform- and normative vs. informational influence:
ance. Englewood Cliffs, NJ: Prentice-Hall, Effects of type of issue and assigned decision
1984. rule. J. Personality Soc. Psychol. 1987; 53:
67. Jacoby I. Evidence and consensus. JAMA 1988; 306–13.
259: 3039. 83. Schittekatte M. Facilitating information ex-
68. Payne JW, Bettman JR, Johnson EJ. Behavioral change in small decision-making groups. Eur. J.
decision research: A constructive processing Soc. Psychol. 1996; 26: 537–56.
perspective. Annu. Rev. Psychol. 1992; 43: 84. Stasser G. Pooling of unshared information dur-
87–131. ing group discussion. In: Worchel S, Wood W,
69. Bettman JR, Kakkar P. Effects of information Simpson J (eds). Group process and productivity.
presentation format on consumer information Newbury Park, CA: Sage, 1992.
Consensus Development Methods for Clinical Guidelines 447
85. Gustafson DH, Shukla RK, Delbecq A, Walstre 100. Rohrbaugh J. Improving the quality of group
GW. A comparative study of differences in judgement: social judgement analysis and the
subjective likelihood estimates made by individ- nominal group technique. Organ. Behav. Hum.
uals, interacting groups, Delphi groups, and Performance 1981; 28: 272–92.
nominal groups. Organ. Behav. Hum. Perform- 101. Nemiroff RM, King DD. Group decision-
ance 1973; 9: 280–91. making as influenced by consensus and self
86. Fischer GW. When oracles fail – a comparison orientation. Human Relations 1975; 28: 1–21.
of four procedures for aggregating subjective 102. Hall J, Watson WH. The effects of a normative
probability forecasts. Organ. Behav. Hum. Per- interaction on group decision-making perform-
formance 1981; 28: 96–110. ance. Human Relations 1970; 23: 299–317.
87. Brightman HJ, Lewis DJ, Verhoeven P. 103. Innami I. The quality of group decisions, group
Nominal and interacting groups as Bayesian verbal-behavior, and intervention. Organiza-
information processors. Psychol. Rep. 1983; 53: tional Behavior and Human Decision Processes
101–2.
1994; 60: 409–30.
88. Soon A, O’Connor M. The effect of group
104. Reagan-Cirincione P, Rohrbaugh J Decision
interaction processes on performance in time
conferencing: A unique approach to the behav-
series extrapolation. Int. J. Forecasting 1991; 7:
ioral aggregation of expert judgment. In: Wright
141–9.
89. Sniezek JA. A comparison of techniques for G, Bolger F (eds). Expertise and Decision Sup-
judgmental forecasting by groups with common port. New York: Plenum Press, 1992.
information. Group Organ. Stud. 1990; 15: 105. Argyle M, Dean J. Eye-contact, distance and
5–19. affiliation. Sociometry 1965; 28: 289–304.
90. Larreche JC, Moinpour R. Managerial judge- 106. Sommer R, Olsen H. The soft classroom.
ment in marketing: the concept of expertise. J. Environment and Behavior 1980; 5: 3–16.
Marketing Res. 1983; 20: 110–21. 107. Worchel S, Shackelford SL. Groups under
91. Herbert TT, Yost EB. A comparison of decision stress: the influence of group structure and
quality under nominal and interacting consensus environment on process and performance. Per-
group formats: the case of the structured prob- sonality Soc. Psychol. Bull. 1991; 17: 640–7.
lem. Decis. Sci. 1979; 10: 358–70. 108. Flowers ML. A laboratory test of some implica-
92. Erffmeyer RC, Lane IM. Quality and acceptance tions of Janis’s groupthink hypothesis. J. Per-
of an evaluative task: the effects of four group sonality Soc. Psychol. 1977; 35: 888–96.
decision-making formats. Group Organ. Stud. 109. Anderson LE, Balzer WK. The effects of timing
1984; 9: 509–29. of leaders opinions on problem solving groups:
93. Nemiroff RM, Pasmore WA, Ford DL. The A field experiment. Group and Organization
effects of the two normative structural inter- Studies 1991; 16: 86–101.
ventions on established and ad hoc groups: 110. Wortman PM, Vinokur A, Sechrest L. Do con-
implications for improving decision making sensus conferences work? A process evaluation
effectiveness. Decis. Sci. 1976; 7: 841–55. of the NIH consensus development program.
94. Burleson BR, Levin BJ, Samter W. Decision- J. Health Politics, Policy and Law 1988; 13:
making procedure and decision quality. Hum. 469–98.
Commun. Res. 1984; 10: 557–74. 111. George JF, Dennis AR, Nunamaker JF. An
95. Van de Ven AH, Delbecq AL. The effective- experimental investigation of facilitation in an
ness of nominal, Delphi, and interacting group EMS decision room. Group Decision and Nego-
decision making processes. Acad. Manage. J.
tiation 1992; 1: 57–70.
1974; 17: 605–21.
112. Davis JH. Group decision and social interaction:
96. White SE, Dittrich JE, Lang JR. The effects of
A theory of social decision schemes. Psychol.
group decision-making process and problem-
Rev. 1973; 80: 97–125.
situation complexity on implementation attempts.
Admin. Sci. Q. 1980; 25: 428–39. 113. Davis JH. Group decision and procedural jus-
97. Jarboe SC. A comparison of input-output, tice. In: Fishbein M (ed.). Progress in Social
process-output, and input-process-output models Psychology, Hillsdale, NJ: Erlbaum, 1980.
of small group problem-solving. Communication 114. Davis JH, Stasson MF, Parks CD, Hulbert L,
Monographs 1988; 55: 121–42. Kameda T, Zimmerman SK, Ono K. Quantita-
98. Miner FC. A comparative analysis of three tive decisions by groups and individuals: voting
diverse group-making approaches. Acad. Man- procedures and monetary awards by mock civil
age. J. 1979; 22: 81–93. juries. J. Exp. Soc. Psychol. 1993; 29: 326–46.
99. Rohrbaugh J. Improving the quality of group 115. Arrow KJ. Social Choice and Individual Values.
judgement: social judgement analysis and the New Haven: Yale University Press, 1963.
Delphi technique. Organ. Behav. Hum. Perform- 116. Coleman A. Game Theory and Experimental
ance 1979; 24: 73–92. Games. Oxford: Pergamon Press, 1982.
448 The Advanced Handbook of Methods in Evidence Based Healthcare
117. Flores BE, White EM. Subjective vs objective variations from an expert panel process. Am. J.
combining of forecasts: An experiment. J. Fore- Public Health 1990; 80: 1246–52.
casting 1989; 8: 331–41. 121. Shrout PE. Analyzing consensus in personality
118. Rowe G. Perspectives on expertise in the aggre- judgments: a variance components approach. J.
gation of judgments. In: Wright G, Bolger F. Personality 1993; 61: 769–88.
(eds). Expertise and Decision Support. New 122. Kozlowski SWJ, Hattrup K. A disagreement
York: Plenum Press, 1992. about within-group agreement: disentangling
119. Huber GP, Delbecq A. Guidelines for combin- issues of consistency versus consensus. J. Appl.
ing the judgments of individual memebers in Psychol. 1992; 77: 161–7.
decision conferences. Academy of Management 123. Murphy MK, Black NA, Lamping DL, McKee
Journal 1972; 15: 161–84. CM, Sanderson CFB, Askham J, Marteau T.
120. Naylor CD, Basinski A, Baigrie RS, Goldman Consensus development methods, and their use
BS, Lomas J. Placing patients in the queue for in clinical guideline development. Health Tech-
coronary revascularization: evidence for practice nol. Assess. 1998; 2(3): 1–88.
Part VI
IDENTIFYING AND
FILLING GAPS IN THE
EVIDENCE
I N T R O D U C T I O N by A N D R E W S T E V E N S
The number of health technologies, including most expensive new technologies, affecting the
drugs, devices, procedures and settings, is seem- largest numbers, with the greatest degree of
ingly endless. There are some 10 000 distin- uncertainty as to their effectiveness and with the
guishable diseases in the International Classifi- shortest ‘window of opportunity’ with which to
cation of Diseases (ICD-10) coding, each of assess them.
which is accompanied by at least 10 significant Healthcare systems therefore need mechan-
interventions, implying there are some 100 000 isms to:
existing interventions. To this an ever-expanding
annual new intake needs to be added. However, 1 identify new and emerging health technol-
the capacity to identify new technologies as they ogies before they are allowed to diffuse
appear, still less to prioritise them for evalu- widely into the system. Such a system might
ation, and then evaluate them seems uncomfort- also build up an inventory of existing tech-
ably finite. Many healthcare systems now recog- nologies and label them by three of these
nise that this gap needs careful managing if criteria, i.e. numbers, unit cost and evidence
base,
they are to avoid the distortion of priorities for
2 select the most important topics for assess-
healthcare spending. The risks are both dis-
ment, i.e. both whittling the possibilities
tortion due to ‘free-for-all’ in an uninformed
down to a manageable number (relative to
healthcare market, and of distortion due to inap- research capacity wherever it operates), and
propriate selection in the control of healthcare ensure that expensive research, itself, gives
technologies’ availability. In the former circum- value for money,
stance the most heavily promoted new technol- 3 respond at the appropriate time and with
ogies (often, but not always, pharmaceuticals) suitable research methods, and
usurp health budgets whether they are cost- 4 operate a system for knowledge dissemina-
effective or not, and indeed sometimes whether tion and implementation.
they are effective or not. In the latter circum-
stance, a selection of some technologies may This is not a simple linear progression through
be evaluated and their diffusion managed, but four stages, but an iterative process involving
with no guarantee that this selection reflected a stages 2, 3 and 4. This section focuses on the
sensible choice. A sensible choice might be one first three stages. The fourth stage including
in which priority in evaluation was given to the issues of dissemination, including incentives,
450 The Advanced Handbook of Methods in Evidence Based Healthcare
education guidelines, feedback and peer pres- evaluations using a number of new criteria,
sure, is beyond the scope of this book, although including the handling of alternative outcomes,
a start is made in this area by Black et al.’s the handling of uncertainty, and the availability
review in Chapter 24 on consensus methods of data to feed the model. The authors propose a
in guidelines. Other useful papers include the new model which is being tested on four case
NHS Centre for Reviews paper on guidelines studies of publicly funded randomised control-
(1992), on dissemination (1999) and Oxman led trials.
et al. (1995).1–3 Even when new technologies have been iden-
As regards the first three stages, this section tified, priorities have been set and the case for a
of the book explores new territory. Robert and health technology assessment been shown to be
colleagues investigate how to scan for new tech- persuasive, the question remains as to when (as
nologies. They note that horizon scanning and well as how) the assessment be undertaken. The
early warning are necessary for the managed problem, identified by Mowatt and colleagues,
entry of new technologies into healthcare sys-
particularly concerns fast-changing technologies.
tems, such that they can be properly evaluated
and planned for prior to diffusion. The difficul- The argument against early evaluation, is that
ties confronting horizon scanning concern the until a technology has stabilised – for example
breadth of the horizon, the unpredictable pattern until surgeons undertaking a new operation have
of diffusion – many technologies do not follow a all passed a certain point on the learning curve –
simple linear diffusion curve – and the tendency the results of the evaluation will not fairly
for new technologies to evolve and be used for reflect its established use. On the other hand, if
different indications as they diffuse. Notwith- an evaluation is undertaken once it has stabilised,
standing these difficulties, a number of countries the potential for getting clinicians to cooperate
have set up early warning systems to accompany with a study will be that much more difficult.
their health technology assessment programmes. Their minds will be made up – equipoise will
The methods that such systems use are still in not be achievable. Mowatt, reviewing the dif-
evolution. Using a combination of literature fusion and development of a number of case
review, survey and case study, the authors studies – laporoscopic cholecystectomy, chorion
conclude on the need for multiple sources in villus sampling, telemedicine, breast cancer, and
horizon scanning – though particularly the use gene therapy for cystic fibrosis, favour alertness
of key pharmaceutical and medical journals and and early assessment, if necessary undertaken
expert networks. iteratively as the technology evolves.
Harper and colleagues consider the pre-
liminary economic evaluation of health tech-
nologies. The process of prioritising health
technology assessment projects has two compo- REFERENCES
nents. The first is to narrow the number of
possibilities, and the second is to undertake a
preliminary economic evaluation which might 1. NHS Centre for Reviews and Dissemination.
inform the value for money of further more Implementing Clinical Practice Guidelines. Effec-
expensive research. Harper identifies a number tive Health Care, 1992.
of criteria already in use to do the former, 2. NHS Centre for Reviews and Dissemination. Get-
including those used by the health technology ting evidence into practice. Effective Health Care
assessment programme in the United Kingdom 1999; 5(1).
and the Institute of Medicine in the United 3. Oxman A, Thomson M, Davis D, Hayes R. No
States. Their main focus, however, is to review magic bullets: a systematic review of 102 trials of
attempts to perform preliminary economic evalu- interventions to improve professional practice.
ation. It is possible to appraise models for such Can. Med. Assoc. J. 1995; 153 (10).
25
Identifying New Healthcare Technologies
judged by the extent to which it facilitates IFN-b could have been identified in April 1993,
timely research based evidence on new tech- or even as early as November 1981.
nologies. This chapter makes further specific The introduction of other types of healthcare
research recommendations. technologies which have had large implications
for the cost and effectiveness of care has also
been uncontrolled. Examples include laparo-
BACKGROUND scopic surgery in the early 1990s (which has
been termed the ‘biggest unaudited free-for-all
in the history of surgery’8).
Brief descriptions of the key concepts that are The rapid speed with which new healthcare
central to this area of study and examined in technologies can diffuse, their potential impact,
this chapter are given in the appendix (see and their increasing numbers mean that there
pp. 467–70). is an urgent need to develop and operate an
approach which singles out those health technol-
Context ogies which might have a significant impact on
the NHS in the near future.
Often new technologies (whether drugs, devices,
procedures or innovative ways of delivering
services) are introduced to a healthcare system Response to Date
in a haphazard and uncontrolled manner, caus-
ing unnecessary confusion or expense. Many Many of the technologies in regular use today
high-cost, high-profile technologies have dif- would have been hard to predict 25 years ago,9
fused rapidly and not always appropriately or in but there is widespread recognition that a long-
a controlled way.1 Their diffusion, evaluated or term perspective is useful in all aspects of
not, is disorganised and occurs at rates that national policy-making. Nevertheless, formal
depend on the strengths of various influences2,3 analysis of the future remains a low priority for
(for example, pharmaceutical company market- most national decision-makers,10 particularly
ing or clinical enthusiasm for a new surgical those in the health sector.11
technique4). This is not surprising given that the Other sectors such as electronics, aviation12
key features of the market for healthcare tech- and agriculture have undertaken extensive studies
nology include a lack of information, and the of technological innovation and diffusion;13 yet it
separation of technology provision from its is in healthcare where major shortcomings in
financial ramifications.5 managing technological change have been iden-
Concerns are both about effectiveness and the tified, possibly because of the unusual way in
impact of the costs of new technologies in a which health technologies diffuse (see below).
fixed-budget, publicly funded healthcare system Such healthcare responses as there have been to
such as the NHS of the United Kingdom. date can be classified according to the time span
Expensive medical technologies are becoming of their perspective and according to whether
more frequent relating to expensive capital they are concerned principally with deferment or
equipment (for example, a whole-body scanner), with encouragement (Figure 1).
highly skilled persons to operate them (for ex- An EWS intended to help set priorities for
ample, renal dialysis) or by their high usage (for HTA research lies in the upper-left quadrant in
example, certain diagnostic tests or drugs). In Figure 1. It intends to help control and rational-
addition, many new drugs are very expensive, in ise the adoption and diffusion of technologies
part at least because of their development cost. that are being promoted by the healthcare indus-
Around 100 new drugs (including major new try and professional opinion-leaders.14 Other
indications) are introduced each year with in- than for HTA, an EWS with a short-term per-
creasing policy and expenditure implications. spective may be used by others needing early
Beta interferon (IFN-b) for multiple sclerosis information on emerging technologies. Commis-
could cost £400 million per annum in the United sioners of healthcare are one such group, but
Kingdom alone, but at the time of its licensing there is often too little evidence for their pur-
many questions about its usefulness continued to poses on the categories of people for whom the
await a trial with generalisable findings with full new intervention will be cost-effective or, for
assessment of cost and benefit. Research based example, on when treatment should be stopped.
evidence on cost–effectiveness (not just safety Futures studies may also take a longer-term
and efficacy) is the only way to establish the perspective and comprise a more ‘cooperative’
appropriateness of uptake of any technology,6 approach with industry (bottom-left and bottom-
and earlier identification could have helped to right quadrants). For example, futurologists and
ensure this was available prior to the marketing researchers brought together by British Tele-
and introduction of IFN-b.7 The development of communications (BT) have tried to look into the
Identifying New Healthcare Technologies 453
future by combing the literature and talking to From the 1970s onwards, studies of bio-
leading practitioners. They produced a timetable medical innovation, and of the diffusion of
for major medical and scientific developments medical technology, have become more fre-
over the period 1998–2030.15 Such initiatives quent, and slowly a base of knowledge is emerg-
may often be part of national attempts at tech- ing,20 but the basic mechanisms underlying
nology forecasting. medical research and development remain largely
Most of these initiatives are very recent. The unknown.21 The evolution of a new biomedical
longest experience is in the Netherlands which technology was initially thought of as a series of
was one of the first countries to identify the technical events which is usually described as
potential benefits of HTA.16,17 Only a very few linear-sequential (known also as the ‘science-
countries,18,19 have established programmes for push’ model) (Figure 2).
identifying, and monitoring the diffusion of, However, in the 1980s the validity of the
new and emerging healthcare technologies. A linear-sequential model was questioned.22,23 Its
Dutch Steering Committee on Future Health basic limitation was its implication that inno-
Scenarios (STG) in the Netherlands18 has been vation is much more systematic than it really is,
an on-going, iterative process. This has con- whereas not only research but also the broader
tributed to the healthcare horizon scanning lit- environment (as expressed through market
erature, but along with the few other exercises forces) influences each stage of the development
has not been evaluated. There is, therefore, no process.6 Thus the development and uptake of
agreed or empirically proven method of identi- an innovation is unpredictable and cannot be
fying and predicting the likely future impact of described in terms of standard processes.24
new healthcare technologies. Some healthcare sector research has pointed
to the difficulties of identifying the point of
Challenges to Horizon Scanning and innovation.13 Stocking found that in 22 innova-
Early Warning tions she studied there was one central person,
often a doctor, who had the idea, developed it,
A major potential problem for horizon scanning and was central in its promotion.25 This would
is the unpredictable pattern of healthcare inno- suggest that knowledge of where such individ-
vation and diffusion. uals announce their achievements (for example,
technology will eventually be used.24 New of the irrational and haphazard introduction of
healthcare technologies often interact with other new healthcare technologies or to promote
technologies in unexpected ways. These inter- the adoption of beneficial and cost-effective
actions frequently cannot be anticipated as the technologies.
complementary technology may not yet have
been invented (for example, day surgery and
anaesthetics). Progress in five different bio-
medical research programmes (X-ray, tomo- METHODS USED
graphic techniques, instrumentation, mathemat-
ics and computers) were required in order to
We used four methods – literature review, tele-
develop computer tomography (CT) scanners26
phone survey, Delphi survey and case studies –
(Figure 3).
Stocking suggests that the important time to in order to achieve the stated aims of the pro-
assess a new technology is ‘at the point when ject which were: (i) to explore the most useful
opinion-leaders become interested in it’, which sources for identifying new healthcare technol-
is at a very early stage, often before clinical ogies; and (ii) to make recommendations to
trials data are available.25 Fineberg likens at- assist the establishment and operation of an
tempts at assessment ‘in this complex of evolu- EWS in the UK.
tion in science, disease, technology and society
to standing on shifting ground and aiming at a
moving target that is also changing shape’.27 Literature Review
Assessing the effectiveness and costs of an
innovative technology before its introduction A systematic review was made of the literature
and diffusion is therefore problematic.28,29 Early on health futures and forecasting in the UK and
assessments may not reflect potential capabil- from healthcare systems overseas to assess
ities or lower costs, and therefore will be of little information sources which have previously been
use either for the researchers or policy-makers. used to identify new healthcare technologies.
Later assessments risk being of little use for The databases searched were: Medline (1966–
decision-makers.30 97), HealthSTAR (1975–97) and ECRI’s HTA
Inevitably, earlier warning of a new technol- Information Service database of grey literature
ogy will not be able to provide as precise and in the US (1990–96).
detailed information as warnings which come
much nearer to the technology’s uptake into a
healthcare system. This trade-off between level Telephone Survey
of accuracy and earlier warning may be perti-
nent when selecting which information sources A semi-structured telephone survey was carried
are to be used as part of an EWS, as some out of coordinators of existing EWS in six
sources will be more likely to provide earlier countries (the Netherlands, Sweden, Denmark,
warning, whilst others will provide greater detail France, United Kingdom and Canada) to iden-
on an already emerging technology. tify which sources are currently being used, and
For prioritising a HTA research programme a to inform recommendations on the establish-
major challenge is to forecast which technologies ment and operation of an EWS in the UK as part
are likely to generate the most policy interest once
of a national HTA system. The questionnaire
they are widely used. Some sources will ensure
that no important technologies are missed, but focused on the aims, methods and level of
the appraisal of such sources will require more support of each of the EWS. The responses have
resources – much of which will be expended on been used to inform the critique of existing
technologies that come to nothing. They will EWS.
also need the development of criteria for select-
ing the technologies most likely to have a large
impact. Alternatively, other sources will require Delphi Survey
less appraisal in order to select the most import-
ant technologies, but run the risk of omitting A Delphi study, involving international experts,
from the research prioritisation exercise technol- to assess potential sources for identifying new
ogies which turn out to have large implications healthcare technologies, was performed. The
for the health service. content of the three stages of the Delphi study
Whilst the outputs of an EWS may be aimed are summarised in Figure 4.
at different audiences, the rationale for their We compiled a list of potential information
existence is the same: ‘managed entry’ either sources for identifying new healthcare technol-
to help prevent the undesirable consequences ogies from existing or previous EWS and other
456 The Advanced Handbook of Methods in Evidence Based Healthcare
1 Key medical journals (i.e. Lancet, British Medical Journal, New England Journal of Medicine)
2 Key pharmaceutical journals (i.e. PharmaProjects, Scrip, InPharma)
3 Key scientific journals (i.e. Science, Nature)
4 The financial press and press cuttings generally
5 Patent literature
6 Pharmaceutical companies
7 Private healthcare providers
8 Biotechnology companies
9 Medical engineering companies
10 Sentinel groups of expert health professionals
11 Patient special interest groups
12 Conference/meeting abstracts
13 The results of other countries’ horizon scanning exercises (e.g. the Netherlands)
similar initiatives and additional sources were criteria, which information sources would be the
suggested by participants (Table 1). Participants most useful for identifying various types of
were asked to suggest, using an agreed set of healthcare technology.
Identifying New Healthcare Technologies 457
* Summarised in Banta HD, Gelijns AC, Griffioen J, Graaff PJ. An inquiry concerning future health care
technology: methods and general results. Health Policy 1987; 8: 251–64; and Banta and Gelijns 1994.30 These
citations relate to the same project undertaken in the Netherlands during the mid-1980s.
The Netherlands ✓ ✓ ✓ ✓
Sweden ✓ ✓ ✓ ✓
United Kingdom ✓ ✓ ✓
Denmark ✓ ✓ ✓ ✓
France ✓ ✓ ✓ ✓
Canada ✓ ✓ ✓ ✓ ✓
Identifying New Healthcare Technologies 459
Timeliness 5 5
Time efficiency 4 4
Agreement with other sources 3 4
Objectiveness 3 4
Sensitivity of source 3 3
Depth of source 3 3
Specificity of source 3 3
Elucidation of likely knock-on effects 3 2
Explicitness of limitations 2 3
460 The Advanced Handbook of Methods in Evidence Based Healthcare
Case studies The three drug case studies all revealed that
pharmaceutical journals (for example, Scrip,
The case studies provide examples of healthcare Hospital Pharmacy, Biotechnology, Drug Ther-
technologies that have diffused without having apy, American Pharmacy) would have provided
been fully evaluated and/or without adequate early warning. In all of the three cases a relatively
consideration of their expenditure and policy large number of published studies appeared at
implications, and the consequent need for an key stages of the licensing process, such as at
EWS. They provide a means of validating the the time of submission of an application to the
results of the other three methods which were FDA or announcement of FDA approval, but
adopted to assess information sources for identi- such events occur relatively late in a drug’s
fying new healthcare technologies (Table 5). development. However, it is problematic to
However, the retrospective case study findings comment on the likely specificity of these sources
cannot be regarded as definitive because each on the basis of retrospective case studies.
source had it been used prospectively might have As evidenced by our case studies, journal
had an unmanageable yield (low specificity). articles in leading medical journals can provide
In the case of CT scanners, patents appeared early warning via:
to be an important source, but the very small
proportion of products for which patents are (a) reports of primary research (for example,
issued that actually reach the healthcare market a report of a Phase III clinical trial of
makes this source inefficient,30 and this source dornase alfa35); or
was not selected by the respondents to the Delphi (b) discursive pieces on the future of a par-
survey. Five other case studies (biosensors, ticular technology, such as those type III
LVADs, IFN-b, dornase alfa and donepezil) papers identified by our literature review
illustrated that monitoring the regulatory control (for example, The Lancet editorial on tele-
of drugs and devices in the United States via the medicine,36 the New England Journal of
FDA would be a useful source. Medicine editorial on dornase alfa,37 the
It is difficult to assess the potential role of Journal of the American Medical Asso-
pharmaceutical and biotechnology companies ciation paper on LVADs38 or series such
and medical engineering companies through as the ‘medical advances’ papers in the
retrospective case studies. However, in some British Medical Journal); or
cases (telemedicine) there has been a strong (c) news sections which may alert the reader to
technology push from manufacturers, although developments in areas of highly prevalent
or serious disease.
there has also been a degree of receptiveness on
the part of healthcare providers. For example, Specialist medical journals were helpful in a
British Telecommunications (BT) in the UK has number of the case-studies (providing particu-
developed the CARE project, initiating a series larly early warning in the cases of LVADs,
of telemedicine trials designed to gain an insight biosensors, IFN-b and donepezil).
into the potential impact of telehealth services. In four of the case studies (CT scanning,
In the cases of IFN-b and dornase alfa, pharma- telemedicine, dornase alfa and laparoscopic
ceutical companies were clearly involved in cholecystectomy) conferences would have been
promoting their products directly to clinicians a useful source as illustrated by Baker’s39 report
prior to licensing in the UK. Having these on the relative importance of sources of infor-
clinicians among an expert panel would have mation amongst early adopters of CT scanners
provided a few months early warning. in the US (Figure 5).
Identifying New Healthcare Technologies 461
Whilst it is problematic to assess retrospect- been identified by the literature review, tele-
ively the benefits of involving experts, six of the phone enquiry, Delphi survey and case studies
nine case studies (biosensors, LVAD technol- can be accessed directly via the Internet. For
ogy, telemedicine, dornase alfa, donepezil and example, many journals are now available on
laparoscopic surgery) were predicted by one or the Internet in some form, and conference
more of the previous studies identified in the reports on specific disease areas can be accessed
literature review (see page 457). These studies through various websites.
had used experts as their main source of infor- Popular media coverage may have helped to
mation. Another of the case studies (IFN-b) was highlight the likely importance of a number of
briefly referred to in the STG report and used as the case studies, such as LVADs and PICUs, but
an exemplar ‘new’ technology by Stevens et al. in it seems that such coverage may only appear
their postal survey.19 The final two case studies after the initial introduction of the technology.
(CT scanning and PICUs) had begun to diffuse
before any of the studies were carried out.
Modern day sources such as drug information Operating an Early Warning System
services could not always be assessed by the
retrospective case studies. They are likely to In 1985 the STG in the Netherlands started a
provide helpful corroboration, as indicated by project on future healthcare technologies, in
the monograph on dornase alfa produced by the collaboration with WHO Europe. The report
Drug Information Pharmacist’s Group. recommended a process for identifying new
Many of the information sources which have healthcare technologies as shown in Figure 6.18
The Netherlands: Both national HTA prioritisation 1–2 years 5–10 experts are used, via postal survey, 50–100 technologies are identified each
and health policy planning before telephone and meetings, to comment on year, 20 are considered in detail and 10
Health Council of the
adoption identified technologies. In addition, nine have reports written or are prioritised for
Netherlands, 1988
standing committees (10 members each) are R & D. The results are used to advise the
part of the routine operation of the EWS. Dutch government on the current level of
knowledge and also disseminated to
parliament, professional groups and the
media. The government uses the results to
inform regulatory measures, research
decisions and the introduction and
adjustment of legislation.
Sweden: Health policy planning ,5 years A scientific board of eight members and 80 technologies are identified each year, 40
before standing committees in certain fields are are considered in detail and brief reports of
‘ALERT’: SBU, 1997
adoption used (by telephone and meetings) both to 5–6 pages are written on 30 of these. The
identify technologies initially and to reports are published in a database available
comment on technologies identified by other on the Internet, and in the SBU newsletter.
sources. One or two experts are used to
advise on each specific technology.
United Kingdom: National HTA prioritisation and ,5 years Two or three experts used to check on each The EWS directly informs the UKs SGHT,
University of health policy planning. before technology identified by other sources (by and thus the NHS R & D strategy, of
Southampton adoption telephone) important new healthcare technologies
(1995–96); University
of Birmingham (1997–)
Denmark: DIHTA, National HTA prioritisation and 1–2 years Experts are used both to identify Results are fed in to the R & D programme,
1997 health policy planning before technologies initially and to check on the health service and industry.
adoption technologies identified by other sources.
France: ANAES, 1997 National HTA prioritisation and Adoption Use 5–8 experts, who are generally Reports are written on less than 10
health policy planning phase proposed by scientific societies, to check on technologies each year. The results are
each technology identified by other sources. disseminated to health policy makers
(French Ministry of Health and insurers)
and scientific societies, to inform coverage
decisions and planning.
Canada: Ongoing 1-year pilot project for 1–2 years Use postal surveys and telephone to access Identify over 1000 technologies each year,
CCOHTA, 1997 purpose of health policy planning before experts both to identify technologies consider 6–12 in detail and write reports on
463
464 The Advanced Handbook of Methods in Evidence Based Healthcare
National HTA organisations have tried to drugs on a regional and national basis, via the
establish an EWS (or at least to systematically Drug Information Service (DIS) and National
identify new healthcare technologies at a given Prescribing. Each of these acts as a supply
time), and Table 6 summarises the aims and source for the new National Horizon Scanning
involvement of experts in the six organisations Centre.
which are currently operating an EWS. Com- Each of the six national initiatives reported
monly, a small number of experts is used to that they are currently disseminating, or are
provide advice on each technology, but in some planning to disseminate, detailed information on
systems formal committee structures have been only a small number of technologies each year,
established as an integral part of the EWS (the usually 10 to 12. This dissemination is carried
Netherlands, Sweden). In the Netherlands, the out via a wide range of mechanisms and prod-
EWS incorporates the expertise of the 170 mem- ucts. This includes providing formal advice to
bers of the Health Council, as well as the nine government (the Netherlands) as well as more
standing advisory boards of the Council, each of informal dissemination to politicians (Sweden,
which have approximately 10 members. The France, the Netherlands) and national and pro-
current initiative in Sweden uses a scientific vincial health policy-makers (Canada, Sweden,
board (with members representing radiology, United Kingdom). Two of the initiatives (Canada,
nursing, physiotherapy, gene technology, oncol- Sweden) have Internet sites which provide up-
ogy, general surgery, general medicine, pharma- dated information on the technologies which
cology and pharmaco-epidemiology) and stand- they have identified and prioritised as being im-
ing committees in certain fields. In Canada, portant. Newsletters are used by three of the
experts – who are nominated by provincial gov- initiatives (Canada, Sweden, the Netherlands).
ernment advisers or otherwise identified through The case studies illustrate the benefit of inter-
Medical Research Council excellence awards
national collaboration. At the international level
and publications – are used via postal surveys
it would be beneficial to collaborate on defini-
and telephone interviews both to identify tech-
nologies initially and to comment on technol- tions, co-ordination, standardisation and com-
ogies identified by other sources (usually, three munication. In the longer term, there may be a
to five experts are consulted per technology). In role for a more formal mode of collaboration,
addition to the six national initiatives described perhaps based within the European Community
below, and although the US does not operate a or through INAHTA (International Network of
national HTA system as such, there are – or Agencies for Health Technology Assessment).
have been – a number of projects undertaken in
the US which are similar in scope to an EWS
(although not necessarily for the explicit pur- RECOMMENDATIONS
poses of HTA). Initiatives similar to the EWS in
Europe and Canada are undertaken at both the
Federal level and within the private sector in the
Recommended Information Sources
United States by organisations with an interest
in the evaluation of healthcare technologies. One of the key assumptions in our approach has
The most striking aspect of all these initiatives been to assume that different types of tech-
is, with the exception of the Netherlands, how nologies will be identified through different,
recently they have been established. It remains although not necessarily mutually exclusive,
to be seen how well established some of the information sources. For instance, in the case of
latest initiatives will become. procedures that are not product-dependent (for
In the UK there are a number of organisations example, arterial operations) the STG relied
and initiatives that either explicitly or implicitly more heavily on expert opinion, informal docu-
have a role in providing early warning of health- mentation of scientific and technological devel-
care technologies within the UK.40 As well as opments, and professional meetings and publica-
the Forecasting Secretariat to the National Stand- tions, than on commercial product development
ing Group on Health Technology, which was databases. A combination of sources will be
established in 1995, there are various activities required in order to ensure that all types of
for clinical early warning, such as the Safety and technologies and all important technologies are
Efficacy Register of New Interventional Pro- identified. Using more than one source will pro-
cedures (SERNIP) and the CMP subcommittee vide corroboration, increase the likely accuracy
of the Government’s SMAC, which aim to allow of any predictions, and increase the amount of
time for the preparation of guidelines, or to useful information regarding a new technology.
act as a brake on unjustified expenditure. In There were some discrepancies between the
addition, there is a well-established network of information sources which were recommended
pharmacists which provides information on new for identifying new healthcare technologies by
Identifying New Healthcare Technologies 465
Primary
Patients ✕ ✕
FDA licensing ✓ ✓ ✓✓
Pharmaceutical and biotechnology companies ✕ ✓✓ ✓
Medical engineering companies ✕ ✓✓
Secondary
Pharmaceutical journals ✓ ✓ ✓✓ ✓✓
Medical journals ✓ ✓✓ ✓✓ ✓✓
Scientific journals ✓ ✓✓ ✕
Specialist medical journals ✓ ✓✓ ✓
Conferences ✓ ✓✓ ✓ ✓
Experts ✓✓ ✓✓ ✓✓ ✓✓
Patient interest groups ✕ ✕ ✓
Private healthcare providers ✕ ✓✓
Drug Information Services ✕ ✓ ✕ ✓
Internet ✕ ✓ ✕
Media ✓ ✓ ✕ ✓
Tertiary
Other countries’ EWS activities ✓ ✓ ✓✓
Key:
Literature review
✓✓ 5 used by all previous studies
✓ 5 used by at least one previous study
✕ 5 not used
Telephone enquiry
✓✓ 5 used by at least 4 of EWS
✓ 5 used by some (1–3) of EWS
Delphi study
✓✓ 5 consensus that this source is a minimum requirement for an EWS
✓ 5 no consensus but from comments received may be useful
✕ 5 not recommended
Case studies
✓✓ 5 in the opinion of the case study reviewer (G.R.) this was the best source for at least one of the case
studies
✓ 5 in the opinion of the case study reviewer (G.R.) this source may have been helpful for at least one of the
case studies
Blank cells indicate no evidence available as method was not used to assess specific source
previous initiatives (papers from the literature best method for doing so is currently either
review), the telephone enquiry of existing assumed or arbitrary.
national EWS, the Delphi study and the retro- In addition to the three sources above,
spective case studies (see Table 7). ‘specialist’ medical journals, FDA licensing
The following three information sources were applications, conferences and liaison with phar-
suggested from all four of our methods: key maceutical and biotechnology companies were
pharmaceutical journals; key medical journals; highlighted, with reservations, as being poten-
and experts. It is hardly surprising that experts tially useful, additional information sources.
seem such an important source, and the pertinent Therefore, for the purposes of identifying
question for an EWS is not whether to use new healthcare technologies we recommend the
experts but how to select and access them. The following approach, using wherever possible
means of selection is particularly crucial, but the resources which are available on the Internet:
466 The Advanced Handbook of Methods in Evidence Based Healthcare
(a) scanning of ‘specialist’ medical journals, with regional DIS, SERNIP, SMAC-CMP)
key medical journals, FDA licensing appli- with the aim of ensuring adequate coverage
cations, key pharmaceutical journals and of all types of technologies and providing
conference abstracts, and liaison with phar- sufficient early warning; and
maceutical and biotechnology companies, (d) that the EWS should be part of a national
to produce a database of potential tech- programme to allow HTA research to be
nologies; and commissioned or run in parallel alongside
(b) regular meetings and/or surveys of sentinel early clinical trials.
groups of expert health professionals in EWSs for identifying healthcare technologies
order to review and comment on the tech- should not be concerned with making exhaustive,
nologies identified by the other information long-term forecasts but with highlighting new
sources. and emerging high-impact technologies that are
It should be noted that some of the potential likely to materialise. Wherever possible, the
sources are changing (for example, HTA agencies required length of early warning of a new tech-
and patient special interest groups), and may nology is 3 years, although this may vary depend-
become capable of playing an increasing role; ing on the type of research that is required (more
they should, therefore, be kept under review. for a randomised controlled trial, less for a
review or modelling exercise) and the technol-
ogy concerned. Current initiatives are concerned
Recommended Methods for Operating mainly with relatively short time-horizons (i.e.
an EWS 1–2 years).
One of the key elements in an EWS is a
The results of the literature review and the system for contacting and eliciting opinions
telephone enquiry suggest that the notion of from experts. Experts can be used both to
early warning has only recently emerged from ‘brainstorm’ new developments and to filter
reflections on the nature and utility of health information from other (documentary) sources.
technology assessments. These have emphasised Having identified new technologies, criteria
the importance of identifying a new technology for selecting those technologies which are in
as early as possible so that an appropriate evalu- most urgent need of evaluation are needed. The
ation can be initiated at a very early stage.b Of six national EWS used different criteria with
the existing national initiatives, the Health which to select which emerging technologies
Council of the Netherlands – which built on the should be highlighted but the following are
work of the STG in the mid 1980s – has had the commonly mentioned:41
most experience of an EWS. Often it may be (a) expected health impact (burden of disease);
possible to make use of existing related schemes (b) efficacy or predicted efficacy of the
or initiatives. However, where such opportun- technology;
ities do not exist specific initiatives are required; (c) type of development (innovativeness, inno-
even when they do exist, they may require vation phase, speed of diffusion);
supplementing. (d) economic consequences (investment cost,
The exact form and operation of the EWS total economic impact); and
(and the sensitivity and specificity, level of (e) policy relevance (regulatory decision,
detail and timeliness which will be required research agenda, controversial, ethical
from the chosen information sources) will ulti- concerns).
mately depend upon the healthcare system of
The value of an EWS to a HTA programme will
which it is a part and the purposes to which the
be determined to a very large extent by the
EWS are to be applied. Important aspects of
responsiveness of the programme to the outputs
the operation of an EWS are:
of the EWS. An EWS should not aim to provide
(a) continuity, so that the important monitor- an exhaustive list of all potential new healthcare
ing function of an EWS can be performed technologies, but select the most important tech-
on those technologies which have a long nologies and concentrate research planning on
development phase; these.
(b) only a relatively small core staffing is Central to the operation of an EWS is the
required as long as there is access to need for consistent methods of updating infor-
experts either through formal committee mation; technologies do not suddenly appear
structures or regular surveys and/or focus with little prior warning, but have been in devel-
meetings; opment for a long time before they begin to
(c) the need for collaboration with national diffuse. For example, the bases for the develop-
and international programmes that already ment of telemedicine and LVADs were first
exist (for example, in the UK collaboration conceived in the 1950s and 1960s, respectively.
Identifying New Healthcare Technologies 467
Often parallel developments in a number of method, and each has disadvantages: the litera-
other technological areas are required prior to ture review revealed very few relevant studies,
the full potential of the innovations being able to the EWS co-ordinators who participated in the
be realised (for example, CT scanners, tele- telephone enquiry are developing their systems
medicine, biosensors). This pattern of tech- by trial and error, the opinions of the partici-
nological development highlights the need for pants in the Delphi study are necessarily sub-
a ‘watchful waiting’ approach by an EWS; jective and open to bias, and the case studies are
monitoring technologies at different stages in historical exemplars only. However, the overall
their innovation, development and adoption. analysis of the results from the four methods (a
There may often be a progression from reports form of triangulation) provides a more robust
of discoveries in scientific journals to reports review of the important issues relating to EWS
of progress in developing the technologies in for identifying new healthcare technologies than
specialist journals and then onto key medical any single method alone.
journals as the technology is adopted. Given the lack of empirical evidence on the
practical value of an EWS it is important that
any new initiative be evaluated and monitored.
CONCLUSIONS Our findings would support further research
on:
The paucity of empirical evidence means that 1 Prospectively recording the information
one must be cautious in deciding which are the sources used to identify new technologies in
most useful sources of information for identify- order that their accuracy can be assessed at a
ing new healthcare technologies and the best later date when the value of the output from
methods for operating an EWS. However, EWS the EWS is known.
are being established simultaneously in a num- 2 Undertaking detailed prospective case studies
ber of countries (often by HTA agencies), and of new healthcare technologies to map the
intuitively they would seem to offer obvious diffusion processes of technologies and to
benefits. assess information sources for identifying
The relative usefulness of the many available them prior to their introduction into health
sources of information about new healthcare services.
technologies depends on the particular types of 3 Estimating the likely ‘payback’ from pro-
technology under consideration. Additionally, viding early warning of new healthcare
each source has its advantages and disadvantages, technologies.
and some provide earlier (and often, as a con- 4 Determining how much early warning is
sequence, less certain) warning of new tech- required for: (i) strategic policy decision
nologies than others. Each will also provide making (e.g. HTA research prioritisation);
information about different aspects of a technol- and (ii) day-to-day operational management
ogy and its likely impact, and some sources will decisions, which will include determin-
provide more detail than others. Thus the choice ing what is the most appropriate balance
of information sources which feed into an EWS between length of early warning and the
may be influenced by the choice between: level of certainty as to the likelihood of
(a) earlier warning of a potential technology the importance of the new technology.
with little certainty of its likely impact in
terms of its precise application and timing
of introduction (examples include confer-
ence abstracts and, perhaps, pharmaceut- ACKNOWLEDGEMENTS
ical and biotechnology companies); and
(b) very clear and precise information of a We are particularly grateful to all the partici-
specific technology but relatively late warn- pants in the international Delphi study, the co-
ing (i.e. shortly before introduction of the ordinators of EWS who responded to our tele-
new healthcare technology) to the health phone enquiry, the respondents to our e-mail
system (examples include key medical survey of ISTAHC members, to Professor Tom
journals and FDA licensing). Walley (University of Liverpool), Professor
Our conclusions and recommendations are based Duncan Colin-Jones (Portsmouth) for providing
on the results of four separate methods which a retrospective analysis of the accuracy of earlier
approached the two study questions from differ- initiatives, to David Hands of the South and
ent perspectives; each of these provided some- West Drug Information Centre (Southampton),
what different findings, which emphasises the to Deborah Anthony at the Wessex Institute for
importance of a multi-faceted approach. We Health Research and Development for her
adopted this approach as there is no single best assistance with the case studies on biosensors
468 The Advanced Handbook of Methods in Evidence Based Healthcare
and telemedicine, to Graham Mowatt and col- health care. Oxford & Anglia Regional Health
leagues for assistance with the case studies on Authority, September 1994.
laparoscopic cholecystectomy and telemedicine, 11. Garrett MJ. A way through the maze. What
and to Dr Pamela Akerman for assistance with futurists do and how they do it. Futures 1993:
the paediatric intensive care case study. 254–74.
12. Chen K, Jarboe K, Wolfe J. Long-range scenario
construction for technology assessment. Techno-
logical Forecasting and Social Change 1981; 20:
NOTES 27–40.
13. Rogers EM. Diffusion of innovations, 4th edn.
New York: Free Press, Macmillan, 1995.
a. Personal communication, Professor D. Banta
14. Vallance P. Separate R&D budget is needed for
and Dr J. Spiby, April 1998. Banta noted that the
monitoring effects of new drugs. Br. Med. J.
predictions in the STG report, whilst being ‘globally’
1998; 316: 939.
accurate, would have ‘problems with timing and
15. Researchers try crystal ball gazing to predict
focus’, citing that the study did not pick up minimally
future. Br. Med. J. 1996; 313: 706.
invasive surgery, although endoscopes and micro- 16. Szczepura A, Kankaanpaa J (eds.). Assessment
surgery were mentioned. of health care technologies. Case studies, key
b. Blume S. ‘Early warning in the light of theories concepts and strategic issues. Cambridge: John
of technological change’, European Workshop: Scan- Wiley & Sons, 1996.
ning the horizon for emerging health technologies, 17. Borst-Eilers E. Assessing hospital technology in
Copenhagen, 12–13 September 1997. the Netherlands. Br. Med. J. 1993; 306: 226.
18. Scenario Commission on Future Health Care
Technology. Anticipating and assessing health
care technology. Volume 1: General Considera-
REFERENCES
tions and Policy Conclusions. The Netherlands:
Martinus Nijhoff Publishers, 1987.
1. Dent THS, Hawke S. Too soon to market. Br. 19. Stevens A, Robert G, Gabbay J. Identifying new
Med. J. 1997; 315: 1248–9. health care technologies in the United Kingdom.
2. Whitted GS. Medical technology diffusion and its Int. J. Technol. Assess. Health Care 1997; 13:
effects on the modern hospital. HCM Review 59–67.
1981; Spring: 45–54. 20. Bonair A, Perrson J. Innovation and diffusion
3. Rutten FFH, Reiser SJ (eds.). The Economics of of health care technologies. In: Szczepura A,
Medical Technology. Proceedings of an Inter- Kankaanpaa J (eds). Assessment of Health Care
national Conference on Economics of Medical Technologies. Case studies, key concepts and
Technology, 1988. strategic issues. Cambridge: John Wiley & Sons,
4. Freiman MP. The rate of adoption of new pro- 1996.
cedures among physicians. The impact of specialty 21. Battista RN. Innovation and diffusion of health-
and practice characteristics. Medical Care 1985; related technologies. A conceptual framework.
Int. J. Tech. Assess. Health Care 1989; 5:
23: 939–45.
227–48.
5. Abel Smith B. The marketplace for medical tech-
22. Gelijns A, Rosenberg N. The dynamics of techno-
nology. In: The Economics of Medical Technol-
logical change in medicine. Health Affairs 1994:
ogy. Proceedings of an International Conference
28–46.
on Economics of Medical Technology, 1988.
23. Smith R. The roots of innovation. Br. Med. J.
6. Peckham M. Towards research based health care. 1987; 295: 1335–8.
In: Newsom-Davis J, Weatherall D (eds). Health 24. Luiten AL. The birth and development of an
policy and technological innovation. London: The innovation: the case of magnetic resonance
Royal Society, 1994. imaging. In: Rutten FFH, Reiser SJ (eds). The
7. Smith L, McClenahan J. Management of the Economics of Medical Technology. Proceedings
introduction of Betaseron. A Developmental of an International Conference on Economics
Evaluation. London: Kings Fund, 1997. of Medical Technology. Berlin: Springer-Verlag,
8. Cuschieri A. Whither minimal access surgery: 1988.
tribulations and expectations. Am. J. Surg. 1995; 25. Stocking B (ed.) Expensive health technologies.
169: 9–19. Oxford: Oxford University Press, 1988.
9. Newsome-Davis J, Weatherall D (eds). Health 26. Battelle. Analysis of selected biomedical research
policy and technological innovation. London: The programs. Vol. 2. Columbus Ohio: Battelle Colum-
Royal Society, 1994. bus Laboratories, 1976.
10. Hadridge P, Hodgson T. The Hemingford 27. Fineberg HV. Effects of clinical evaluation on the
Scenarios: Alternative futures for health and diffusion of medical technology. In: Institute of
Identifying New Healthcare Technologies 469
Medicine. Assessing medical technologies. Wash- Faculty of Public Health Medicine Part II Exam-
ington DC: National Academy Press, 1985. ination, 1995.
28. Greer AL. Medical technology: assessment, 45. Gelijns AC, Thier SO. Medical technology
adoption and utilization. J. Med. Syst. 1981; 5: development: an introduction to the innovation-
129–45. evaluation nexus. In: Gelijns AC (ed.). Medical
29. Foote SB. Managing the Medical Arms Race. Inno- innovation at the crossroads. Vol. I: Modern
vation and Public Policy in the Medical Devices methods of clinical investigation. Washington
Industry. Berkeley: University of California Press, DC: National Academy Press, 1990.
1992. 46. Warner KE. The need for some innovative con-
30. Banta HD, Gelijns AC. The future and healthcare cepts of innovation: an examination of research
technology: implications of a system for early on the diffusion of innovations. Policy Sciences
identification. WHO Statist. Q. 1994; 47: 140–8. 1974; 5: 433–51.
31. Food & Drug Administration. Forecast of emerg-
ing technologies. US Dept of Health & Human
Services, June 1981.
32. Spiby J. Advances in medical technology over the APPENDIX
next 20 years. Community Medicine 1988; 10:
273–8. Definitions
33. Technology Foresight Programme, Health & Life
Sciences Panel. Notes on the Delphi survey. Com- Futures studies and futurology
panion paper B to the Health & Life Sciences Futures is an extremely wide field, and futures studies
Panel report. Office of Science & Technology, fulfil many, and quite different, purposes.11,42 Figure 7
London: HMSO, 1995. sets the scope of this chapter in the context of the
34. Spiby J. Future and emerging medical technology. discipline of futurology. This chapter is concerned
A study to investigate advances in medical tech- principally with new healthcare technologies in the
nology in the United Kingdom. Submission to short term, and making plausible predictions as to
Membership of the Faculty of Community which are likely to be important and introduced into a
Medicine, 1987. healthcare system:
35. Fuchs HJ, Borowitz DS, Christiansen DH et al.
Effect of aerosolized recombinant human Dnase
on exacerbations of respiratory symptoms and on New and emerging healthcare technologies
pulmonary function in patients with cystic fibrosis. Healthcare technology ‘encompasses all methods used
N. Engl. J. Med. 1994; 331: 637–42. by health professionals to promote health, prevent and
36. Telemedicine: fad or future? Lancet 1995; 345: treat disease, and improve rehabilitation and long-
73–4. term care. It includes the activities of the full range of
37. Davis PB. Evolution of therapy for cystic fibrosis. healthcare professionals, and the use of equipment,
N. Engl. J. Med. 1994; 331: 672–3. pharmaceutical and healthcare procedures gener-
38. Berger RL, McCormick JR, Stetz JD et al. Suc- ally’.43 New healthcare technologies are those which
cessful use of a paracoroperal left ventricular are only just about to be introduced, or have only
assist device in men. JAMA 1980; 243: 46–9. recently been introduced, to clinical practice44 (includ-
39. Baker SR. The diffusion of high technology ing new applications of existing technologies). Emerg-
medical innovation: the computed tomography ing technologies are earlier in the pre-adoption phase,
scanner example. Soc. Sci. Med. 1979; 13: for example pharmaceuticals in Phase II and III
155–62. trials.
40. Stevens A, Packer C, Robert G. Early warning of Thus, new and emerging technologies comprise
new health care technologies in the United King- those technologies in the applied research stage, about
dom. Int. J. Technol. Assess. Health Care 1998; the time of initial clinical testing, and those past the
14: 680–6. stage of clinical trials but not yet in widespread use.
41. Carlsson P, Jorgensen T (eds). European work- They may include technologies localised to a few
shop: scanning the horizon for emerging health centres, and as a general rule tend to be relatively
technologies. Copenhagen, September, 1997. unevaluated.
42. Bezold C. The future of health futures. Futures
1995; 27: 921–5.
Innovation
43. NHS Executive. Report of the NHS Health Tech-
nology Assessment Programme 1996. London: An innovation is an idea, practice or object that is
HMSO, June, 1996. perceived as new by an individual or other unit of
44. Rosen R. Introducing new medical technologies: adoption.13 Technological innovation in medicine
what role for purchasers? Membership of the covers the wide range of events that includes the
470 The Advanced Handbook of Methods in Evidence Based Healthcare
discovery or invention, development and dissemina- tion, budgetary planning) can then be performed if
tion of a new healthcare technology.45 needed.
in the future. Whilst health technology assess- is that once stability has been identified, these
ment (HTA) seeks to inform and influence this same proponents are no longer willing to con-
evolution to encourage the adoption of effective duct rigorous evaluation (such as through a
and safe forms of healthcare, it is only one, still randomised trial) because by then they are con-
relatively weak, factor influencing this pro- vinced, rightly or wrongly, of the value of the
cess. Much more often, changes occur for other new technology. This problem is exacerbated by
reasons, such as political, social, economic and the fact that technology change often occurs in
commercial pressures. The instability in health an unpredictable way, such that stability is only
technologies that results commonly causes diffi- recognisable in retrospect some significant time
culties for HTA, even undermining it. By the after it has actually occurred. The consequence
time a formal assessment has been completed, of waiting for stability is therefore that any
the technology may have changed so much that health technology assessment performed is com-
the results are considered to be irrelevant to monly not sufficiently rigorous to be really use-
current practice. These issues and the ways they ful, or is too late to inform diffusion. This then
may best be addressed are the focus of this is the dilemma for evaluating fast-changing
chapter. Given that all technology is subject to technologies: early assessment may lead to
change, albeit at markedly varying rates, there is a distorted evaluation which under-values (or
no simple definition of ‘fast-change’. We have over-values) the technology; on the other hand,
chosen to use the term to characterise a technol- late assessment may be methodologically so
ogy whose time-scale of development is so weak (and so late) that it has little if any impact
quick as to jeopardise the value of any HTA. on the technology’s adoption and use.
The most obvious reason for fast-change is It was against this backdrop that we con-
developments in the technology itself. Sometimes ducted a systematic review of research on these
such change is step-wise, reflecting occasional issues, drawing on detailed case studies of
fundamental alterations in the technology; at six specific health technologies to inform the
other times it is categorised by multiple small review.1 After describing the methods of our
changes giving the appearance of a smooth research in the next section, we go on to
change over time. A particular example of the summarise our findings, before drawing their
latter type of change is the learning associated implications together in the final section of the
with a new technology, such as a surgical tech- chapter. Whilst observations are made about
nique, that requires operator skill. The term HTA in general, our perspective was based on
‘learning curve’ is commonly used to describe the situation in the United Kingdom.
this. Whilst it is such changes in the technology
itself which are the focus of this chapter, it is
important to recognise that other types of ‘fast-
changes’ may occur, which are independent of
METHODS
the technology itself. For example, there may be
rapid changes in the types of healthcare prob- A series of literature searches was undertaken in
lem to which the technology is applied, or in an attempt to identify papers focusing on, firstly,
alternative approaches to healthcare and parallel general principles involved in the timing of
types of management, or in the costs of the HTAs, and secondly, reported assessments
technology. of six specific medical applications. Reported
Whatever the reason for fast-change, such assessments of laparoscopic cholecystectomy,
developments raise questions about the value of chorion villus sampling, teleradiology, tele-
any evaluation performed prior to a significant dermatology, diagnosis of genetic susceptibility
change. Whilst the impact of such changes will to breast cancer, and gene therapy for cystic
always be a matter of opinion, it ranges from fibrosis (Table 1) were analysed in order to try
such fundamental change that the performance to identify factors which influenced the timing
of the technology has obviously altered, through of those assessments.
to subtle or minor change such that the technol- The generic technologies and applications
ogy is essentially unchanged, from an assess- were selected to allow meaningful comparisons
ment point of view. between a range of technologies, of various ages
Change tends to be most rapid during the and at various stages of evaluation, so as to
early life of new technology. Taken at face value provide insights into the problems of how and
it may therefore seem most sensible to wait for when to assess new and fast-changing technol-
stability before any formal assessment is under- ogies. The technologies chosen allowed com-
taken. Indeed, this is a commonly adopted parison between those that were unregulated/
position. Proponents of new technologies often highly regulated, those with a low/high media
argue that to evaluate a new method before it is profile, those that were commercially driven/
stable is to put it to an unfair test. The difficulty publicly funded, those that were diagnostic/
Fast-Changing Health Technologies 473
therapeutic, and those that were invasive/non- timing. The Medline search (1966–96) retrieved
invasive. 1477 references in total, but in order to keep
Systematic searches for papers on the timing numbers at a manageable level was restricted to
of HTA and on each of the six medical applica- the period 1991–96, resulting in 605 references.
tions were made primarily on the electronic The EMBASE search from 1980–96 retrieved
databases Medline and EMBASE. Other data- 260 references. The titles (and abstracts where
bases searched included the Cochrane Library, available) of references retrieved by the searches
PROMT (national and international business were scanned, and those concerned with the
journals and newspapers), System for Informa- principles of HTA (176) obtained and analysed
tion on Grey Literature in Europe (SIGLE), and for any information on timing of assessment.
the Index of Scientific and Technical Proceed- Searches carried out on the other electronic
ings. Reference lists of important papers identi- databases were unsuccessful in retrieving refer-
fied were hand-searched. The following journals ences on timing of assessment.
were also hand-searched during 1996: Bandolier, The literature on the timing of HTA does not
British Medical Journal, Controlled Clinical appear to be extensive; of the 176 papers con-
Trials, Evidence Based Medicine, Health Service cerning HTA in general which were obtained
Journal, International Journal of Technology and analysed, 47 contained some mention of
Assessment in Health Care, Journal of Health timing. No papers were identified which focused
Services Research and Policy, The Lancet, and solely on the issue of timing; where timing was
Quality in Health Care. Key individuals were mentioned, comments were of a general rather
also interviewed. than specific nature and within the context of a
Literature-based indicators may provide use- broad-ranging discussion on HTA.
ful information on the activity around a tech- There was a broad consensus that new health
nology.2,3 Therefore, in addition to carrying out technologies should be evaluated before they
literature searches, the publication trends of the were allowed to diffuse, unevaluated, into clin-
six applications were also analysed. Using data ical practice.4 In addition to safety, effectiveness
from a Medline search, these trends were dis- and cost issues, ethical issues should also be
played graphically in an attempt to relate the addressed.5 To be regarded as successful, assess-
patterns generated to events in the development ments had to be done early enough to affect
and diffusion of the technologies, with the aim decision-making.6 There was a perception that
of identifying key points in time that could be the results of evaluations were generally avail-
used as indicators to trigger assessment. able too late, when decisions on adoption had
already been taken and the technology had dif-
fused into practice.7,8
WHEN TO ASSESS Several authors proposed the idea of an iden-
tifiable point in time at which an evaluation
General Principles should be undertaken, as opposed to being
merely ‘early’. The importance of the ‘critical
Initially, attempts were made to implement a point’ was such that if evaluations were under-
search strategy specific enough to retrieve papers taken too early they would be largely forgotten,
solely on the timing of HTA, but this proved whereas if they were undertaken too late, they
unproductive. Therefore, a more sensitive search would be of little value.9 This raised the issue of
strategy was devised and papers were sought on timing of evaluation of fast-changing technol-
HTA at a broad level, with the intention of ogies, in the context of determining when they
analysing these for any information included on were sufficiently stable for this to be done.
474 The Advanced Handbook of Methods in Evidence Based Healthcare
Stocking maintained that the first group to adopt about one of the procedures.17 Chalmers’ argu-
an idea would be the innovators involved in ment was that where surgical procedures had
designing the new procedure or technology.10 been tested by trials, such trials had been pre-
This in itself, however, does not lead to wide- ceded by many consecutive series of patients
spread adoption. Stocking argued that it was receiving the technique in which it might have
only when the opinion leaders of the relevant been slowly modified. He maintained that it was
group of clinicians began to take up the idea that unethical to wait until the technique had been
the majority of clinicians would follow. The modified before evaluating it. This was because
important period for initiating an evaluation was physicians would be asking certain patients to
the time when the opinion leaders were still give up their right to the standard accepted
considering the new technology.10 During this therapy in favour of a procedure not yet suffi-
phase there was clearly an emerging technology, ciently developed to justify comparing it with
but it had not yet diffused out of control. There- that standard therapy. Chalmers argued that it
fore the period when opinion leaders were con- would be more ethical to randomise from the
sidering the technology might also signal the beginning and explain to patients that they had
time when it had become sufficiently stable to an equal chance of receiving whichever was the
be evaluated. more beneficial therapy.17
There was also agreement that assessment Other authors, however, have drawn attention
should be an iterative process rather than a one- to the fact that during the initial stages the
off study.9,11,12 This raised the possibility of the practitioner’s skills and expertise with a pro-
existence of a number of critical points to guide cedure were still evolving, with the result that
first and subsequent evaluations. Policy-making the risks and benefits associated with it might
often depended on understanding the implica- change considerably.18 Gelijns stated that in
tions of a new technology at several points in its view of this ‘learning curve’ phenomenon, the
diffusion.13 A change in the nature of the condi- initial assessment of a new procedure would
tion, expanded professional knowledge, a shift probably need to involve non-experimental
in clinical practice, or publication of a new, studies.18 She argued that such early reporting of
conflicting assessment were some of the factors clinical experience might form the basis for the
that might trigger a decision to re-assess.14 design of subsequent RCTs or otherwise well-
Assessment could also be linked to the life cycle controlled trials to determine a procedure’s
of a technology, starting with prospective safety and efficacy.18
assessment of an emerging technology, assess-
ment of the technology in practice, then later
assessments as it changed, was used for new Diffusion and Development of Health
indications, or to decide whether it should be Technologies
replaced or abandoned.13,15
The issues outlined above underline the com- If the pattern of development and diffusion of
plex environment in which health technologies health technologies could be predicted, this
are developed and applied. One of the key information would be useful in helping to
requirements of rigorous assessment that is inform the decision of when to initiate an evalu-
influenced by rapid environmental change is that ation. However, other than for highly regulated
of ‘equipoise’, defined by Lilford and Jackson as technologies such as pharmaceuticals, this pro-
a situation in which there is no preference cess does not generally follow a simple linear
between the treatment options to be compared.16 progression. A more complex model in which
They argued that randomised controlled trials clear transition points can no longer be easily
(RCTs) were only ethical under conditions of identified has now largely replaced the earlier,
equipoise.16 One of the implications of equipoise conventional view that technologies had a
was that, practically and ethically, the period simple, consistent pattern of development and
during which rigorous RCTs could be under- diffusion. In addition, the same technology
taken might be brief since the length of time might also be modified in different ways by
during which a clinician had no preference different adopters, without rigorous evaluation
between the treatment options to be compared taking place. Health technology development is
was likely to be short. now generally seen as an iterative process, with
Some authors have argued for randomisation evolving scientific knowledge interacting with
of the first patient to receive a new treatment, as market demand to generate a particular pattern
opposed to delaying until randomisation could of innovation.18,19
be carried out with no further changes to the The introduction into healthcare of technol-
treatment in question.17 Chalmers maintained ogies already developed for other purposes, in
that physicians could never undertake a con- different sectors, further complicates this situ-
trolled trial if they were consciously enthusiastic ation. A high percentage of new medical devices
Fast-Changing Health Technologies 475
are not generated from biomedical research but include technological ‘push’ from instrumenta-
emerge through the transfer of technologies tion manufacturers, and the ‘pull’ of demand,
developed outwith the healthcare sector.4 One both from health professionals and patients.
example of this process is that of adapting At the time LC was first introduced there was
videoconferencing for ‘telemedicine’ applica- no requirement for surgeons to undergo special-
tions. In such cases, not only the early phases ised training before adopting it. This situation for
but also much of the later development of the new surgical procedures contrasted markedly
technology is unknown to organisations respons- with that governing the introduction of new
ible for clinical evaluation. Thus the course that drugs. New drugs had to conform to strict,
emerging, ‘fast-changing’ technologies take is centrally imposed regulations, which required
unpredictable; they are able to evolve independ- rigorous testing in animals according to strict
ently of any health service constraints. experimental designs, followed by carefully
The way in which healthcare is organised and controlled testing in humans with appropriate
delivered may also impact on the innovation protocols and follow-up observation. In contrast,
process. Traditionally, the developers of new no formal government regulatory system existed
health technologies looked upon physicians as for the development and evaluation of clinical
their primary customer base. More recently, procedures; their development had traditionally
however, other groups such as policy-makers, taken place in the context of clinical autonomy
hospital managers, patients, the general pub- and the relationship of trust between physician
lic and regulators have begun to be seen as and patient, with evaluation depending to a great
significant stakeholders in the demand for extent on professional self-regulation.18 This had
technology.20 led to a situation where the potential safety,
Following adoption, development does not efficacy and effectiveness of many surgical pro-
cedures had not been evaluated systematically
suddenly come to an end. Adoption of a technol-
during their development, allowing a ‘tidal
ogy may be the start of a long-term process of
wave’ of new healthcare technologies to diffuse
modification and redesign, based on the experi-
through healthcare systems before proper evalu-
ences of the technology’s new-found users. ation was undertaken.21
Health technologies therefore may continue to This was the case with LC, where diffusion
evolve for some time after their initial introduc- was not preceded by adequate evaluation of the
tion into practice. This raises the difficult ques- technique, and much of the reported assessments
tion of when exactly is the best time to evaluate consist of uncontrolled, descriptive studies.22
a ‘fast-changing’ health technology, bearing in Searches of Medline and EMBASE identified
mind that the eventual results of an evaluation 496 and 194 papers respectively. There has been
may be made irrelevant by developments in the no large-scale RCT of laparoscopic versus open
technology since the evaluation was initiated. cholecystectomy (OC). A few small RCTs have
been undertaken of LC versus OC.23–26 There
have been a similar number of RCTs of laparo-
The Case Studies scopic versus mini-cholecystectomy (MC).27–30
The observational studies and RCTs of LC
Systematic reviews of reported assessments of versus OC generally reported a longer operating
the following six medical applications were time, shorter hospital stay, less postoperative
undertaken in an attempt to clarify factors which pain, faster return to normal activity, and a much
had influenced the timing of those assessments, smaller scar. Whilst early RCTs of LC versus
and which might prove relevant to the wider MC found in favour of LC, later RCTs con-
debate on timing. The search strategy aimed to cluded that LC offered no clear advantage over
identify RCTs, major observational studies, sys- MC. A systematic review of the effectiveness
tematic reviews and also other overviews of the and safety of LC reported that the most serious
literature. Titles (and abstracts where available) complication was an increased likelihood of
of papers retrieved by the searches were scanned injury to the bile duct, and one of the review’s
and those that met the above criteria were recommendations was that surgeons should not
obtained and analysed. be encouraged to replace MC with LC.31
Evaluating LC posed several major problems.
Laparoscopic cholecystectomy There were difficulties associated with varying
levels of surgical skill. For example, a lack of
Laparoscopic cholecystectomy (LC) is a form of skill due to being at an early point on the
minimal access surgery (MAS) for removing the ‘learning curve’ might result in poor perform-
gallbladder. Since its introduction, the procedure ance by surgeons. As early as 1990 it was being
has diffused rapidly into routine clinical prac- argued that LC should be confined to specialised
tice. Factors influencing the rate of diffusion centres that would participate in prospective
476 The Advanced Handbook of Methods in Evidence Based Healthcare
studies.32 However this call was not heeded. curve’ period led to ethical objections to under-
Training issues were eventually addressed by a taking RCTs, despite a lack of reliable evidence
Working Group on the Implications of MAS for about rare or longer-term outcomes.
the NHS.33 Another problem was that MAS
technologies were potentially fast-changing, Chorion villus sampling
which was partly a function of the parallel
development of similar technology for non- Chorion villus sampling (CVS) and amnio-
medical applications. One of the features of the centesis are the most common methods of pre-
‘assessment’ part of HTA, as defined by the natal diagnosis of chromosomal abnormalities.
NHS Executive, was that a health technology CVS involves retrieving cells for chromosomal
should be sufficiently stable before it was analysis from the developing placenta, and was
evaluated.34 The value of an RCT would be originally performed at about 9 weeks gestation
diminished if, midway through the study, in the first trimester of pregnancy, with a
clinicians felt that the procedure should be diagnosis available around 2 weeks later. CVS
can be undertaken either transabdominally or
altered in some way to reflect new and important
transcervically. Amniocentesis involves trans-
developments.35 abdominal sampling of fetal cells from the
Media reports of the apparent benefits of LC amniotic fluid which surrounds the fetus and is
increased public awareness of the procedure and usually performed at about 16 weeks gestation
led to patients requesting it in preference to OC. in the second trimester of pregnancy, with a
It was argued that patient demand, fostered by diagnosis available around 3 weeks later.
press reports, was the most important factor in CVS was established in 1982 as a method for
facilitating the diffusion of minimal access ther- prenatal diagnosis,39 and has been used clinically
apies such as LC.36 This led to problems with since early 1984.40 Transabdominal CVS was
patient recruitment to RCTs. Given the apparent introduced by Smidt-Jensen et al. in 1984.41
benefits of the procedure, it became difficult to Searches of Medline and EMBASE identified
persuade patients to enter a clinical trial where 209 and 34 papers, respectively. Three major
they had an equal chance of receiving the con- RCTs have been undertaken comparing first-
ventional form of therapy.35 However it could be trimester CVS with second-trimester amnio-
argued that for this situation to have arisen in centesis, with the aim of measuring pregnancy
the first place, patients were not being fully outcome, antenatal complications and diagnostic
informed about the potential risks involved in accuracy. These included a Canadian trial from
undergoing the LC procedure. It was in such 1984–88,42 a Danish trial from 1985–90,43 and a
circumstances of patient-led demand, where European multi-centre trial from 1985–89.44 In a
MAS procedures were diffusing rapidly, that large US trial from 1985–86 randomisation was
RCTs were most needed to evaluate the benefits attempted but then abandoned because of patient
and risks involved. reluctance to be randomised to the amnio-
The long-term implications of LC also needed centesis arm.45 A prerequisite for clinicians tak-
to be taken into account when considering how ing part in the three major RCTs and in the large
to evaluate MAS technologies. RCTs would US study was that they should have undertaken
have had to be large in patient numbers and the procedure a number of times prior to the
based on a long period of follow-up in order to commencement of the trials.
detect important, but rare, adverse events. The The Canadian study was initiated in 1984 to
cost and logistical difficulties involved generally ensure that CVS was not introduced before its
risks and safety had been assessed.42 The reason
inhibited such trials from taking place.35
given for undertaking the Danish trial in 1985
As early as 1991 voices were being raised that was that, since CVS was a recent technical
the obvious benefits of LC rendered an RCT innovation, it was important to test the pro-
unethical,37 and by 1993 it was being argued that cedure critically before it was used routinely.43
the opportunity to conduct a large-scale con- The MRC European trial was initiated in the
trolled trial had passed.38 With hindsight, sys- same year because of the emergence of CVS as
tematic evaluation of LC should have started an alternative to amniocentesis; although CVS
from the earliest stages of its introduction. The allowed earlier diagnosis, its safety and diag-
reasons for this not being the case included that nostic accuracy were unknown.44
there was no consensus to withhold new tech- A Cochrane systematic review of CVS com-
niques until properly evaluated, neither was pared with amniocentesis for prenatal diagnosis
there centralised control to prevent diffusion of described the three major RCTs as being of
unevaluated techniques. Media reports fuelled generally good quality.46 The review concluded
initial demand for the new treatment, and that second-trimester amniocentesis was safer
favourable clinical audit during the ‘learning than CVS, and therefore the benefits of earlier
Fast-Changing Health Technologies 477
diagnosis by CVS had to be set against its of interest, mainly due to factors such as further
greater risks. If earlier diagnosis was required, technological advances, reduced costs, an empha-
then transabdominal CVS was seen as being sis by healthcare programmes on improving
preferable to the transcervical approach. efficiency, and a demand by rural patients and
The Canadian trial was remarkable in that physicians for equal access to high-quality
CVS was only available in Canada to women healthcare.
who participated in the study. Women who were The main driving force in telemedicine
eligible, but declined to take part, frequently appears to have been developments in com-
cited physician influence in helping them reach munications technology.54 The latest resurgence
a decision.47 Some women, who initially took in interest was at least partly due to technology
part and were allocated to the amniocentesis providers keen to expand their markets. By fund-
arm, found this unacceptable and subsequently ing telemedicine research in medical organisa-
travelled to the USA to obtain CVS.48 tions and academic institutions, telecommunica-
The possibility of a causal link between CVS tions companies have succeeded in stimulating
and limb defects was first reported in 1991,49 interest among the medical community and the
though none of the three major RCTs had identi- general public alike, thereby generating new
fied such a link. A number of other studies markets for their products.55
produced conflicting results, with some arguing Legal issues have also influenced the develop-
in favour of a link,50,51 whilst others argued ment of telemedicine, especially in the United
against.39,52,53 States, where most activity has taken place.
A number of factors appear to have influ- Were litigation to be successful, this might lead
enced when and how evaluations have been to reduced funding of projects and restrict future
carried out. Reports of the RCTs indicated that telemedicine development. US physicians are
there was a perceived need to ascertain the licensed by state, which raises questions over
safety, efficacy and diagnostic accuracy of CVS. their professional status if they undertake inter-
Clinical experience in using the technique was state remote diagnosis of a patient. The issue
regarded as an important factor in its safety and of professional status also extends to inter-
success. The ability to conduct successful RCTs national programmes, where the patient is located
was dependent on factors such as patient and in a different country from the telemedicine
physician preferences, and the extent to which consultant.
CVS was available outwith the confines of an Reported assessments of teleradiology and
RCT. teledermatology systems have been primarily
descriptive. Factors which appear to have influ-
enced the timing of these projects include the
Teleradiology and teledermatology technology-driven, as opposed to needs-based,
Telemedicine can be defined as the use of nature of telemedicine development in general.
transmitted images, voice and other data to Projects have mainly focused on technical feas-
permit consultation, education and integration in ibility, not clinical efficacy or cost–effectiveness,
medicine over a distance. Such systems can be and taken place during periods when commer-
used to deliver, for example, radiology and cial providers have made funding available. The
dermatology services to patients who are located short time scale and limited funding of many
in a different place from the specialist radiologist projects would have made it difficult to imple-
or dermatologist. ment RCTs. In addition, telemedicine projects
Searches of Medline and EMBASE for tele- have generally served sparsely populated areas,
radiology identified 158 and 100 papers respect- resulting in insufficient patient numbers to pro-
ively, and for teledermatology 10 and five papers vide statistically reliable data.
respectively. A search on the bibliography of
the Telemedicine Information Exchange on the Diagnosis of genetic susceptibility to
internet identified 231 papers on teleradiology,
and 21 on teledermatology. Although tele-
breast cancer
medicine has been around since the 1950s, the There are around 25 000 new cases of breast
early programmes failed to meet physician and cancer in the UK each year, with up to 10%
patient expectations and were not cost-effective. having an inherited basis. Mutations of the
When external funding was withdrawn the pro- breast cancer genes BRCA1 and BRCA2 may
jects came to an end and popular interest account for at least 80% of families with a
declined. A cycle of technological development history of breast cancer. Genetic tests can deter-
continued approximately every decade, resulting mine whether an individual is predisposed to
in renewed activity followed by a waning of develop a disease such as breast cancer.
interest when expectations were not realised. Searches of Medline and EMBASE identified
From around 1990 there has been a resurgence 81 and 47 papers, respectively. Evaluation of
478 The Advanced Handbook of Methods in Evidence Based Healthcare
genetic diagnosis for predisposition to breast therapy for cystic fibrosis.56–59 These were Phase
cancer is still at an early stage. In the United I studies involving a small number of patients.
States in 1994 the National Center for Human Most of the US studies have used adenovirus-
Genome Research (NCHGR), along with the mediated gene transfer, while a liposome-
National Cancer Institute, the National Institute mediated vector was the method of preference
of Nursing Research, and the National Institute in Europe. The studies generally reported a
of Mental Health, jointly awarded more than temporary improvement in function, followed
$2.5 million to research groups to answer some by a return to pre-treatment levels. By mid 1996
of the questions surrounding genetic testing. A there were five protocols in Europe, and 11 in
major European Union-funded project, integrat- the United States, for studies involving gene
ing several European centres and co-ordinated therapy for cystic fibrosis.
by the University of Aberdeen, UK, was in its Regulation of gene therapy is still evolving.
early stages. The aims of the project included The Medicines Control Agency (MCA) is the
documentation and evaluation of clinical ser- statutory body in the UK responsible for all
vices, management of familial breast cancer and aspects of medicinal products for human use.
preparation of guidelines. Assessment of protocols by this agency is cur-
Genetic testing has raised a number of ethical, rently a prerequisite of all gene therapy trials in
legal and social issues, and these concerns have the UK. In the United States, the Recombinant
been widely covered by the media. The useful- DNA Advisory Committee (RAC) exists to
ness of providing a test for BRCA1 and BRCA2 develop and oversee the application of guide-
for the general population remains controversial. lines for different types of genetic manipulation.
These two forms of genetic predisposition The US and European regulators are imposing a
account for less than 10% of all breast cancers. similar regime as that applied to pharmaceut-
In the event of a positive result, there is no clear icals, while considering the ethical issues raised
action to take which would effectively prevent by the invasive use of genetic technologies.
the disease from developing. In addition, some Gene therapy is a high-risk undertaking with
women who test positive for BRCA1 or BRCA2 no products on the market, and no guarantees of
will not go on to develop breast cancer. Also, a commercial success. A number of large pharma-
negative test result does not guarantee that ceutical companies have made investments in
breast cancer will not occur. Notification of test the technology, and many biotechnology com-
results, both positive and negative, can cause panies have also made major commitments.
psychological distress. In addition, individuals Most of the biotechnology companies are
who opt for genetic testing may inadvertently relatively small, the research is labour-intensive,
leave themselves open to insurance and employ- and costs are high. There is an expectation that
ment discrimination. Even while it was being gene therapy will eventually realise a commer-
argued that genetic testing should remain in the cial return, and there will be pressure from this
research setting until these issues were resolved, sector to translate research findings into market-
commercial organisations were in the process able products at the earliest opportunity.
of developing commercial diagnostic testing Genetic technology in medicine has been
services to identify mutations of BRCA1 and receiving wide coverage in the media in recent
BRCA2. years, with gene therapy one of the applications
attracting much interest. However, concern
has been expressed that over-optimistic media
Gene therapy for cystic fibrosis
reports on gene therapy might lead patients to
Cystic fibrosis (CF) is one of the most common make potentially life-threatening decisions based
lethal hereditary disorders, occurring once in on false expectations.60
approximately every 2500 births. The disease is Little information was provided in published
caused by a chromosomal abnormality, which reports about when trials were undertaken or
results in the creation of abnormally thick reasons for timing. One major factor that might
mucus. Individuals who have CF tend to suffer affect timing was that gene therapy development
from chronic lung infections, and lung damage was being controlled much like that of a drug,
is the usual cause of premature death, with few with a strict, centralised, regulatory process
patients surviving much beyond 30 years of age. through which clinical protocols had to pass
Gene therapy involves replacing the defective before receiving approval. Thus, although tech-
gene with a copy of a normal gene. nical capability might be determining the point
Searches of Medline and EMBASE identified at which trials were planned, the evolving regu-
248 and 355 references, respectively. There latory processes were deciding the actual dates
have been a small number of evaluations of gene at which they would be implemented.
Fast-Changing Health Technologies 479
Lessons From the Case Studies societal issues that can be presented by an
emerging (and changing) technology.
The six medical applications illustrated various
difficulties of evaluating health technologies that
were developing and changing, and the lessons PUBLICATION TRENDS ANALYSIS
to be learned. Laparoscopic cholecystectomy
was a classic example of the problems involved
in evaluating a fast-changing surgical technique. Our starting assumption was that the publication
LC demonstrated the importance of distinguish- trends associated with a technology might in
ing the process of using a technology from its some way be related to its rate of development
outcome, and of finding methods when evaluat- and diffusion. Therefore, we analysed the pub-
ing that took into account learning effects and lication trends for the six medical applications to
differing levels of underlying skill. Learning establish whether there might be identifiable
issues have since been addressed.61 ‘critical’ points in time that could be used as
Chorion villus sampling exemplified a tech- indicators to initiate an evaluation, particularly
nology where step-wise changes of uncertain in the context of a fast-changing technology
becoming sufficiently stable to allow assessment
significance were taking place, demonstrated
to take place.
by the shift away from the transcervical to the
We searched Medline from 1970–95, using
transabdominal approach. CVS also illustrated
exploded MeSH terms combined with textword
the challenge in health technology assessment of
terms for each of the six applications. The total
identifying rare but serious adverse effects and
numbers of references retrieved for each applica-
relating them to the technology. CVS demon- tion were then subdivided by year of publication
strated the value of RCTs that straddle changes and plotted on graphs (x-axis 5 year of publica-
in the technology. The RCTs allowed the impact tion; y-axis 5 number of references). Numbers
of the differences to be evaluated, but also given are for all references retrieved on each
highlighted the fact that these kinds of study application, not just reports of clinical trials.
design may not be the most appropriate for long- Figure 1 shows the publication trends for all
term safety and monitoring issues. six medical applications. Whilst the numbers of
Teleradiology and teledermatology were ex- references per year for LC and CVS appear to
amples of technologies developed originally for have peaked and are now declining, those for the
purposes other than healthcare. They demon- remaining four applications, although far fewer
strated constantly changing technical perform- in number, continue to increase. It can be seen
ance, such that image quality continued to that LC has generated a far higher number of
improve significantly during the course of references over a much shorter period of time
trials; telemedicine developments were primarily that CVS. This may indicate the level of interest
technology-driven applications trying to find a generated by the technology, or the amount of
place in healthcare. controversy surrounding it.
Diagnosis of genetic susceptibility to breast The decline in the annual numbers of refer-
cancer illustrated the development of tests that ences for LC and CVS suggests that these health
were thought to be inherently beneficial, before technologies have reached a reduced level of
a policy was established for their management activity in terms of research and evaluation. This
and for handling the wider ethical and societal may be because they are now perceived by
issues. The technology involved was also origin- healthcare professionals as having been suffi-
ally developed for a different purpose, that of ciently well evaluated to resolve any remaining
gene mapping. Its application to genetic risk uncertainty over their safety and effectiveness.
assessment was unpredictable in terms of bene- The graph suggests that the remaining four
fits and risks. There was, nevertheless, an in- applications have not yet reached this situation;
creasing public perception of and demand for their annual numbers of publications, although
this technology. by comparison low, continue to increase.
Gene therapy for cystic fibrosis showed a Figure 2 shows the annual number of refer-
technology at a very early stage of development ences, represented by columns, for LC for the
that eventually could have profound implica- period 1989–1995. The principal RCTs, rep-
tions in terms of benefits, risks, and the ethical resented by horizontal lines, are also included;
issues raised. Gene therapy demonstrated the these compare laparoscopic versus open chol-
challenge of evaluating emerging technologies ecystectomy (LC-OC) and laparoscopic versus
by choosing the right research methods at the mini-cholecystectomy (LC-MC). (In both Fig-
appropriate time, how regulation affected this ures 2 and 3 the horizontal lines of the RCTs
process, and also the challenging ethical and relate to the x-axis only.)
480 The Advanced Handbook of Methods in Evidence Based Healthcare
One reference appeared in 1989. This was ‘window’ may open from around the time that
followed by a dramatic increase in numbers over the first papers on a new procedure appear in the
the next few years, peaking at over 600 refer- clinical literature; it may close when coverage in
ences each year for 1993 and 1994. A decline the clinical literature and the wider media leads
in numbers began to appear in 1995. Over a to reduced uncertainty about the procedure’s
relatively short period of time a significant safety and effectiveness.
amount of publishing activity was generated However, the publication trends analysis cov-
on the subject of LC. The fact that trials were ered too few applications and was too crude to
being initiated during the period to 1992 demon- give a clear indication as to whether this might
strated the existence of continuing uncertainty be an approach that would yield useful indicator
over LC, and also that it was still feasible to points to trigger HTAs. In addition, the applica-
undertake RCTs (even if small-scale) during this tions other than LC and CVS had such a limited
time. literature (for example teledermatology) that the
Figure 3 shows the annual number of refer- absolute numbers of publications were too small
ences, represented by columns, for CVS for the to be meaningful. The retrospective nature of
period 1983–1995. The principal clinical trials, this type of analysis renders it less useful than if
represented by horizontal lines, are also included; it could be applied prospectively, so that the
these compare CVS with amniocentesis. The information was then available when most
first references to CVS appeared in 1983, with needed. It should also be borne in mind that
numbers on the whole steadily increasing until some databases are more ‘immediate’ than
reaching a plateau over the period 1989–1993, others in the length of time taken to incorporate
followed by a decline in 1994–1995. All the new bibliographic references. For example,
large trials in Figure 3 were initiated by the end Internet databases such as the Telemedicine
of 1985. Information Exchange may reflect a more ‘real-
Information from the publication trends for time’ picture of publication trends than conven-
LC (Figure 2) and CVS (Figure 3) does suggest tional databases such as Medline and EMBASE.
that there may be a narrow and relatively early They may therefore be more useful for monitor-
‘window of opportunity’ when it is possible ing technology development with a view to
to initiate an RCT of a new procedure. This timing evaluation.
482 The Advanced Handbook of Methods in Evidence Based Healthcare
organizations, and third-party payers in the 26. Berggren U, Gordh T, Grama D, Haglund U,
United States. Int. J. Technol. Assess. Health Care Rastad J, Arvidsson D. Laparoscopic versus open
1995; 11: 79–92. cholecystectomy: hospitalization, sick leave,
9. Banta HD, Andreasen PB. The political dimen- analgesia and trauma responses. Br. J. Surg. 1994;
sion in health care technology assessment pro- 81: 1362–5.
grams. Int. J. Technol. Assess. Health Care 1990; 27. Kunz R, Orth K, Vogel J, et al. [Laparoscopic
6: 115–23. cholecystectomy versus mini-lap-cholecystect-
10. Stocking B. Factors influencing the effectiveness omy. Results of a prospective, randomized study].
of mechanisms to control medical technology. In: Chirurg 1992; 63: 291–5.
Stocking B (ed.). Expensive health technologies. 28. McMahon AJ, Russell IT, Baxter JN, et al.
Oxford: OUP, 1988. Laparoscopic versus minilaparotomy cholecystect-
11. Franklin C. Basic concepts and fundamental omy: a randomised trial. Lancet 1994; 343:
issues in technology assessment. Intensive Care 135–8.
Med. 1993; 19: 117–21. 29. McGinn FP, Miles AJG, Uglow M, Ozmen M,
12. Advisory Council on Science and Technology. A Terzi C, Humby M. Randomized trial of laparo-
report on medical research and health. London: scopic cholecystectomy and minicholecystect-
HMSO, 1993. omy. Br. J. Surg. 1995; 82: 1374–7.
13. Banta HD, Thacker SB. The case for reassess- 30. Majeed AW, Troy G, Nicholl JP, et al. Ran-
ment of health care technology. Once is not domised, prospective, single-blind comparison of
enough. JAMA 1990; 264: 235–40. laparoscopic versus small-incision cholecystect-
14. Donaldson MS, Sox HC. Setting priorities for omy. Lancet 1996; 347: 989–94.
health technology assessment. Washington: 31. Downs SH, Black NA, Devlin HB, Royston
National Academy Press, 1992. CMS, Russell RCG. Systematic review of the
15. Steering Committee on Future Health Scenarios. effectiveness and safety of laparoscopic chol-
Anticipating and assessing health care technol- ecystectomy. Ann. R. Coll. Surg. Engl. 1996; 78:
ogy. Dordrecht: Martinus Nijhoff, 1987. 235–323.
16. Lilford RJ, Jackson J. Equipoise and the ethics of
32. Cuschieri A, Berci G, McSherry CK. Laparo-
randomization. J. R. Soc. Med. 1995; 88: 552–9.
scopic cholecystectomy. Am. J. Surg. 1990; 159:
17. Chalmers TC. Randomization of the first patient.
273.
Med. Clin. North Am. 1975; 59: 1035–8.
33. Cuschieri A. Minimal access surgery: implica-
18. Gelijns AC. Modern methods of clinical invest-
tions for the NHS. Report from a Working Group
igation. Washington: National Academy Press,
chaired by Professor Alfred Cuschieri. Edin-
1990.
burgh: HMSO, 1994.
19. Deber RB. Translating technology assessment
34. NHS Executive. Report of the NHS Health Tech-
into policy. Conceptual issues and tough choices.
Int. J. Technol. Assess. Health Care 1992; 8: nology Assessment Programme. London: NHS
131–7. Executive, 1996.
20. Bower DJ. Innovation in health care: develop- 35. Sculpher M. A snip at the price? A review of the
ments in NHS trusts. J. Manage. Med. 1994; 8: economics of minimal access surgery. Uxbridge:
54–61. Brunel University, 1993.
21. Sheldon TA, Faulkner A. Vetting new technol- 36. Banta HD. Minimally invasive therapy in five
ogies: those whose efficacy and safety have not European countries: diffusion, effectiveness and
been established will now be registered and cost-effectiveness. Health Policy 1993; 23:
evaluated. Br. Med. J. 1996; 313: 508. 1–178.
22. Pearson VAH. Minimal access surgery: a review. 37. Neugebauer E, Troidl H, Spangenberger W,
Bristol: Health Care Evaluation Unit, Department Dietrich A, Lefering R. Conventional versus
of Epidemiology and Public Health Medicine, laparoscopic cholecystectomy and the randomised
University of Bristol, 1994. controlled trial. Br. J. Surg. 1991; 78: 150–4.
23. Barkun JS, Barkun AN, Sampalis JS, et al. 38. Macintyre IMC, Wilson RG. Laparoscopic chol-
Randomised controlled trial of laparoscopic versus ecystectomy. Br. J. Surg. 1993; 80: 552–9.
mini cholecystectomy. The McGill Gallstone 39. Froster UG, Jackson L. Limb defects and chorionic
Treatment Group. Lancet 1992; 340: 1116–19. villus sampling: results from an international
24. Trondsen E, Reiertsen O, Andersen OK, Kjaers- registry, 1992–94. Lancet 1996; 347: 489–94.
gaard P. Laparoscopic and open cholecystectomy. 40. Jackson LG, Zachary JM, Fowler SE, et al. A
A prospective, randomized study. Eur. J. Surg. randomized comparison of transcervical and
1993; 159: 217–21. transabdominal chorionic-villus sampling. The
25. Jan YY, Chen MF. [Laparoscopic versus U.S. National Institute of Child Health and
open cholecystectomy: a prospective randomized Human Development Chorionic-Villus Sampling
study]. J. Formos. Med. Assoc. 1993; 92(Suppl. 4): and Amniocentesis Study Group. N. Engl. J. Med.
S243–9. 1992; 327: 594–8.
484 The Advanced Handbook of Methods in Evidence Based Healthcare
SUMMARY INTRODUCTION
This chapter presents a review of the existing This chapter appraises models for prioritising
methodologies for prioritising health technology health technology assessment (HTA) projects,
assessment projects. It sets out criteria for and assesses criteria currently used to prioritise
appraising models based on the EUR-ASSESS proposed assessments. Conclusions are drawn
project guidelines and the priorities of the UK from the review and used to recommend a
National Health Service Research and Develop- model to assess the expected costs and bene-
ment Health Technology Assessment (HTA) fits from HTA projects that can be used to
Programme. Most HTA organisations currently prioritise research or appraise a specific research
employ criteria based systems for prioritisa- proposal.
tion which do not include the specific cost– Any organisation funding HTA is faced with
effectiveness calculations which are the main the problem of prioritisation. The continuing
subject of this review. The literature reviewed development of healthcare technologies has
here includes models that allow an a priori increased the supply for potential assessment,
quantitative assessment of likely payback to and the move towards evidence based healthcare
HTA. Seven major models were selected as has increased the demand for assessments in
appropriate, and each was appraised against 13 terms of effectiveness and cost–effectiveness to
criteria, including whether the model explicitly inform and influence practice. However, just as
considered alternative outcomes, the counter- resources for providing new technologies are
factual and the costs and benefits of implementa- limited, so are resources for their assessment,
tion, and whether it was transparent and flexible and some form of prioritisation is necessary.
to adjust to different scales of research. Models Funding bodies frequently consider likely costs
that seem most appropriate for funding bodies to and benefits in general terms, but rarely attempt
use are identified. The chapter discusses issues of to estimate systematically the cost–effectiveness
how to deal with uncertainty, operationalising a of undertaking a research project. If such esti-
model, resource requirements, the timing of costs mates were available, HTA funding organisa-
and benefits, and the likely impact on practi- tions could use the calculations to aid funding
tioners’ behaviour. It takes the view that if the decisions.
ultimate purpose of applied HTA is to improve The increased demand for evidence to inform
healthcare, it is insufficient to see the outcome in practice and decision-making has resulted in the
terms of the reduction of uncertainty alone, but rapid growth of HTA programmes, and of
needs to include the implications of the resultant assessments proposed for funding. The funds
likely use of the technology. available limit the number of technologies that
can be evaluated: in 1997, approximately 1800
486 The Advanced Handbook of Methods in Evidence Based Healthcare
topics were proposed as candidates for assess- which the proposed research topics and projects
ment under the UK National Health Service can be prioritised.
R&D HTA programme; only 2% secured These criteria-based systems are inherently
research funds in that funding cycle.1 HTA sensitive to the interpretation of the criteria by
organisations need to assess the potential returns the panel members. Implicit within the decisions
to the proposed research to allocate limited concerning research proposals are judgements
research funds efficiently. The concept of cost– based upon potentially incomplete or inaccurate
effectiveness needs to be applied to the choice information and opinion. Due to the nature of
of trials, as well as to the use of the technologies these processes, there is no way of controlling
themselves. for the potential variation in interpretation. It is
HTA organisations such as the National for this reason that a number of researchers have
Health Service R&D HTA programme and the undertaken the task of developing a systematic,
Medical Research Council in the UK currently quantitative model to assess research proposals,
employ a criteria-based system for the prioritisa- to work alongside the conventional process of
tion of research projects.1,2 However, these sys- scientific peer review.
tems do not include an explicit calculation of the
cost–effectiveness of undertaking the research.
The panels assessing research proposals are METHODS
given a check-list of criteria (for example,
would the project lead to a reduction in uncer-
tainty, are the costs justified, are the objectives Search strategies were developed for use with
clearly stated) against which research proposals Medline, HealthSTAR and BIDS SSCI. These
are judged. search strategies were designed for maximum
The EUR-ASSESS project (an international retrieval, using both index terms and free text
group designed to stimulate and co-ordinate searches. Once titles had been selected, papers
developments in HTA in Europe and to improve were selected according to whether or not the
decision-making concerning adoption and use of article contained a proposed theoretical model
health technology) produced a set of guidelines for relating the costs (at whatever level) of
for the prioritisation of HTA projects.3 These undertaking a health services research project
guidelines suggest that with any form of projected benefits of health
services research. This selection was undertaken
(a) it should be clear to all involved in an HTA by two of the authors (G.H. and J.T.). In addi-
programme how priorities are identified; tion, more general papers concerning the pay-
(b) there should be agreement regarding the back to research, whilst not explicitly relating
approach, method and criteria for assess- the costs of the research to the payback, were
ment between those responsible for priority also considered for providing the context for the
setting; discussion below. The selected articles were
(c) the approach should reflect the goals of the collected and their models appraised using the
programme and the resources available; following criteria, adapted from earlier work by
(d) the aim of the priority setting process the authors.5
should be to ensure that the priorities
reflect the likely costs and benefits of the
possible assessments; Appraisal Criteria
(e) where possible, the method for priority
setting should allow possible assessments These were as follows:
to be rated in a systematic way against
1 Does the proposed model state its aims? In
explicit criteria;
particular, is it concerned with identifying
(f) information on priorities should be shared
technologies for assessment or for prioritis-
between those responsible for HTA pro-
ing individual trials, or both?
grammes; and
2 Does the proposed model consider the
(g) the processes and outcomes of priority set-
alternative outcomes of a trial, and the
ting should be evaluated to assess whether,
inherent uncertainty surrounding the out-
and if so how, it has affected the topics
comes in a probabilistic framework?
assessed.
3 Can the model be used within an iterative
These guidelines have been considered in devel- economic evaluation framework, using data
oping a set of criteria for assessing the literature from earlier stages and also informing later
below. The Institute of Medicine (IoM) in the stages?
USA has developed a more formalised method- 4 Does the process appear transparent and
ology which assigns values and weights to seven easily interpretable by a research funding
criteria in order to achieve a global score,4 with body and potential funding applicants?
Ex Ante Evaluation of Likely Payback (PATHS Study) 487
5 Does the model consider the counterfactual assessment. Medical Care 1990; 28: 703–23;
– what would happen if the trial were not to ISSN: 0025-7079.
be funded? ● Thompson MS. Decision-analytic determina-
6 Does the model consider the likely imple- tion of study size. The case of Electronic
mentation of the intervention following the Fetal Monitoring. Medical Decision Making
reporting of the trial’s conclusions? 1981; 1: 165–79.
7 Is there a data source for the model to be ● Townsend J, Buxton M. Cost effectiveness
operationalised? scenario analysis for a proposed trial of
8 Are other forms of payback to the research hormone replacement therapy. Health Policy
considered? If so, how? In particular, are 1997; 39: 181–94.
weights assigned to a range of potential
benefits to research? Each of these studies will be described briefly
9 How is the uncertainty and subjectivity and then systematically appraised using the cri-
inherent in the data collection and model- teria above.
ling process considered? In the paper by Thompson,7 the optimal sample
10 How efficient is the process likely to be? In size of a trial is estimated using decision analy-
particular, will the process itself consume sis. The aim of the model is to identify the
resources (researchers time, costs of data break-even sample size of a proposed trial, by
collection), and will the outcome be useful valuing benefit of reduced mortality in terms of
to research-funding bodies? estimated life earnings, with the cost of the trial
11 Is the process flexible to adjust to different containing a fixed component plus a variable
scales of research? component per subject. Thompson applied the
12 Can the model accommodate alternative methodology to a trial for electronic fetal
outcome measures of the trial, if necessary? monitoring of low-risk babies.
13 Can the model accommodate the analysis The Detsky8 model is based on the prior
of particular subgroups of patients? distribution of the risk reduction of the inter-
vention under consideration and the effect of a
trial (with a given power and significance level)
on the distribution of this risk reduction. By
LITERATURE REVIEW combining this effect with, for example, the
number of people currently dying per year when
treated with the traditional therapy, Detsky pro-
Seven major studies were selected as appro- duced a measure of effectiveness of a clinical
priate from the articles identified in the literature trial. By relating this to the cost of the trial,
search. The selection criterion was that the article Detsky was able to estimate a cost–effectiveness
contained a proposed model or a quantitative
ratio of the research. Detsky’s model assumed
example of assessing the benefits of a research
100% implementation following the conclusion
project in relation to the proposed costs. Meth-
of the trial.
ods which simply rated a proposed HTA against
Eddy9 developed the Technology Assessment
a set of criteria were excluded. The selected
Priority Setting-System (TAPSS) which models
articles were:
the effect of a range of factors on a single
● Claxton K, Posnett J. An economic approach research outcome measure. These factors in-
to clinical trial design and research priority- clude the number of potential candidates for the
setting. Health Economics 1996; 5: 513–24. intervention under evaluation, the probability
● Detsky AS. Using economic analysis to that the assessment will arrive at a particular
determine the resource consequences of result, the implementation following that result,
choices made in planning clinical trials. and the number likely to receive the technology.
Journal of Chronic Diseases 1985; 38: By adjusting these parameters, TAPSS can sys-
753–65. tematically account for any variation in the key
● Drummond MF, Davies LM, Ferris III FL. variables, and in the inevitable uncertainty that
Assessing the costs and benefits of Medical exists around the estimation of some of these
Research: the Diabetic Retinopathy Study. variables. Repeating the exercise for other pos-
Social Science and Medicine 1992; 34: sible outcomes and combining these probabil-
973–81. istically into a single figure can determine the
● Eddy DM. Selecting technologies for effects on more than one outcome. The TAPSS
assessment. International Journal of Tech- model forms the basis for the development of a
nology Assessment in Health Care 1989; 5: number of later models.
485–501. Phelps and Parente6 developed a model to
● Phelps CE, Parente ST. Priority setting in account for unexplained variation in medical
medical technology and medical practice expenditure in counties in New York State,
488
Does the proposed The model’s aim is to The procedure aims to The objective is to The model developed To account for To estimate the To develop a
model state its ensure that a trial demonstrate that ‘develop and test a was ‘a simple framework unexplained optimal sample methodology for
aims? In design is efficient. ‘economic principles can methodology for and quantitative method variation in size of a trial using the ex ante
particular, is it The objective is to be used to make explicit assessing the social costs for estimating the value medical decision analysis. evaluation of a
concerned with promote consistency the consequences of and benefits of research.’ of assessing different expenditure proposed
identifying in decision-making planning or management It is designed to consider technologies.’ It between New York technology
technologies for and priority setting decisions . . .’ The prospectively whether a considers a proposed counties and thus assessment,
assessment or for between research and model relates the cost research project should assessment and evaluates to estimate an comparing the
prioritising service provision. of the trial (but not the be funded. how it will affect health implicit value of costs and benefits
individual trials, treatment) to the sample and economic outcomes. assessing the following an
or both? size and the clinically The assessment can be technology. assessment’s
important difference compared between conclusion with the
identified in the trial. different proposed likely practice in
The aim is to consider assessments. the absence of the
the marginal productivity assessment.
of research funding,
when expanding or
contracting trial sizes.
Does the proposed The model compares To calculate the power The model is constructed Yes, the alternative No. Yes – net gain or Alternative
model consider the cost of a trial with of the trial, the model using a decision- conclusions of the loss of life. outcomes are
the alternative the expected net uses an expected level of theoretic approach, with assessment are reflected considered
outcomes of a benefits for different risk reduction, taking all the alternative outcomes in the range of Delta explicitly, with
trial? sample sizes. The possible risk reductions of the trial represented results selected for the positive, negative
alternative outcomes (2100% to 1100%) into by the probabilities assessment. For each and inconclusive
are incorporated in account, by weighting assigned to the different Delta result (the result of outcomes defined
the determination of them with the probability events and through an assessment that can in terms of varying
The Advanced Handbook of Methods in Evidence Based Healthcare
the expected benefits. that each of these sensitivity analysis potentially change the health gains (or
outcomes will occur. applied to these use of a technology) a losses) and
probabilities. probability needs to be weighted with the
estimated that reflects probabilities of the
the likelihood that the trial reaching those
assessment will reach outcomes.
that Delta result.
Can the model be The calculation of the Unlikely. Calculation of Yes. The cost- Potentially yes, as the No, it is a highly Yes. Yes.
used within an expected value of effectiveness depends on effectiveness analysis is detailed claculations specific macro type
iterative economic sample information the change in the risk in terms of the involved in the TAPSS model.
evaluation (EVSI) and the reduction following the effectiveness measure of model would be of
framework, using expected value of trial, which may be of the trial (vision years benefit in identifying
data from earlier perfect information little value to the gained). Such important sub-groups for
stages and also (EVPI) may be of researchers. However, information can inform specific analyses during
informing later little value in the later the marginal approach the full evaluation. later stages of the
stages? stages of economic would be of value assessment.
evaluation. should further funds be
requested by the
researchers
Does the process It is a highly The concept of the prior The decision-analytic Uncertain, as the model It is a broad field Probably, although Yes.
appear theoretical model, distribution of the risk approach clearly shows involves the interaction methodology and dependent on prior
transparent and using techniques and reduction is highly the processes of the between a number of the strong distribution.
easily concepts that may subjective, and the evaluation, the factors. The means of assumptions
interpretable by a mean little to implications of such may assumptions made and calculation may not be between
research funding potential research not be fully understood. what adjustments have particularly transparent. methodology and
body and applicants. been made. interpretation mean
potential funding it is not really
applicants? transparent.
Does the model Not explicitly. No. Yes. No. The methodology No. Yes.
consider the centres on local
counterfactual – practice moving
what would directly to current
happen if the trial average, but it is
were not to be not incremental.
Ex Ante Evaluation of Likely Payback (PATHS Study)
funded?
Does the model Not explicitly. No. Yes. Yes. No, it is implied Yes. Yes.
consider the likely to be perfect.
implementation of
the intervention
following the
reporting of the
trial’s
conclusions?
489
490
Table 1 Continued
Claxton and Drummond Phelps and Thompson Townsend and
Posnett (1996)11 Detsky (1985)8 et al. (1992)10 Eddy (1989)9 Parente (1990)6 (1981)7 Buxton (1997)12
Is there a data Theoretical model, Published literature. Published literature and Estimated by author. Published literature Expert opinion and Published
source for the with no example expert opinion. and estimates published literature.
model to be used. based on existing literature. Benefits
operationalised? US state data. calculated from
published estimates
of lifetime
earnings.
Are other forms of The benefits are based No. No. Health gains only. No, but model can be re- No. No. No.
payback to the on patient utilities, estimated using other
research converted into outcomes.
considered? If so, population benefits.
how? In Other forms of
particular, are payback are not
weights assigned explicitly considered.
to a range of
potential benefits
to research?
Can and how is Not explicitly, Yes, through sensitivity Yes, through sensitivity The model could be re- No. Could be Yes, through
the uncertainty although the EVPI analysis around the prior analysis in calculating calculated using different incorporated into adjusting the
and subjectivity can be re-calculated, distribution of the risk the cost-effectiveness values for the variables. the decision- proportion of
inherent in the with adjustments in reduction and the power ratios. theoretic people receiving
data collection the probabilistic curve. framework. treatment following
and modelling framework. a trial’s conclusion.
process
The Advanced Handbook of Methods in Evidence Based Healthcare
considered?
How efficient is Uncertain. Process is relatively Modelling exercise The data required to Interesting Dependent on prior If data or informed
the process likely simple, but subject to the would be relatively calculate results approach, but the distribution and opinion, is
to be? In interpretation of the prior straightforward. either requires categories are strong assumptions, available, then yes.
particular, will the distribution. estimation or resource- probably too broad but would be
process itself intensive collection, to be useful and usable.
consume resources for example, a effective in
(researchers time, trial. changing practice.
costs of data
collection), and
will the outcome
be used?
Is the process Uncertain – not Yes – the model was Yes, although the Unlikely, as data on No. Yes. Potentially more
flexible to adjust applied to data. applied retrospectively to application was to the small trials will be applicable to big
to different scales a number of trials NEI Diabetic unavailable to the extent trials, such as the
of research? ranging in costs from Retinopathy Study, the required by the TAPSS proposed HRT to
$78 000–$150 000. cost of which was $10.5 model. which the model is
million. For an applied in the
application to a smaller article.
trial, the sensitivity Implementation
analysis would need to following trial
be more precise. outcomes will
require precise
sensitivity analysis
for smaller trials.
Can the model It can accommodate a Difficult, as effectiveness Use of QALYSs is Yes, but the model No – potential In a limited way if Yes, application
accommodate number of patient is based on change in possible, which combine would need to be re- benefits are based they can be can use QALYs,
alternative based outcomes, as risk reduction, requiring a number of health calculated, and the on the assumed translated into combining a
outcome measures the effectiveness is one health related related outcomes. results combined with welfare losses from monetary terms to number of health
of the trial, if based around patients’ outcome. However, other forms of some (arbitrary) deviation from feed into the break- outcomes.
necessary? utilities. However, it benefits may be difficult weighting procedure. current average even equation. However, other
may be difficult to to use in the model. practice. forms of benefits
include other forms may be difficult to
Ex Ante Evaluation of Likely Payback (PATHS Study)
Can the model Not explicitly Not explicitly Yes, within the decision- Yes. No, it is relevant to Not explicitly Not explicitly
accommodate the considered, but would considered, but would be analysis framework. average county considered, but considered, but
analysis of be possible. possible. practice. would be possible. would be possible.
particular
subgroups of
patients?
491
492 The Advanced Handbook of Methods in Evidence Based Healthcare
which estimates an implied value of assessing not, the authors argue that expert opinion can be
the technology. The model assumes that any used to provide estimates of the gaps and to
variation in expenditure not explained by demo- provide the necessary data. Three trial outcomes
graphic or socio-economic factors is due to (positive, negative and inconclusive) were iden-
uncertainty about the outcome of the interven- tified, and changes in the long-term use of HRT
tion. The model operates under the assumptions were predicted for each outcome scenario,
that the ‘correct’ level of expenditure is approx- together with the net effects of these changes
imated by the adjusted mean level of expenditure in terms of discounted streams of costs and
and that the elasticity of demand for all health benefits.
technologies is 0.15. These factors are combined
to determine the welfare losses associated with
the unexplained variation in expenditure. These
welfare losses are then compared across Diag- APPRAISAL OF THE STUDIES
nosis Related Groups (DRGs) to identify the
highest payoff from potential research spending.
The analysis is focused at a technology rather The seven major studies identified from the
than a research proposal level, and is limited by literature search are appraised in Table 1.
the fact that a DRG may require more than one
particular technology.
Drummond et al.10 used a decision analytical Commentary
model to assess the economic returns from a
piece of applied research. The model was Certain issues arise from the review. Although a
applied retrospectively to the Diabetic Retino- number of papers consider the process by which
pathy Study, which was funded by the National research funds are allocated, very few actually
Eye Institute from 1972–1981. Adjustments to propose a systematic model to determine
the probabilities within the decision tree (the whether a research proposal is likely to be a
relative probabilities of whether a patient has cost-effective investment. Only seven such
diabetic retinopathy before and after the trial) models are identified above and are reviewed
and also to the extent of the population who according to the criteria listed.
might benefit from the intervention following
the trial, the benefits they might receive and Aims of the models, their methods and
the costs of implementation ensure that un-
certainty surrounding the modelling processes is levels of information
sufficiently covered. The criteria used to assess the models are based
Claxton and Posnett11 used a Bayesian con- on how practical they are in terms of effective-
cept of the value of information within their ness and efficiency. In general, the models had
model. Their model used the expected value of similar aims, namely to assess the ‘value’ of
perfect information (EVPI) and the expected undertaking a particular piece of health ser-
value of sample information (EVSI), expressed vices research or health technology assessment,
in monetary units, in two ‘hurdles’ designed to although the method used vary in each case. The
assess first the technologies for assessment and predominant methods used are variants of a
then the specific research proposals. Firstly, the
decision–analysis approach, with the costs and
EVPI is used to assess whether a reduction in
benefits for each trial outcome (and subsequent
the uncertainty surrounding a technology area
would be valuable. The criteria set by Claxton implementation pattern) combined probabilistic-
and Posnett is that the EVPI must exceed the ally. Such an approach has the advantages that
fixed costs of the research. Secondly, to ensure the benefits of the research are clearly defined in
the design of the trial is cost-effective when terms of the health outcomes of the trial’s effect
conducted at the optimal scale, the trial design on practice. The models that require a calcula-
must clear a second ‘hurdle’, namely that the tion of the expected value of information have
expected net benefits of the research (ENBR) is two major problems. The level of information
greater than zero, at the optimal sample size. available before the trial needs to be combined
Townsend and Buxton12 developed a decision into a prior distribution, for example of the
analysis model applied to an example of a perceived mortality risk reduction of undergoing
proposed trial of the long-term use of hormone a surgical procedure. Also, some models attempt
replacement therapy (HRT). The authors were to convert the increased value of information
able to draw upon detailed empirical data and into financial terms, and the problems of this
published estimates of the long-term benefit of have long been recognised. As Buxton and
HRT, but recognised that for other technologies, Hanney argue, such problems are ‘probably
such detail might not be available. Where it is no more readily soluble for research payback
Ex Ante Evaluation of Likely Payback (PATHS Study) 493
assessment than for the assessment of the inter- and institutional rigidities. There will also be
ventions themselves’.13 uncertainty surrounding the data used and the
models reviewed could be recalculated, with
Staged approach sensitivity analysis for this.
A major test of the efficiency of a model is Models such as those of Drummond et al. and
likely to be the ease with which the model could Townsend and Buxton, that use direct health out-
fit into an HTA funding organisation’s existing comes within a probabilistic, decision–analysis
processes of funding allocation. Sculpher et al. framework, and clear measures of the proportion
described the possibility of using economic of the population likely to benefit from the
evaluation iteratively as part of the HTA pro- intervention following the trial’s conclusion, are
cess,14 possibly with earlier preliminary analysis likely to be able to handle uncertainty more
being used to inform later decisions, for ex- systematically. With models that use a more
ample in identifying subgroups on which the full ‘theoretical’ approach, such as those of Claxton
evaluation might concentrate. Data collected for and Posnett and Phelps and Parente, it may be
analysis at an earlier stage might be used product- more difficult to handle the uncertainty that
ively at a later stage. For example, that collected surrounds variables such as the extent of imple-
to assess whether the technology requires mentation and the variation of the effect of size
assessment might be used to assess a specific in different subgroups of the population. In our
trial. The earlier stages of the iterative pro- context, the major criteria of the efficiency of
cess involve the prioritisation of technologies. the model is how likely it will be able to
Buxton and Hanney13 described a three-stage influence decisions made by a research-funding
approach to the prioritisation in terms of esti- organisation. This may raise problems for models
mating the payback from research. The first two using concepts that may appear abstract to
stages (theoretical potential payback and expec- potential funding applicants.
ted potential payback) consider topics or tech-
nologies, whilst the third stage (likely realisable The incremental effects of a trial:
payback) considers specific research proposals considering the counterfactual
designed to reduce specific elements of un-
certainty surrounding the technology that had To consider the incremental effects of the trial
been identified in the previous two stages. Our on subsequent practice, a model needs to include
review throws doubts on whether the more theor- what practice would be were the trial not
etical models have the requisite characteristics to funded. It may not be sufficient to assume that
satisfy this overall three-stage approach. current practice continues. The incremental
effects of the trial relating to changes in practice
Dealing with uncertainty need to be explicitly accounted for in terms of
It is the existence of uncertainty that prompts health gains. A trial that finds in favour of an
the initial demand for some form of HTA. The intervention that would have been introduced or
nature of the uncertainty is most appropriately continued in use anyway, but would not have
investigated at the technology stage, to see prevented its adoption had it found against the
whether the existence of this uncertainty is detri- intervention is clearly unnecessary. Few models
mental and if ameliorated, whether net benefits satisfy the necessary criteria of explicitly calcu-
might accrue. The extent of the uncertainty lating the incremental costs and benefits of a
is important, and the greater the number of trial.
‘unknowns’, the higher will be the likely cost of
the necessary specific research projects. The cost The likely impact on practitioners’
of a specific proposal will depend upon what it behaviour
is we are uncertain about. Uncertainty may
surround the cost of a particular intervention, the We take the view that, given the ultimate pur-
effectiveness of a particular intervention, or both pose of HTA is improved healthcare, the out-
the costs and the effectiveness. Identification of comes need to include the implications of the
the nature and extent of the uncertainty will resultant likely use of the technology, and so it
influence both the research proposal and the full would be insufficient for outcomes to be in
evaluation stages. There will be uncertainty also terms of the reduction of uncertainty alone. That
and some degree of subjectivity about imple- is, an assessment’s health-related benefits must
mentation – how the research results will influ- consider the extent to which an assessment’s
ence practice. This will depend on the effective- conclusions will affect health service practi-
ness of existing means of dissemination of tioners’ subsequent behaviour. A number of
information amongst practitioners, commission- models, including those of Detsky and Phelps
ers and policy-makers, and on the professional and Parente assume full implementation of the
494 The Advanced Handbook of Methods in Evidence Based Healthcare
intervention following the report of a trial’s uncertainty, the requisite data may not be avail-
findings. This is unlikely: organisational bar- able and a surrogate alternative such as using
riers, ineffective information dissemination and expert opinion may be necessary. Data require-
poor management of change are likely to operate ments of a model may be demanding or too
at varying levels, and prevent or delay the costly for a modelling exercise to be efficient.
immediate adoption of the more effective or Sheldon16 described modelling as ‘a way of
cost-effective technology or intervention in some representing the complexity of the real world in
organisations. A more realistic model would a more simple and comprehensible form’. How-
allow some means of assessing the extent of ever, if the model itself requires data collection
the likely adoption of the more cost-effective that is likely to consume significant resources –
technology, perhaps through the use of expert as might be the case with the TAPSS model –
opinion. the cost of the preliminary modelling exercise
might begin to approach the costs of under-
taking the assessment itself, in which case, the
Non-health benefits efficiency of the model and its usefulness to
A conflict arises between the multi-dimensional HTA funding organisations comes into question.
nature of payback from health services research A model that requires a wide range of vari-
and the need for a precise quantitative model. ables and considers more than one outcome may
Buxton and Hanney15 have identified five differ- appear attractive, but the data requirements
ent categories of potential benefits from health might be too great for it to run effectively,
services research: knowledge benefits; benefits especially where little research has been under-
to future research; political and administrative taken previously.
benefits; health benefits; and broader econ-
omic benefits. According to their objectives and The timing of the costs and benefits of the
sources of funding, HTA organisations may
place different emphases on these, and it is technology
possible to argue that a model designed to The timing of the costs and benefits relating to the
prioritise research projects should relate these HTA will affect its overall cost–effectiveness. It
potential benefits from research directly to is unlikely that the payback, and the costs of
the resource implications of undertaking the changes in implementation, will occur immedi-
research. As stated earlier, such a model would ately after the trial. All the models considered
require estimates of the relative importance of above are able to incorporate appropriate dis-
each form of benefit. For some HTA pro- counting; however, a number of the applications
grammes, it is clear that the prime motivation is on which the models were tested did not, but
to improve the cost–effectiveness of the health rather assumed immediate implementation.
system. Any other form of benefit or payback, Related to this is the potential lifespan of a
albeit important, would be considered second- technology. If a particular technology or pharma-
ary. Therefore, preliminary cost–effectiveness ceutical considered for assessment is expected to
studies using ‘equation-based models’ to evalu- be superseded shortly by a rival, which is at an
ate a research proposal have understandably earlier stage of development, then the ‘shelf-life’
concentrated on the health benefits, to reduce of the technology being considered for assess-
subjectivity in terms of weighting alternative ment might be relatively short, and a substantial
forms of payback from research. Other forms of investment in the assessment may not be effi-
benefits however should not be ignored, and cient. The benefits of the research over a given
whereas the majority of applications of the time period needs therefore to be adjusted for
models concentrate upon the health benefits, varying ‘shelf-life’. The value of an assessment
minor modifications of models such as TAPSS may also be reduced or negated if results of a
would allow other forms of payback to be similar or related assessment of the same tech-
assessed and non-health benefits could and nology is impending.
should be considered alongside the model.
The information needed to operationalise the Highly theoretical models using abstract meth-
above model would be: ods to determine the benefits of the research, are
(a) the likely development of policy or prac- unlikely to demonstrate this transparency.
tice, in the absence of the trial; The models based on a decision analysis
(b) the likely implementation of the experi- framework in which the cost–effectiveness of
mental intervention following the report- the research is expressed in terms of the health
ing of the trial results, given alternative intervention appeared to be the most transparent.
scenarios for the results: These models tend to rely on published data to
– the extent to which the technology or inform the process, and providing that a com-
service might be adopted; petent literature search is undertaken, such data
– how existing policy might affect the (costs and health-related benefits) should be
level of implementation; and accessible and translation of health benefits
– whether the technology or service into monetary form is not necessary. By the
would be introduced for all potential nature of the problem, there will be information
candidates, or whether it would be tar- required that will not be available, but which
geted at specific groups. might best be obtained by appropriate use of
(c) how other relevant services might be affec- expert opinion, for example through Delphi
ted following the adoption or change in groups.
practice of the experimental intervention; Use of a range of expert opinion would help
(d) whether the outcomes used within the trial to obviate bias, and sensitivity analysis is
are appropriate, whether they are likely to relatively straightforward in models based upon
influence future policy and/or practice and decision analysis theory, as the costs, benefits
whether there are any alternative outcome and probabilities of each arm can be easily
measures that might be more appropriate; adjusted. On the other hand, models in which
(e) whether the expert considers there are the benefits are difficult to disentangle may not
threshold levels for costs, effectiveness be easily adjusted by sensitivity analysis to
or cost–effectiveness, that might trigger account for uncertainty.
adoption of the experimental intervention; With these considerations in mind we devel-
and oped our model, aiming to include the various
(f) the probabilities that the overall outcome necessary characteristics identified, and attempt-
of the trial will be positive, negative or ing to strike a balance between the issues of
inconclusive. rigour, transparency and resource requirements.
Ideally to test this model it would be used
prospectively on a number of randomly selected
CONCLUSIONS trials, and the predicted outcomes compared
with a number of trials which had been funded
This chapter has considered a range of character- but not subjected to the analysis. However, such
istics that a quantitative model for the prioritisa- an evaluation would be lengthy and difficult to
tion of HTA projects needs to portray. There is undertake. As a realistic approximation there-
often a trade-off between the ‘completeness’ of fore we are currently testing and evaluating the
the model (and therefore the resource implica- model on four case studies of four publicly
tions of the modelling exercise) and the ease of funded randomised trials, identified by the
operationalising the model. A balance needs to authors in liaison with the UK NHS R&D HTA
be struck between these two factors. The models programme and the Medical Research Council.
presented here do not replace the existing pro- These case studies are running concurrently with
cesses of research prioritisation, but would be the preliminary modelling exercise and are
complementary to the process of peer review being evaluated prior to completion, and the
and criteria-based approaches that consider also evaluation compared with what professionals
benefits that are not applicable to direct quantifi- think the reported results imply for future use
cation. A model therefore needs to fit within the of the technology. On completion of the project
existing processes and be able to use earlier we hope to provide a practical guide to
analyses within an iterative framework. HTA organisations, particularly on quantitative
Transparency is an essential characteristic of assessment that can be used as an adjunct to
an acceptable model. Potential applicants need prioritise research or to appraise specific research
to know how their proposal is to be assessed. proposals.
absolute risks, 66 and HTA studies, 276 benefit and harm, 60, 62–4, 67–8,
administrative databases, 222 and observational studies, 287–8 68
Adverse Drug Reaction and randomised controlled Beta-Blocker Heart Attack Trial,
notification system, 144 trials, 283–7 104
aggregated datasets, 143 Bayes’ theorem, 277 between-cluster variation, see
analysis of variance, 185, 301–2 Bayesian philosophy, 30, 277 variation and cluster trials
analytical methods, developments computational issues, 280; see bias, 7, 39, 96, 105, 254, 371–90,
in, 273–4 also Markov Chain 416
aneurysmal subarachnoid Monte Carlo methods ascertainment, 39
haemorrhage (SAH) confidence profile method citation (reference), 375
case study, 152–5 cross-design synthesis, 288 co-intervention, 39
APACHE III, 163 decision and policy-making, 288 country, 373–4
applicability of results, 56–69 decision–theoretic approach, detection, 41
definition, 57 280 duplicate (multiple) publication,
methods for applying results, decision–theoretic Bayesian 375
61–8 approach, 285 grey literature, 373
area-based health interventions, default priors, 282 information, 141
methods for evaluating, language, 373–4
diagnostic tests, 279
295–313 minimization, 39, 41
eliciting opinion, 281
check-list, 309 outcome reporting, 7, 374–5
empirical approach, 280 performance, 41
cohort design, 301
design effect, 302 epidemiological studies, 287 publication, see publication bias
intervention studies, 298–300 exchangeability, 283 randomisation, 41
observational designs, 298 formal frequentist approach, 284 response, 253–4
paved designs, 300 frequentist properties, 285 selection, 7, 12, 39, 41
repeat cross-sectional design, historical controls, 287–8 binary outcomes, 393–4
301 informal Bayesian approach, biomedical innovation model, 453
restricted randomisation, 300 285 blinding, 7, 12, 20, 27, 30, 39, 41,
sample size requirements, 301 informal non-Bayesian 44, 79, 96–7, 110, 111,
stratified designs, 300–1 approach, 284 121, 141, 217, 410, 413,
study designs, 297–301 institutional comparisons, 288 417, 418, 420, 421, 423
within cluster and between interim analysis and predictions, double blinding, 411, 416, 420
cluster variation, 301–2 279, 284–5 blobbogram, 128, 129
see also organisation-based interpretation of probability, 278 bootstrapping, 331, 361
health interventions in meta-analysis, 288, 402–3 BUGS, 403
ascertainment bias, 39 multiplicity, 279–80, 288 cancer clinical trials, quality of
Asthma Quality of Life phase 1 trials, 287 life measures, 338–51
Questionnaire, 182 phase 2 trials, 287 cancer registries, 139, 143
attribution, 219 power, 284 CARE project, 460
autonomy, 29, 42 prior beliefs, see prior case-control studies, 126, 128,
Bayesian methods/statistics, 13, distribution 129, 130
19, 66, 273, 275–94, 399, proper approach, 280 case registers, 139
402 reference approach, 280 case record forms, 222
alternative hypotheses, 284 sensitivity analysis, 286 ceiling and floor effects, see
alternative trial designs, 286 Beck Depression Inventory, 183 patient-based outcome
crossover trials, 286 behavioural interventions, 20, 21; measures
equivalence trials, 286 see also ethical issues censored data, 223, 356
factorial designs, 287 benefit and cost, measurement of, censored survival times, 356
N-of-1 trials, 286–7 179–80 Census of Population, 138, 139
500 The Advanced Handbook of Methods in Evidence Based Healthcare
criterion validity, see validity Discounted Utility (DU) model, eligibility criteria, 102–3, 110
critical appraisal, 226, 360, 363, 232 eligible population, 57
368, 392 discounting, 231–46, 282 enthusiastic priors, 282; see also
Cronbach’s alpha, 126, 184–5 discrete choice modelling, 238–9; Bayesian methods
cross-design synthesis, 59, 288 see also closed-ended statistics, prior distribution
crossover trials, 286 questions epidemiological studies meta-
disease/condition-specific analysis of, 403; see also
instruments, 182 case-control studies, cohort
data collection, 45
disease registers, 144 studies
questionnaire surveys, 247–71
Doctors Independent Network epistemological uncertainty, 12, 18
resource use data, see cost data
and costing (DIN), 143 equipoise, 11, 13–19
data-dependent allocation, 286 double-blind trials, 44 collective, 8, 13–14
data extraction, 79, 80 double consent randomised design, concept, 28
data linkage between routine 107–8; see also Zelen constraint, 18–20
datasets, 139–40 randomisation corollaries, 18
Data Monitoring Committees, 45, drop-out, 42, 44 fundamental nature, 15
49 informative drop-out, 352–4, individual, 13–14
data transformation, 328–31 363 personal, 15, 20
databases, 26–7, 80, 118, 138, duplicate (multiple) publication, requirement, 11
297, 318, 322–6, 372, 473; 375; see also publication equity, assessing issues of using
see also under specific bias routine datasets, 142–4
databases equivalence trials, 286
datasets early warning systems (EWS), for error, 44, 184, 222
administrative, 222 new healthcare measurement, 222
aggregated, 143 technologies, 451–70, 458, random, 184, 254, 302
census-based, 137, 144 464, 465; see also horizon sampling, 302, 398, 399
enhanced administrative, 143 scanning systematic, 254, 413
health-status-only, 144 aim of, 470 type I, 275, 276
health-technology-only, 143–4 case studies, 457, 460–1 type II, 276
registry-type, 142–3 challenges to, 453–5
routine, 142–4 ethical issues, 11–21
operating an EWS, 461–4 adaptive designs, 12, 19, 21
see also routine data recommended information
decision-making, 33, 39, 41, 42, altruistic patients, 11, 15–16, 20
sources, 464–6 behavioural interventions, 20,
82, 97, 279, 394, 427, 433, recommended methods for
437–8 21
operating an EWS, blinding, 7, 12, 20
task-analysis model, 252 466–7
see also consensus development cluster trials, 8, 12, 20, 21
timescale and purposes, 453 competent patients, 14, 16, 20
methods economic evaluation, 2, 273
decision models/modelling, consent, 8, 11, 12, 16, 17, 20
and costing, 219
488–91, 492, 493, 497 incompetent patients, 16, 20
and meta-analysis, 404
decision theory, 11 interim analysis, data and
and RCTs, 48
Declaration of Helsinki, 28 results, 12, 19, 20
handling uncertainty in, 314–37
default priors, 282; see also prior placebo controlled trials, 12,
distribution, Bayesian health status measures in,
195–214 18–21
methods power calculations, 19
Delphi method, 428, 429, 434, preliminary evaluation of health
technologies, 485–98 ‘prior’ probabilities, 8, 14, 15,
444; see also consensus 17
development methods economic welfare theory, 219
editorial policy of journals, 376; randomised controlled trials,
Delphi study, for identifying new design and conduct of,
health care technologies, see also publication bias
effect modification, definition, 64 8, 11–24
455, 456, 459, 465 restrictions on trial treatments,
deontological approaches, 28, 30 effect size, 128, 129, 186–7
effectiveness, 40 17
design effect, 302–3, 305–6
assessment of, 141 sample size, 19
Detsky model, 487
Diagnosis Related Groups cost see cost effectiveness trial effect, 17
(DRGs), 492 datasets for assessing, 142–4 trial entry, 17
diagnostic tests, Bayes's theorem differences in measured, 97 underpowered clinical trials, 12,
in, 279 of healthcare interventions, 18, 20
diary cards, 222 study designs for the ethics
diffusion, 141, 142–4, 451, 470, evaluation of, 116–35 and randomisation, 283–4
474–5 efficacy, 46 randomised controlled trials, 8,
dimension-specific instruments, assessment of, 141 11–24, 25–37
183 routine datasets for the cultural factors, 25–37
disaggregation of results, 225 assessment of, 142–4 economic factors, 25–37
discount rates, derivation, 239, studies, 107 social factors, 25–37
241 see also self-efficacy Ethics Committees, 31, 34
502 The Advanced Handbook of Methods in Evidence Based Healthcare
European Organisation for lessons from case studies, 479 for presenting questions, 260,
Research and Treatment in publication trends analysis, 261–3
Cancer Questionnaire 479–81 for presenting response
(EORTC QLQ-C30), 190, chorion villus sampling, categories, 260, 263
338, 339, 341, 344, 348 476–7, 479, 481, 482 for quality assessment of
EuroQol EQ-5D questionnaire, gene therapy for cystic randomised trials, 413,
183, 196, 197, 205, 206, fibrosis, 478 414, 423
210 when to assess, 473–9 for question ordering, 260, 263
EuroQol health state, 201, 237, 15-D, 197, 205, 206 for systematic review/meta-
238, 239 Fisherian approach, 276–7 analysis practice, 392
Evidence, 367–9 fixed effect models, meta-analysis, quality assessment, 410–11
applying, 60, 62, 63 393 reporting, 47
critical appraisal of, 368 fixed effect regression model, 401 GUSTO trial and the Bayesian
from randomised controlled follow-up, 41, 44, 46 approach, 283
trials, 39 cohorts of registered studies,
generalised synthesis of, 403 373
grading, 368 design issues, 301 Hawthorne effect, 11, 17
identifying and filling gaps in, loss to, 44 Health Assessment Questionnaire
449–50 quality of life (QoL) and (HAQ), 189
in primary care, 62 survival studies, 356 healthcare innovations, early
integrating, 368 Food and Drug Administration warning system (EWS) for,
interpretation and delivery, 368 (FDA), 377, 457, 460, 465 451–70
synthesis, 288 forest plot, 397, 398 healthcare interventions see
evidence based healthcare, freedom of information, 377 intervention studies
overview, 1–5 frequentist approaches, 66 ‘Health of the Nation’ initiative,
evidence based medicine, funnel plot, 379–82, 385 33–4
overview, 1–5, 391 futures studies, 469–70 health-related quality of life
exchangeability, concept of, 283 (HRQoL), 182, 237
exclusion criteria, 8, 31, 34, 56, health service costs see cost data
57, 58, 59, 68, 103, 155, Galbraith radial plot, 381 and costing
217, 249, 257, 410, 418 General Health Rating Index, 190 health state, 237
exclusive trials, versus inclusive General Household Survey for movement between, 361–2
trials, 107 England and Wales, 183 valuation techniques, 197, 198,
expectancy, 80–3 General Practice Research 200–1, 202–4
criteria for verifying presence Database (GPRD), 143 see also health status
of, 81 generalisability, 7, 39, 58, 61, 106, health status, 136, 142, 144, 182,
definitions, 80 110, 170, 391 186–9
treatment-related, 80 and selective participation, 110 datasets, 142–4
types of, 81–2 definition of terms relating to, health status measures, 181–94,
expected 57 195–214
incremental cost–effectiveness see also external validity and criteria for assessing, 183–94
ratio (EICER), 495 applicability of results in economic evaluation,
net benefits of the research generalised estimating equations 196–214
(ENBR), 492 (GEEs), 306–8, 310 multi-attribute utility scales,
value of perfect information generalised synthesis of evidence, 204–6, 210
(EVPI), 492 403 non-preference based measures,
value of sample information generic instruments, 183; see also 206–8, 210
(EVSI), 492 patient based outcome QALY and non-QALY
experimental interventions, 297 measures measures, 208–9
external validity, 7, 59, 95–115 GREAT (Grampian Region Early selecting a measure for use in
assessment of, 107 Anisteplase Trial), 277–9 economic evaluation,
improving, 107–8 grey literature bias, 373; see also 200, 209–10
see also generalisability, publication bias techniques for health state
validity group consent, 33 valuation, 200–4, 210
extrapolation, definition, 57 group decision-making, 434–5; see types of, 182–3, 195–6
extreme scenario analysis, 317 also consensus validity, 199–200, 202
development methods see also individual measures of
face and content validity, 185–6; group judgement, 435 health status
see also validity reliability of, 437 health surveys, 144; see also
factorial designed studies, 8 guidelines questionnaire surveys
factorial trials, 8, 287 consensus development methods health technology assessment
fail-safe N method, 385 and clinical guidelines, (HTA), 1–5, 92, 179, 275,
fast changing health technologies, 426–48 367, 453, 486–95
471–84 CONSORT, 8, 39, 47, 50, 107, and routine datasets, use of,
case studies, 475–8 424 136–48
diffusion and development of, development of clinical practice Bayesian philosophy and HTA,
474–5 guidelines, 68 277–80
Index 503
classical statistical approach in individualised measures of health, magnitude estimation (ME), 201;
HTA, 276–7 183 see also health status
definition of, 141 information sources (for measures
diffusion and equity of HTAs, identifying new health care Mann–Whitney U-test, 346
141 technologies), 456, 457–61 Mantel–Haenszel method, 395,
meta-analysis in HTA, 391–408 criteria for assessing, 459 397
methods, 2, 4, 7 for identifying case studies, 460 Markov Chain Monte Carlo
overview, 1–5 primary, secondary and tertiary, (MCMC) methods, 280,
prioritising, 485–98 470 288, 333, 403
programmes, 1, 4, 182, 462–3, recommended, 464–6 Medicines Control Agency
485 information technology and (MCA), 478
qualitative methods in HTA, routine data, in healthcare, mental hospital admissions data,
166–78 139 139
key issues and controversies, informative censoring, 360 meta-analysis, 96, 275, 280, 282,
167–9 informative drop-out, 352–4, 363 391–408, 410–25
qualitative research in HTA, informed consent, 12, 16, 17, 20, and Bayesian methods, 288,
169–75 29–33, 42 402–3
timing of HTAs, 471–84 cultural factors, 32–3 and economic evaluation, 404
health technology only, datasets, innovation, 469–70 and fixed effect models, 393–8,
see datasets intention-to-treat analysis, 39, 44, 401
Health Utility Index (HUI-I, HUI- 45–7 and publication bias, 372–3,
II, HUI-III), 197, 205, 206, interaction 378–83, 385, 404
210 definition, 64 Binary outcomes, 393–7
heterogeneity self-efficacy, 81, 82 continuous outcomes, 398
assessing, 398–9 interim analysis of RCTs, 12, 19, cumulative, 404
definition, 64 20, 284–5 epidemiological studies, 403
in study findings, 110 internal consistency, 184 generalised synthesis of
in treatment effect, 65 internal validity, 95–115; see also evidence, 403
test for, 398 validity guidelines for good practice,
underlying causes of, 400–1 International Classification of 392
see also meta-analysis Diseases (ICD), 140, 449 heterogeneity, assessing, 398–9
hierarchical modelling, 273, 283 International Society for Health individual patient data, 404
hierarchical structure of data, 296 Technology Assessment, 1 inverse variance method, 394
historical controls, 287–8; see also intervention studies, 297, 298 meta-regression, 400
Bayesian methods area-based, or organisation- missing data, 404
homogeneity, test for, 398 based, 295–313 of observational studies, 403
horizon scanning, 451–70; see interviewing, structured, 281 ordinal outcomes, 398
also early warning intra-class correlation coefficient, quality assessment of
systems; new healthcare 302–5 randomised trials,
technologies inverse variance method, 394–5 included in, 404, 409–25
Hospital Anxiety and Depression development of guides for,
Scale (HADS), 338–40, judgements in concensus
development methods, 415–21
343, 345, 347–8 random effects models,
Hospital Episode Statistics (HES), 433–4, 438, 443–4
individual, 438, 443 399–400, 401
138, 143 search strategy to identify, 411
HUI see Health Utility Index group, 435
methods of synthesising, 440–4 meta-regression, 120, 400–2
Human Fertility and Embryology
justice minimal clinically important
Authority, 138, 142
in healthcare, 33 difference (MCID), 189
to participants after trial, 34 Minimum Data Set for outpatients,
incidental factors of treatments, 74 143
Kaplan–Meier survival analysis, missing data, 223, 363, 404
inclusion criteria, 8, 43, 45, 56, 223, 357, 358, 359
57, 58, 66, 155, 208, 214, modified nominal group technique
Kappa statistics, 120, 126, 132, (NGT), 429
216, 217, 219, 221, 225, 420, 422
238, 249, 250, 257, 259, morbidity data, 139, 151
265, 392 L'Abbe, plot, 398 mortality statistics, 138
inclusive trials, versus exclusive life-table analysis, 223 multi-attribute utility scales, 197,
trials, 107 Likert scale, 188, 190, 429 198, 204–6, 210
incompetent patient, 16, 20; see limited disease registers, 144; see characteristics of, 205
also informed consent also datasets/databases, practicality of, 204
incremental cost-effectiveness health status only reliability of, 205
ratio (ICER), 315, 316, linear regression analysis, 381, validity of, 205–6
322, 323, 326, 327, 331–3 382 valuation of, 206
league table of analysed ICERs, logistic regression analysis, 67, see also individual multi-
324–5 160, 161, 308, 409, 411 attribute utility scales
individual-level data analysis, 296, lump-and-smear priors, 282; see multi-centre analysis, and
305 also priors distribution Bayesian methods, 286
504 The Advanced Handbook of Methods in Evidence Based Healthcare
multi-level analysis, 307–10 open-ended questions, 237–8, 240, competent, 14, 16–17, 20
multi-state models, in survival 241, 246, 264 incompetent, 16, 20
analysis, 352, 355, 361–3, opportunity costs, 18, 141, 203, Patient's Expected Event Rate
364 232 (PEER), 66–7
multivariate models of risk, 108 organisation-based health patient preferences, 98, 105, 109
N of 1 trials, 286–7 interventions, methods for payback models, 286
see also Bayesian methods evaluating, 295–313 peer reviewing and publication
National Prospective Monitoring check-list, 309 bias, 376
Scheme for HIV, 142 cluster-level evaluation, 296 percutaneous transluminal
negative outcome expectancy, 80, individual-level evaluation, 296 coronary angioplasty
82 problems of organisation-based (PTCA), case studies, 98,
new healthcare technologies, evaluations, 296 99, 104–5, 152, 155–62,
identifying, 451–70 study designs, 297–301 157–63
impact and uptake, 150 see also area-based health performance bias, 41
information sources on, 456–61 interventions person trade-off (PTO), 201–4
see also early warning systems, outcome expectancy, 80, 82; see Peto method, 397
horizon scanning also placebo effect pharmacokinetics, 287
Neyman–Pearson approach, 276–7 outcome frequencies, 122, 123 Phase I, II, III trials, 287, 384
NHS Central Register, 139 outcome measures, 41 placebo controlled trials, ethical
NHS Centre for Reviews, 450 see also patient based outcome issues, 12, 19–20, 28
NHS Research and Development measures placebo effect, 9, 73–90
Programme, 1, 4, 49, 182, outcomes defining, 73
485 binary outcome data, 393–4, determinants of, 75–6
nominal group technique (NGT), 401 expectancy, 80–4; see also
429, 434, 438; see also combining outcomes, 398 expectancy
consensus development defining, 40–1 implications for health service
methods trade-offs between, 67 staff and practitioners,
non-health service costs, 217–18 see also health status, patient- 85
non-informative prior based outcome measures, mechanisms of, 75
distributions, 282 quality of life model of, 77
non-patient-specific resource use outcomes-only data sets, 136, 144 patient characteristics, 76
methods, 223 P-values, 275, 276–7, 282, 297, practitioner factors, 76
non-preference based measures of 336, 379, 393 practitioner–patient interaction,
health, in economic paired designs, 300 76
evaluation, 206–8, 209, pairwise comparisons, 238, 239 treatment characteristics, 76
210 partisanship in research, 174 placebo sag, 75
non-randomised studies, 95–115, partitioned survival analysis, pooled estimates, 128, 129
299 359–61 pooling data, 394–5, 399–400
and RCTs, differences between PATHS model, 495–7 population effectiveness,
results, 99–105 patient-based outcome measures, aggregated datasets for,
the role of, 111 181–94 143
threats to internal and external acceptability, 189–90 populations involved in research
validity, 95–115 appropriateness, 184 and clinical practice,
see also ‘quasi-experimental’ ceiling and floor effects, 187 definitions, 57
and observational studies definitions, 182 PORT studies, 138
(QEO) feasibility, 190–1 postal surveys, 255
non-resource costs, 218 internal consistency, 184 power calculation, 19, 150, 151,
null hypotheses, 284, 379 interpretability, 188–9 284, 488, 490
number needed to treat (NNT), precision, 187–8 practitioner–patient interaction, 76
65–7, 348 reliability, 184–5 pragmatic studies, 107
Nuremberg Code, 33 reproducibility, 185 predictive distribution, 279; see
observational evidence, 91–3 responsiveness, 186–7 also Bayesian methods
observational studies, 298 scoring methods, 188 preference-based measures of
and Bayesian methods, 287–8 time to complete, 189–90 health, 200; see also health
and meta-analysis, 403 types of instruments, 182–3 status measures
for the evaluation of validity, 185–6 preference trials, 109
effectiveness of health patient data, individual, 404 Prescription Pricing Authority, 143
care interventions, patient-level cost data, data sets, primary care, using evidence in,
116–35 328 62
see also quasi-experimental and patient participation, barriers to, primary research, 2, 4, 40, 74, 79,
observational studies 42–3 80, 249, 258, 259, 260,
(QEO) patient preferences, 30 261, 264, 310, 460
odds ratio (OR), 346–7, 393–7, patient-related self-efficacy prior distribution, 279–83, 402–3
420 expectations, 81 default priors, 282
Office of Population Census and patients enthusiastic priors, 282
Surveys (OPCS), 140, 153 autonomy, 29 lump and smear priors, 282
Index 505
ratio of odds ratios (ROR), 415, case studies, 152–62 social desirability bias, 254; see
416 Routine datasets, 136–48 also bias
Read codes, 140 for efficacy/effectiveness, and social support mechanism, and the
regression analysis, 204, 235, 241, equity diffusion, 142–44 placebo effect, 75
305, 306, 307, 308, 400 census related, 144 socio-cultural factors in RCTs,
regression methods for clustered health technology and patient 25–37
data, 306–7 and/or health status Sociofile, 26, 27, 426, 432
regression plot, 402 datasets, 142–3 standard error, 305–10, 327, 328,
relative risk, 62, 63, 64, 65, 66, health technology only, 143–4 357, 359, 380, 395, 397,
67, 106, 120, 122, 124, health status only, 144 401
158, 161, 162, 394 standard gamble (SG), 201, 203,
reliability 204, 210
group judgement, 436–7 Safety and Efficacy Register of standardised coding systems, 140
of patient based outcome New Interventional standardised response mean
measures, 184–5 Procedures (SERNIP), 140, (SRM), 187
religion, and ethics of clinical 464 STATA Version 4.0, 120
sample size, 19, 41, 284, 338–51 stochastic analysis, 316
trials, 27, 28, 30, 31
calculation, 220–1, 304, 340; stratified designs, 299, 300–1
Renal Registry, 142
see also power study design, 297–301, 363–4
repeated cross-sectional design,
calculation study protocols, 39, 44
301
comparison, 347–8 study quality, 404
reporting, quality of, 47–8
confounding by, 125–6 subgroup analysis, 5, 7, 8, 153–4,
reproducibility, of patient based determination, QoL measures,
outcome measures, 185 155, 158–9, 161
338–51 of treatment effects, 60
Research Ethics Committees, 31, methods, 345–7
34 subset analysis, 286
requirements in organisation or survival analysis
resource use data area-based evaluations,
analysis of, 223–4 multi-state, 361–3
301–5 Proportional Hazards
collection of, 215–30 scales
methodological issues, 216–25 Regression model
for measuring quality, 417–20 (PHREG), 362
presentation of, 224–5 versus components, 420–1
Resource Utilisation Survey quality-adjusted, 352, 355–61
sceptical prior distributions, 282, survival data
(RUS), 222 285–6
respondent validation, 171 simultaneous analysis of Quality
see also prior distribution of Life and, 352–66
response bias, 253–4; see also bias Schedule for the Evaluation of
restricted allocation, 299, 300 System for Information on Grey
Individual Quality of Life Literature in Europe
restricted randomisation, 299, 300 (SEIQoL), 183
restrictive studies, 17–18, 96 (SIGLE), 473
Scottish Morbidity Record Form 1 systematic reviews, 2, 47, 63,
risk-based individualisation of (SMR1), 151, 152, 153
benefit and harm, 60 367–9
secondary research, 2 and meta-analysis, 391–3
risk-benefit approach, 60, 62 selection bias, 7, 12, 39, 41, 46,
risk difference, 122 and publication bias, 378–83
131, 141, 160, 221, 222, applicability of results, 56–72
risk reductions, absolute, 65 223, 298, 385; see also
individuals, 66–7 conduct and reporting of, 68
bias guidelines for good practice,
robust standard errors, 306 selection models, 382
Rosser disability and distress 392
self-efficacy expectations, 80, 81; search strategy to identify, 411
classification, 197, 205 see also placebo effect,
Rotterdam Symptom Checklist survey of published, 383
expectancy
(RSCL), 338, 339, 340–1, semi-Markov process, 362
344, 346–8 sensitivity analysis, 224, 286, Taylor series method, 333
routine data, 136–48, 149–65 316–17, 320, 359, 421 Technology Assessment Priority
definition, 138–9 SF-36 health survey status Setting-System (TAPSS)
in assessment of costs of measure, 183, 188, 195–6, model, 487, 494
technologies, 141–2, 144–5 207 telephone interviews, 255
in assessment of diffusion of Sickness Impact Profile, 190 TEMPUS project (The Estimation
health technologies, 141, simultaneous analysis of quality of of Marginal Time
142–4, 150–1 life (QoL) and survival Preference in a UK-wide
in assessment of efficacy or data, 352–69 Sample), 231–46; see also
effectiveness, 141, 142–4, case study – the MIC study, time preferences for health
150 353–4 tertiary research, 2
information technology and, 139 methods, 354–63 thermometer rating scale, 201; see
limitations of, 163–4 multi-state survival analysis, also visual–analogue scale
linking of data, 139–40, 144 361–3 threshold analysis, 421; see also
uses of, 162–3 quality adjusted survival sensitivity analysis
using to complement and enhance analysis, 355–61 threshold utility analysis, 359, 361
the results of RCTs, simultaneous modelling, 363 time horizon and measurement of
149–65 simultaneous modelling, 363 costs, 219–20
Index 507
time preferences for health, treatment restrictions in trials, threshold utility plane, 359
eliciting 17–18 see also cost–utility analysis;
closed-ended method: discrete treatment target population, 57 Health Utility Index; multi-
choice modelling, 238–9, trial entry and distress, 17 attribute utility scales
240 triangulation, in qualitative utility–attributes relationship, 238
constant timing aversion, 240 research, 172
decreasing timing aversion, 240 trim and fill method, 381–2
discount rates, 239–40, 241 t-test, 187, 240, 305, 328, 411 validity, 185–6
estimation of time preferences, Type I error, 275, 276, 284–5 assessment methods, 432–3
232–6 Type II error, 276 construct validity, 186
health state, 237 criterion, 185
literature, 234 face and content validity, 185–6
uncertainty, 11, 13, 14, 42 internal and external, threats to,
method of, 235 absolute, 13
non-fatal changes in health, 235 95–115
collective, 14, 20 of qualitative research, 171–2
open-ended method, 237, 240 dealing with, 493
stated preference, using to valuation techniques for health
epistemiological ‘uncertainty’ status measures, 210
estimate, 233–5 principle, 13
see also TEMPUS project empirical validity of, 204
in economic evaluation, 224, variance inflation factor, 301
time trade-off (TTO) technique, 314–37
201, 203, 204 variation, 110, 127, 138, 297, 298,
under-powered trials, 18, 20
timing of surgery, 154–5 353, 398, 400, 403, 486,
univariate individual-level
transferability of results, 61–6; see analyses, 305–6 487, 498
also applicability of results unpublished trials, 378–9 between-cluster variation,
transition rates, 362–3 estimating number of, 378–9 301–2, 304, 306, 308
Transplant Support Services locating, 378 cultural, 33
Authority (TSSA), 142 see also publication bias within-cluster, 301–2
treatment distribution population, U.S. Panel on Cost–Effectiveness visual–analogue scale (VAS)
57 in Health and Medicine, techniques, 200–4, 206
treatment effect, 7, 12, 30, 41, 58, 323
59, 64 utility warranted assertability, 171
interpretation of, 348–9 discounted utility (DU) model, within-cluster variation, 301–2
sources of heterogeneity in, 65 232
subgroup analysis of, 60 measures, 183
variance with baseline risk threshold utility analysis, 359, Zelen randomisation, 12, 20, 107,
level, 65 361 286