Skip to main content
UKPMC Funders Author Manuscripts logoLink to UKPMC Funders Author Manuscripts
. Author manuscript; available in PMC: 2013 Oct 15.
Published in final edited form as: J Epidemiol Community Health. 2012 May 10;66(12):1182–1186. doi: 10.1136/jech-2011-200375

Using natural experiments to evaluate population health interventions: new MRC guidance

Peter Craig 1,*, Cyrus Cooper 2, David Gunnell 3, Sally Haw 4, Kenny Lawson 5, Sally Macintyre 6, David Ogilvie 7, Mark Petticrew 8, Barney Reeves 9, Matt Sutton 10, Simon Thompson 11
PMCID: PMC3796763  EMSID: EMS54487  PMID: 22577181

Abstract

Natural experimental studies are often recommended as a way of understanding the health impact of policies and other large scale interventions. Although they have certain advantages over planned experiments, and may be the only option when it is impossible to manipulate exposure to the intervention, natural experimental studies are more susceptible to bias. This paper introduces new guidance from the Medical Research Council to help researchers and users, funders and publishers of research evidence make the best use of natural experimental approaches to evaluating population health interventions. The guidance emphasises that natural experiments can provide convincing evidence of impact even when effects are small or take time to appear. However, a good understanding is needed of the process determining exposure to the intervention, and careful choice and combination of methods, testing of assumptions and transparent reporting is vital. More could be learnt from natural experiments in future as experience of promising but lesser used methods accumulates.

Introduction

Natural experimental studies are often recommended as a way of understanding the impact of population-level policies on health outcomes or health inequalities.[1-4] Within epidemiology there is a long tradition, stretching back to John Snow in the mid nineteenth century,[5] of using major external shocks such as epidemics, famines or economic crises to study the causes of disease. A difficulty in applying similar methods to the evaluation of population health policies and interventions, such as a ‘fat tax’ or a legal minimum price per unit of alcohol, is that very often the change in exposure is much less extreme, and its effects may be subtle or take time to emerge. Although they have certain advantages over planned experiments, for example by enabling effects to be studied in whole populations,[6] and may be the only option when it is impossible to manipulate exposure to the intervention, natural experimental studies are more susceptible to bias and confounding. It is therefore important to be able to distinguish situations in which natural experimental approaches are likely to be informative from those in which some form of fully experimental method such as a randomised controlled trial (RCT) is needed, and from those in which the research questions are genuinely intractable.

The Medical Research Council has recently published guidance to help researchers and users, funders and publishers of research evidence make the best use of natural experimental approaches to evaluate population health interventions (www.mrc.ac.uk/naturalexperimentsguidance). Following the model of the MRC complex interventions guidance,[7] it was written by a multidisciplinary team with experience of evaluation using a wide range of research designs. The ideas were developed and tested in two specially convened workshops of population health researchers. Drafts were reviewed by workshop delegates and by the MRC’s Methodology Research Panel. The guidance is meant to help researchers plan and design evaluations of public health interventions, journal editors and reviewers to assess the quality of studies that use observational data to evaluate interventions, and policy-makers and others to recognise the strengths and limitations of a natural experimental approach. In this paper we summarise the main messages of the guidance.

What are natural experiments?

The term ‘natural experiment’ lacks an exact definition, and many variants are found in the literature.[8-10] The common thread in most definitions is that exposure to the event or intervention of interest has not been manipulated by the researcher. Outside an RCT it is rare for variation in exposure to an intervention to be random, so special care is needed in the design, reporting and interpretation of evidence from natural experimental studies, and causal inferences must be drawn with care.

Why are natural experiments important?

Alternatives to RCTs have been advocated by policy-makers and researchers interested in evaluating population-level environmental and non-health sector interventions[11] and their impact on health inequalities.[4] Such interventions may be intrinsically difficult to manipulate experimentally – as in the case of national legislation to improve air quality, or major changes in transport infrastructure[12] – or implemented in ways that make a planned experiment difficult or impossible, for example with short timescales or extreme variability in implementation.[13] It may also be unethical to manipulate exposure in order to study effects on health if an intervention has other known benefits, if it has been shown to be effective in other settings, or if its main purpose is to achieve non-health outcomes.[14] Even if such ethical and practical restrictions are absent, an RCT may still be politically unwelcome.[15]

Natural experimental approaches are important because they widen the range of interventions that can usefully be evaluated beyond those that are amenable to planned experimentation. For example, suicide is rare in the general population, occurring at a rate of about 1/10,000 per annum. Even in high risk populations, such as people treated with antidepressants, the annual incidence is only around 1/1000. Clinical trials would have to be enormous to have adequate power to detect even large preventive effects, but natural experiments have been used effectively to assess the impact of measures to restrict access to commonly used means of suicide[16-18] and inform the content of suicide prevention strategies in the UK and worldwide.[19]

These and other studies in which a natural experimental approach has produced clear cut evidence of health impacts are summarised in Supplemental Table 1. They illustrate the diversity of interventions that have been evaluated as natural experiments, and the wide range of methods that have been applied. Many of the studies have benefited from the availability of high quality, routinely collected data on exposures, potential confounders and outcomes and substantial, rapid changes in exposure across a whole population, which reduces the risk of selective exposure or confounding by secular trends and increases the confidence with which changes in outcomes can be attributed to the interventions. However, it is misleading to assume that whenever a planned experiment is impossible, there is a natural experimental study waiting to happen. Some but not all of the ‘multitude of promising initiatives’[1] are likely to yield good natural experimental studies. Care, ingenuity and a watchful eye for good opportunities are needed to realise their potential.

When should natural experiments be used?

The case for adopting a natural experimental approach is strongest when: there is a reasonable expectation that the intervention will have a significant health impact, but scientific uncertainty remains about the size or nature of the effects; an RCT would be impractical or unethical; and the intervention or the principles behind it have the potential for replication, scalability or generalisability.

In practice, natural experiments are highly variable, and researchers face difficult choices about when to adopt a natural experimental approach and how best to exploit the opportunities that do occur. The value of a given natural experiment for research depends on a range of factors including the size of the population affected, the size and timing of likely impacts, the processes generating variation in exposure, and the practicalities of data gathering. Quantitative natural experimental studies should only be attempted when exposed and unexposed populations (or groups subject to varying levels of exposure) can be compared, using samples large enough to detect the expected effects, and when accurate data can be obtained on exposures, outcomes and potential confounders. Resources should only be committed when the economic and scientific rationale for a study can be clearly articulated.

Design, analysis and reporting of natural experiments

Planned and natural experiments face some of the same threats to validity, such as loss to follow-up and inaccurate assessment of exposure and outcomes. The key difference is that RCTs have a very general and (if used properly) effective method of minimising the bias that results from selective exposure to the intervention, i.e. the tendency for exposure to vary according to characteristics of participants that are also associated with outcomes. In the case of non-randomised studies, there is no such general solution to the pervasive problem of confounding.[20] Instead there is a range of partial solutions which can be used in some, often very restricted, circumstances but not others. Understanding the process that produces the variation in exposure (often referred to as the ‘assignment process’ even when there is no deliberate manipulation of individuals’ exposure[21]) is therefore critical to the design of natural experimental studies.[9]

Design

A study protocol should be developed, and ideally published, whatever design is adopted. Good practice in the conduct of observational studies, such as prior specification of hypotheses, clear definitions of target populations, explicit sampling criteria, and valid and reliable measures of exposures and outcomes, should apply equally to natural experimental studies.

All natural experimental studies require a comparison of exposed and unexposed groups (or groups with varying levels of exposure) to identify the effect of the intervention. The examples of suicide prevention,[16-17] indoor smoking bans[22-24] and air pollution control[25-26] show that simple designs can provide convincing evidence if a whole population is abruptly exposed to an intervention, and if the effects are large, rapidly follow exposure and can be measured accurately at population level using routinely available data. This combination of circumstances is rare, and more complex designs are usually required.

Natural experiments can also be used to study more subtle effects, so long as a suitable source of variation in exposure can be found, but the design and analysis of such studies is more challenging. In any case, what is often required is an estimate of effect size, and a large observed effect may incorporate a large element of bias due to selective exposure to the intervention. Whatever the expected effect size, care should be taken to minimise bias in the design and analysis of natural experiments.

Design elements that can strengthen causal inferences from natural experimental studies include the use of multiple pre/post measures to control for secular changes, as in an interrupted time series design;[27] multiple exposed/unexposed groups that differ according to some variable that may affect exposure and outcome to assess whether selection on that variable is likely to be an important source of bias;[9] accurate measurement of multiple potential confounders and combinations of methods to address different sources of bias. In a study that exemplifies many of the features of a rigorous approach to identifying relatively small effects, Ludwig and Miller[28] used variation in access to support for obtaining Headstart funding to model exposure, and compared a variety of outcomes among children who were above or below the age cutoff for access to Headstart services (Box 2; Table 1).

Box 2. Selection on unobservables.

Difference in differences

This method compares change over time in exposed and unexposed groups.[36] The differencing procedure controls for unobserved individual differences, and for common trends. It assumes that the unobserved characteristics are fixed, and that the outcomes in each group would change in the same way in the absence of the intervention, so is vulnerable to changes in the composition of the groups and to external influences that differentially affect the exposed and unexposed groups.

Instrumental variables

An instrumental variable (IV) is a factor, such as treatment assignment in a well-designed RCT, which is associated with outcomes only via its association with exposure to the intervention and is independent of other factors associated with exposure. IVs have been used to identify the impact of treatment from routine data.[37-39] In these studies, variables such as distance from a specialised centre have been used to evaluate novel treatments, the assumption being that patients living close to a specialised centre are more likely to receive the novel treatment, but are otherwise similar to other patients.

Regression discontinuity designs

This approach exploits a step change or ‘cutoff’ in a continuous variable used to assign treatment or otherwise determine exposure to an intervention. The assumption is that units (individuals, areas, etc.) just below and just above this threshold will otherwise be similar in terms of characteristics that may influence outcomes, so that an estimate of treatment effect can be obtained by comparing regression slopes on either side of the cutoff. When the Headstart programme to improve the health of disadvantaged children was first implemented in the US, help with applying for funding was targeted on the 300 poorest counties, and a higher proportion of those counties received funding. Ludwig and Miller[28] compared regressions of child mortality on poverty for counties either side of the cutoff, and found lower than expected mortality in those that qualified for assistance.

Table 1.

Examples of natural experimental studies to evaluate population health interventions

Population Intervention Comparator Outcome Analysis method Results Reference
Sri Lanka -
whole
population
Legal restriction on
pesticide imports
Suicide rates
pre-ban;
method specific
suicide rates;
rates of non-
fatal self-
poisoning
Overall suicide
mortality and
mortality
mortality from
suicide by self-
poisoning with
pesticide
Graphical
analysis of
trends
Import bans reduced method-specific and
overall suicide mortality
Gunnell et al
2007[16]
UK - 12-17
year olds
Restriction on
prescribing SSRIs to
people aged <18
Suicidal
behaviour pre-
restriction
Hospitalisations
for self-harm;
suicide
mortality
Joinpoint
regression
Restriction on prescribing of SSRIs was
not associated with changes in suicidal
behaviour
Wheeler et
al 2008[18]
Finland -
men and
women
aged >15
Reduction in alcohol
taxes
Alcohol related
mortality before
the tax change
Alcohol-related
mortality
Poisson
regression to
obtain relative
rates; timeseries
analysis
Alcohol-related mortality increased,
especially among the unemployed
Herttua et al
2008,[40]
Herttua et al
2011[41]
Hong Kong
- whole
population
Legislation to restrict
sulphur content of fuel
Mortality pre-
restriction,
overall and by
district
All-cause and
cause specific
mortality
Poisson
regression of
change in
seasonally
adjusted death
rates
Cardiovascular, respiratory and overall
mortality fell post-restriction; decline
greater in districts with larger falls in SO2
concentrations
Hedley et al
2002[26]
Dublin –
whole
population
Ban on coal sales Mortality pre-
ban, and in the
rest of Ireland
Non-trauma,
respiratory and
cardiovascular
mortality
Interrupted
time-series
Non-trauma, respiratory and
cardiovascular death rates fell post-ban
Clancy et al
2002[25]
Scotland –
patients
admitted to
9 hospitals
Legislative ban on
smoking in public
places
Hospitalisations
pre-ban and in
England
Hospitalisations
for acute
coronary
syndrome
Comparison of
numbers of
admissions pre
and post ban
Admissions for acute coronary syndrome
fell among both smokers and nonsmokers
post-ban
Pell et al
2008[23]
England –
patients
aged >17
Legislative ban on
smoking in public
places
Hospitalisations
pre-ban
Emergency
admissions for
myocardial
infarction
Interrupted
time-series
Small but significant fall in emergency
admissions in the first year post-ban
Sims et al
2010[24]
India –
pregnant
women
Cash incentives to use
a health facility to give
birth
Districts with
low rates of
take up; births
to women not
receiving
payments
Use of health
facilities; infant
and maternal
mortality
Matched and
unmatched
comparisons of
recipient and
non-recipient
births;
difference-indifferences
analysis of
district level
usage and
mortality rates
Higher rates of take-up of the incentives
were associated with higher proportions of
births within health care facilities; use of
health care facilities to give birth was
associated with fewer perinatal and
neonatal deaths. There was a non-
significant reduction in maternal deaths
Lim et al
2010[30]
England –
general
practitioners
Abolition of GP
fundholding
Non-
fundholding
practices; pre-
abolition
admission rates
Referral for
elective and
emergency
admissions
Difference-indifference
analysis of
referrals from
fundholders and
non-fundholders
Fundholders had lower rates of elective
referral while fundholding was in operation
and their rates of referral increased more
than those of non-fundholders following
abolition. There was no difference in
emergency admissions pre or post
abolition.
Dusheiko et
al 2003[31]
USA – low
income
families
with
children
aged 3-5
Headstart - help with
parenting, nutrition,
health and social
services and schooling
US counties
with poverty
levels above the
cutoff used to
allocate help
with accessing
Headstart
funding
Mortality from
causes of
death
amenable to
Headstart
services
Regression
discontinuity
design,
comparing
regressions of
mortality on
poverty for
counties above
and below the
cutoff
For Headstart-related causes, there was a
discontinuity in mortality rates at the
cutoff, but no difference in deaths from
other causes or among children too old to
qualify for Headstart services
Ludwig and
Miller
2007[28]
USA –
patients
admitted to
hospital
with acute
myocardial
infarction
Invasive cardiac
treatment
(catherisation followed
by revascularisation)
Non-invasive
cardiac
treatment
Long term (i.e.
7-year)
mortality
Instrumental
variable
analysis using
regional
catherisation
rates as the
instrument
Cardiac catherisation was associated with
lower mortality; instrumental variable
analyses produced smaller estimates of
effect than analyses (multivariate risk
adjustment, propensity score-based
methods) that only adjust for observed
prognostic factors
Stukel et al
2007[39]

Analysis

The defining feature of a natural experiment is that manipulating exposure to the intervention is impossible. There are a few examples where assignment is by a ‘real life’ lottery, but selection is the rule and a range of methods is available for dealing with the resulting bias.

Where the factors that determine exposure can be measured accurately and comprehensively, matching, regression and propensity scores can be used to reduce confounding (Box 1). Bias will remain if there are unobserved or imperfectly measured factors that influence both exposure and outcomes. Given the difficulty of measuring accurately all of the characteristics associated with exposure to an intervention, methods such as difference in differences, instrumental variables and regression discontinuity designs that deal with unobserved factors are a potentially valuable advance on those that only deal with observed factors (Box 2).

Box 1. Selection on observables.

Matching

This involves finding unexposed individuals (or clusters of individuals) which are similar to those receiving the intervention, and comparing outcomes in the two groups.

Regression adjustment

Measured characteristics that differ between those receiving the intervention and others can be taken into account in multiple regression analyses.

Propensity scores

The likelihood of being exposed to an intervention given a set of covariates can be estimated by logistic regression[35] and used to match exposed with unexposed cases, or for covariate adjustment.

In practice, none of these approaches provides a comprehensive solution to the central problem of selective exposure to the intervention.[20] Methods of controlling for observed factors associated with exposure are vulnerable to selection on unobservables. Methods for dealing with selection on unobservables require strong and untestable assumptions[29] and their use is restricted by the often very limited availability of variables that can be used to model exposure. They are therefore best used in conjunction with additional tests for the plausibility of any causal inferences.

Combining methods that address different sources of bias and comparing the results is one such approach and there are several examples in Table 1. In their evaluation of a conditional cash transfer scheme to encourage women to use health facilities to give birth, Lim et al[30] combined methods for dealing with selection on both observable and non-observable characteristics. Another useful technique is to analyse outcomes that are not expected to change. Dusheiko et al[31] used trends in emergency admissions as a non-equivalent dependent variable to test whether changes in elective admissions could plausibly be attributed to GP fundholding, while Ludwig and Miller[28] compared mortality from causes that were likely or unlikely to respond to Headstart services.

Given the difficulty of eliminating bias, single studies are unlikely to be definitive. Replication and careful synthesis of evidence across studies will be needed to support confident inferences about effectiveness. Exact replication of a natural experiment is unlikely, but partial replication is often possible and may be more informative. Consistent findings from studies using varying designs makes it less likely that common biases are present, and consistent findings across settings or populations increase confidence in the generalisability of causal inferences. For example, a number of studies in different countries have shown that legal restrictions on smoking in public places reduce hospital admissions for heart attacks. Although the size of the effect varies widely, as might be expected given variation in smoking rates and the extent of partial restrictions prior to outright bans, the balance of evidence suggests a real effect.[22]

Reporting

Transparent reporting of natural experimental studies is vital. Established guidelines such as STROBE[32] should be followed, with particular attention to: clearly identifying the approach as a study of a natural experiment; providing a clear description of the intervention and the assignment process; and explicitly stating the methods used to estimate impact. Procedures used to reduce bias should be discussed in a detailed and balanced way. Ideally, qualitative judgements about the risk of bias, and how well it has been dealt with, should be supplemented by a quantitative assessment.[33-34] If a study has used multiple methods, variation in the estimates should be highlighted. The context within which the intervention was implemented should be described as this may affect interpretation and help users assess the generalisability of the findings. Wherever possible, the results should be compared with those of other evaluations of similar interventions, paying attention to any associations between effect sizes and variations in evaluation methods and intervention design, content and context.

Conclusion

There are important areas of public health policy – such as suicide prevention, air pollution control, public smoking bans and alcohol taxation – where natural experimental studies have already contributed a convincing body of evidence. Such approaches are most readily applied where an intervention is implemented on a large scale, the effects are substantial and good population data on exposure and outcome are available. But they can also be used to detect more subtle effects where there is a suitable source of variation in exposure.

Even so, it would be unwise to assume that a particular policy or intervention could be evaluated as a natural experiment without very detailed consideration of the methodological challenges. Optimism about the use of a natural experimental approach should not be a pretext for discounting the option of conducting a planned experiment, where this would be possible and more robust.

Research effort should focus on addressing important and answerable questions, taking a pragmatic approach based on combinations of research methods and the explicit recognition and careful testing of assumptions. Priorities for the future are to build up experience of promising but lesser used methods, and to improve the infrastructure that enables opportunities presented by natural experiments to be seized, including good routine data from population surveys and administrative sources, good working relationships between researchers and policy makers, and flexible forms of research funding.

Supplementary Material

NIHMS54487-supplement-supporting_info.doc

What is already known.

Natural experimental approaches widen the range of interventions that can usefully be evaluated, but they are also more prone to bias than randomised controlled trials. It is important to understand when and how to use natural experiments and when planned experiments are preferable

What this study adds

The UK Medical Research Council has published new guidance on the use of natural experimental approaches to evaluate public health policies and other interventions that affect health.

Natural experimental approaches work best when the effects of the intervention are large and rapid, and good quality data on exposure and outcomes in a large population are available.

They can be also used to study more subtle effects, so long as a suitable source of variation in exposure can be found, but the design and analysis of such studies is more demanding.

Priorities for the future are to build up experience of promising but lesser used methods, and to improve the infrastructure that enables research opportunities presented by natural experiments to be seized.

Acknowledgments

Funding Preparation of this paper was supported by the MRC Population Health Sciences Research Network, and the MRC Methodology Research Panel.

Footnotes

The final published version of this article can be found at http://jech.bmj.com/content/66/12/1182.long doi: 10.1136/jech-2011-200375

Competing interests None declared.

Contributor Information

Peter Craig, MRC Population Health Sciences Research Network and Chief Scientist Office, Scottish Government Health Directorates.

Cyrus Cooper, MRC Lifecourse Epidemiology Unit, University of Southampton, UK.

David Gunnell, School of Social and Community Medicine, University of Bristol, UK.

Sally Haw, Centre for Public Health and Population Health Research, University of Stirling, UK.

Kenny Lawson, Institute of Health and Wellbeing, University of Glasgow, UK.

Sally Macintyre, MRClCSO Social and Public Health Sciences Unit, University of Glasgow, UK.

David Ogilvie, MRC Epidemiology Unit and UKCRC Centre for Diet and Activity Research (CEDAR), Cambridge, UK.

Mark Petticrew, Department of Social and Environmental Medicine, London School of Hygiene and Tropical Medicine, UK.

Barney Reeves, Clinical Trials and Evaluation Unit, University of Bristol, UK.

Matt Sutton, Health Methodology Research Group, School of Community-based Medicine, University of Manchester, UK.

Simon Thompson, Department of Public Health and Primary Care, University of Cambridge, UK.

Bibliography

  • 1.Wanless D. Securing good health for the whole population. HM Treasury; London: 2004. [Google Scholar]
  • 2.Foresight . Tackling Obesities: Future Choices. Department for Innovation, Universities and Skills; London: 2007. [Google Scholar]
  • 3.Department of Health . Healthy Weight, Healthy Lives: a research and surveillance plan for England. Department of Health; London: 2008. [Google Scholar]
  • 4.Kelly MP, Morhan A, Bonnefoy J, et al. Final report to World Health Organisation Commission on the Social Determinants of Health. National Institute for Health and Clinical Excellence, UK/Universidad de Desarollo; Chile: 2007. The social determinants of health: developing an evidence base for political action. [Google Scholar]
  • 5.Davey Smith G. Commentary: Behind the Broad Street pump: aetiology, epidemiology and prevention of cholera in mid 19th century Britain. Int J Epidemiol. 2002;31:920–32. doi: 10.1093/ije/31.5.920. [DOI] [PubMed] [Google Scholar]
  • 6.West SG. Alternatives to randomized experiments. Curr Dir Psych Sci. 2009;18:299–304. [Google Scholar]
  • 7.Craig P, Dieppe P, Macintyre S, et al. Developing and evaluating complex interventions: the new Medical Research Council guidance. British Medical Journal. 2008;337:a1655. doi: 10.1136/bmj.a1655. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Academy of Medical Sciences . Identifying the environmental causes of disease: how should we decide what to believe and when to take action. Academy of Medical Sciences; London: 2007. [Google Scholar]
  • 9.Meyer BD. Natural and quasi-experiments in economics. Journal of Business and Economic Studies. 1995;13:151–61. [Google Scholar]
  • 10.Shadish WR, Cook TD, Campbell DT. Experimental and quasi-experimental designs for generalised causal inference. Houghton Mifflin Company; Boston, Mass: 2002. [Google Scholar]
  • 11.West SG, Duan N, Pequegnat W, et al. Alternatives to the randomised controlled trial. Am J Public Health. 2008;98:1359–66. doi: 10.2105/AJPH.2007.124446. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Ogilvie D, Mitchell R, Mutrie N, et al. Evaluating health effects of transport interventions: methodologic case study. Am J Prev Med. 2006;31:118–26. doi: 10.1016/j.amepre.2006.03.030. [DOI] [PubMed] [Google Scholar]
  • 13.House of Commons Health Committee . Health Inequalities. Third Report of Session 2008-9. Volume 1: Report, together with formal minutes. The Stationery Office; London: 2009. [Google Scholar]
  • 14.Bonell CP, Hargreaves J, Cousens S, et al. Alternatives to randomisation in the evaluation of public health interventions. J Epidemiol Community Health. 2010 doi: 10.1136/jech.2008.082602. doi:10.1136/jech.2008.082602. [DOI] [PubMed] [Google Scholar]
  • 15.Melhuish E, Belsky J, Leyland AH, et al. Effects of fully established Sure Start Local Programmes on 3 year old children and their families living in England: a quasi-experimental observational study. Lancet. 2008;372:1641–7. doi: 10.1016/S0140-6736(08)61687-6. [DOI] [PubMed] [Google Scholar]
  • 16.Gunnell D, Fernando R, Hewagama M, et al. The impact of pesticide regulations on suicide in Sri Lanka. Int J Epidemiol. 2007;36:1235–42. doi: 10.1093/ije/dym164. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Kreitman N. The coal gas story. United Kingdom suicide rates, 1960-71. British Journal of Preventive Medicine. 1976;30:86–93. doi: 10.1136/jech.30.2.86. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Wheeler B, Gunnell D, Metcalfe C, et al. The population impact on incidence of suicide and non-fatal self harm of regulatory action against the use of selective serotonin reuptake inhibitors in under 18s in the United Kingdom: ecological study. BMJ. 2008;336:542–5. doi: 10.1136/bmj.39462.375613.BE. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Taylor S, Kingdom D, Jenkins R. How are nations trying to prevent suicide? An analysis of national suicide prevention strategies. Acta Psychiatr Scand. 1997;95:457–63. doi: 10.1111/j.1600-0447.1997.tb10132.x. [DOI] [PubMed] [Google Scholar]
  • 20.Deeks J, Dinnes J, D’Amico R, et al. Evaluating non-randomised intervention studies. Health Technol Assess. 2003;7 doi: 10.3310/hta7270. [DOI] [PubMed] [Google Scholar]
  • 21.Rubin DR. For objective causal inference, design trumps analysis. Annals of Applied Statistics. 2008;2:808–40. [Google Scholar]
  • 22.Mackay D, Pell J, Irfan M, et al. Meta-analysis of the influence of smoke-free legislation on acute coronary events. Heart. 2010 doi: 10.1136/hrt.2010.199026. (in press) [DOI] [PubMed] [Google Scholar]
  • 23.Pell J, Haw SJ, Cobbe S, et al. Smokefree legislation and hospitalisations for acute coronary syndrome. JAMA. 2008;359:482–91. doi: 10.1056/NEJMsa0706740. [DOI] [PubMed] [Google Scholar]
  • 24.Sims M, Maxwell R, Bauld L, et al. Short term impact of smoke-free legislation in England: retrospective analysis of hospital admissions for myocardial infarction. BMJ. 2010;340:c2161. doi: 10.1136/bmj.c2161. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Clancy L, Goodman P, Sinclair H, et al. Effect of air pollution control on death rates in Dublin, Ireland: an intervention study. Lancet. 2002;360:1210–14. doi: 10.1016/S0140-6736(02)11281-5. [DOI] [PubMed] [Google Scholar]
  • 26.Hedley A, Wong C, Thach T, et al. Cardiorespiratory and all-cause mortality after restrictions on sulphur content of fuel in Hong Kong: an intervention study. Lancet. 2002;360:1646–52. doi: 10.1016/s0140-6736(02)11612-6. [DOI] [PubMed] [Google Scholar]
  • 27.Campostrini S, Holtzman D, McQueen D, et al. Evaluating the effectiveness of health promotion policy: changes in the law on drinking and driving in California. Health Promot Int. 2006;21:130–5. doi: 10.1093/heapro/dak005. [DOI] [PubMed] [Google Scholar]
  • 28.Ludwig J, Miller DL. Does Head Start improve children’s life chances? Evidence from a regression discontinuity design. Q J Econ. 2007:159–208. [Google Scholar]
  • 29.Hernan M, Robins JA. Instruments for causal inference. An epidemiologist’s dream?. Epidemiology. 2006;17:360–72. doi: 10.1097/01.ede.0000222409.00878.37. [DOI] [PubMed] [Google Scholar]
  • 30.Lim SS, Dandona L, Hoisington JA, et al. India’s Janani Suraksha Yonana: a conditional cash transfer scheme to increase births in health facilities: an impact evaluation. Lancet. 2010;375:2009–23. doi: 10.1016/S0140-6736(10)60744-1. [DOI] [PubMed] [Google Scholar]
  • 31.Dusheiko M, Gravelle H, Jacobs R, et al. Discussion papers in economics. University of York; York: 2003. The effect of budgets on doctor behaviour: evidence from a natural experiment. [Google Scholar]
  • 32.von Elm E, Altman D, Egger M, et al. Strengthening the reporting of observational studies in epidemiology (STROBE) statement: guidelines for reporting observational studies. BMJ. 2007;335:335–8. doi: 10.1136/bmj.39335.541782.AD. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Fox MP. Creating a demand for bias analysis in epidemiological research. J Epidemiol Community Health. 2009;63:91. doi: 10.1136/jech.2008.082420. [DOI] [PubMed] [Google Scholar]
  • 34.Turner R, Spiegelhalter D, Smith G, et al. Bias modelling in evidence synthesis. J R Stat Soc Ser A Stat Soc. 2009;172:21–47. doi: 10.1111/j.1467-985X.2008.00547.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.D’Agostino R. Propensity score methods for bias reduction in the comparison of a treatment to a non-randomized control group. Stat Med. 1998;17:2265–81. doi: 10.1002/(sici)1097-0258(19981015)17:19<2265::aid-sim918>3.0.co;2-b. [DOI] [PubMed] [Google Scholar]
  • 36.Angrist J, Pischke J-S. Mostly harmless econometrics: an empiricist’s companion. Princeton University Press; Princeton: 2009. [Google Scholar]
  • 37.Brookhart MA, Rassen JA, Schneeweiss S. Instrumental variable methods in comparative safety and effectiveness research. Pharmacoepidemiol Drug Saf. 2010;19:537–54. doi: 10.1002/pds.1908. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.McClellan M, McNeil BJ, Newhouse JP. Does more intensive treatment of acute myocardial infarction in the elderly reduce mortality? Analysis using instrumental variables. JAMA. 1994;272:859–66. [PubMed] [Google Scholar]
  • 39.Stukel TA, Fisher ES, Wennberg DE, et al. Analysis of observational studies in the presence of treatment selection bias. Effects of invasive cardiac management on AMI survival using propensity score and instrumental variable methods. JAMA. 2007;297:278–85. doi: 10.1001/jama.297.3.278. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Herttua K, Makela P, Martikainen P. Changes in alcohol-related mortality and its socioeconomic differences after a large reduction in alcohol prices: a natural experiment based on register data. Am J Epidemiol. 2008;168:1110–8. doi: 10.1093/aje/kwn216. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Herttua K, Makela P, Martikainen P. An evaluation of the impact of a large reduction in alcohol prices on alcohol-related and all-cause mortality: time series analysis of a population-based natural experiment. Int J Epidemiol. 2011;40:441–54. doi: 10.1093/ije/dyp336. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

NIHMS54487-supplement-supporting_info.doc

RESOURCES