Abstract
Natural experimental studies are often recommended as a way of understanding the health impact of policies and other large scale interventions. Although they have certain advantages over planned experiments, and may be the only option when it is impossible to manipulate exposure to the intervention, natural experimental studies are more susceptible to bias. This paper introduces new guidance from the Medical Research Council to help researchers and users, funders and publishers of research evidence make the best use of natural experimental approaches to evaluating population health interventions. The guidance emphasises that natural experiments can provide convincing evidence of impact even when effects are small or take time to appear. However, a good understanding is needed of the process determining exposure to the intervention, and careful choice and combination of methods, testing of assumptions and transparent reporting is vital. More could be learnt from natural experiments in future as experience of promising but lesser used methods accumulates.
Introduction
Natural experimental studies are often recommended as a way of understanding the impact of population-level policies on health outcomes or health inequalities.[1-4] Within epidemiology there is a long tradition, stretching back to John Snow in the mid nineteenth century,[5] of using major external shocks such as epidemics, famines or economic crises to study the causes of disease. A difficulty in applying similar methods to the evaluation of population health policies and interventions, such as a ‘fat tax’ or a legal minimum price per unit of alcohol, is that very often the change in exposure is much less extreme, and its effects may be subtle or take time to emerge. Although they have certain advantages over planned experiments, for example by enabling effects to be studied in whole populations,[6] and may be the only option when it is impossible to manipulate exposure to the intervention, natural experimental studies are more susceptible to bias and confounding. It is therefore important to be able to distinguish situations in which natural experimental approaches are likely to be informative from those in which some form of fully experimental method such as a randomised controlled trial (RCT) is needed, and from those in which the research questions are genuinely intractable.
The Medical Research Council has recently published guidance to help researchers and users, funders and publishers of research evidence make the best use of natural experimental approaches to evaluate population health interventions (www.mrc.ac.uk/naturalexperimentsguidance). Following the model of the MRC complex interventions guidance,[7] it was written by a multidisciplinary team with experience of evaluation using a wide range of research designs. The ideas were developed and tested in two specially convened workshops of population health researchers. Drafts were reviewed by workshop delegates and by the MRC’s Methodology Research Panel. The guidance is meant to help researchers plan and design evaluations of public health interventions, journal editors and reviewers to assess the quality of studies that use observational data to evaluate interventions, and policy-makers and others to recognise the strengths and limitations of a natural experimental approach. In this paper we summarise the main messages of the guidance.
What are natural experiments?
The term ‘natural experiment’ lacks an exact definition, and many variants are found in the literature.[8-10] The common thread in most definitions is that exposure to the event or intervention of interest has not been manipulated by the researcher. Outside an RCT it is rare for variation in exposure to an intervention to be random, so special care is needed in the design, reporting and interpretation of evidence from natural experimental studies, and causal inferences must be drawn with care.
Why are natural experiments important?
Alternatives to RCTs have been advocated by policy-makers and researchers interested in evaluating population-level environmental and non-health sector interventions[11] and their impact on health inequalities.[4] Such interventions may be intrinsically difficult to manipulate experimentally – as in the case of national legislation to improve air quality, or major changes in transport infrastructure[12] – or implemented in ways that make a planned experiment difficult or impossible, for example with short timescales or extreme variability in implementation.[13] It may also be unethical to manipulate exposure in order to study effects on health if an intervention has other known benefits, if it has been shown to be effective in other settings, or if its main purpose is to achieve non-health outcomes.[14] Even if such ethical and practical restrictions are absent, an RCT may still be politically unwelcome.[15]
Natural experimental approaches are important because they widen the range of interventions that can usefully be evaluated beyond those that are amenable to planned experimentation. For example, suicide is rare in the general population, occurring at a rate of about 1/10,000 per annum. Even in high risk populations, such as people treated with antidepressants, the annual incidence is only around 1/1000. Clinical trials would have to be enormous to have adequate power to detect even large preventive effects, but natural experiments have been used effectively to assess the impact of measures to restrict access to commonly used means of suicide[16-18] and inform the content of suicide prevention strategies in the UK and worldwide.[19]
These and other studies in which a natural experimental approach has produced clear cut evidence of health impacts are summarised in Supplemental Table 1. They illustrate the diversity of interventions that have been evaluated as natural experiments, and the wide range of methods that have been applied. Many of the studies have benefited from the availability of high quality, routinely collected data on exposures, potential confounders and outcomes and substantial, rapid changes in exposure across a whole population, which reduces the risk of selective exposure or confounding by secular trends and increases the confidence with which changes in outcomes can be attributed to the interventions. However, it is misleading to assume that whenever a planned experiment is impossible, there is a natural experimental study waiting to happen. Some but not all of the ‘multitude of promising initiatives’[1] are likely to yield good natural experimental studies. Care, ingenuity and a watchful eye for good opportunities are needed to realise their potential.
When should natural experiments be used?
The case for adopting a natural experimental approach is strongest when: there is a reasonable expectation that the intervention will have a significant health impact, but scientific uncertainty remains about the size or nature of the effects; an RCT would be impractical or unethical; and the intervention or the principles behind it have the potential for replication, scalability or generalisability.
In practice, natural experiments are highly variable, and researchers face difficult choices about when to adopt a natural experimental approach and how best to exploit the opportunities that do occur. The value of a given natural experiment for research depends on a range of factors including the size of the population affected, the size and timing of likely impacts, the processes generating variation in exposure, and the practicalities of data gathering. Quantitative natural experimental studies should only be attempted when exposed and unexposed populations (or groups subject to varying levels of exposure) can be compared, using samples large enough to detect the expected effects, and when accurate data can be obtained on exposures, outcomes and potential confounders. Resources should only be committed when the economic and scientific rationale for a study can be clearly articulated.
Design, analysis and reporting of natural experiments
Planned and natural experiments face some of the same threats to validity, such as loss to follow-up and inaccurate assessment of exposure and outcomes. The key difference is that RCTs have a very general and (if used properly) effective method of minimising the bias that results from selective exposure to the intervention, i.e. the tendency for exposure to vary according to characteristics of participants that are also associated with outcomes. In the case of non-randomised studies, there is no such general solution to the pervasive problem of confounding.[20] Instead there is a range of partial solutions which can be used in some, often very restricted, circumstances but not others. Understanding the process that produces the variation in exposure (often referred to as the ‘assignment process’ even when there is no deliberate manipulation of individuals’ exposure[21]) is therefore critical to the design of natural experimental studies.[9]
Design
A study protocol should be developed, and ideally published, whatever design is adopted. Good practice in the conduct of observational studies, such as prior specification of hypotheses, clear definitions of target populations, explicit sampling criteria, and valid and reliable measures of exposures and outcomes, should apply equally to natural experimental studies.
All natural experimental studies require a comparison of exposed and unexposed groups (or groups with varying levels of exposure) to identify the effect of the intervention. The examples of suicide prevention,[16-17] indoor smoking bans[22-24] and air pollution control[25-26] show that simple designs can provide convincing evidence if a whole population is abruptly exposed to an intervention, and if the effects are large, rapidly follow exposure and can be measured accurately at population level using routinely available data. This combination of circumstances is rare, and more complex designs are usually required.
Natural experiments can also be used to study more subtle effects, so long as a suitable source of variation in exposure can be found, but the design and analysis of such studies is more challenging. In any case, what is often required is an estimate of effect size, and a large observed effect may incorporate a large element of bias due to selective exposure to the intervention. Whatever the expected effect size, care should be taken to minimise bias in the design and analysis of natural experiments.
Design elements that can strengthen causal inferences from natural experimental studies include the use of multiple pre/post measures to control for secular changes, as in an interrupted time series design;[27] multiple exposed/unexposed groups that differ according to some variable that may affect exposure and outcome to assess whether selection on that variable is likely to be an important source of bias;[9] accurate measurement of multiple potential confounders and combinations of methods to address different sources of bias. In a study that exemplifies many of the features of a rigorous approach to identifying relatively small effects, Ludwig and Miller[28] used variation in access to support for obtaining Headstart funding to model exposure, and compared a variety of outcomes among children who were above or below the age cutoff for access to Headstart services (Box 2; Table 1).
Box 2. Selection on unobservables.
Difference in differences
This method compares change over time in exposed and unexposed groups.[36] The differencing procedure controls for unobserved individual differences, and for common trends. It assumes that the unobserved characteristics are fixed, and that the outcomes in each group would change in the same way in the absence of the intervention, so is vulnerable to changes in the composition of the groups and to external influences that differentially affect the exposed and unexposed groups.
Instrumental variables
An instrumental variable (IV) is a factor, such as treatment assignment in a well-designed RCT, which is associated with outcomes only via its association with exposure to the intervention and is independent of other factors associated with exposure. IVs have been used to identify the impact of treatment from routine data.[37-39] In these studies, variables such as distance from a specialised centre have been used to evaluate novel treatments, the assumption being that patients living close to a specialised centre are more likely to receive the novel treatment, but are otherwise similar to other patients.
Regression discontinuity designs
This approach exploits a step change or ‘cutoff’ in a continuous variable used to assign treatment or otherwise determine exposure to an intervention. The assumption is that units (individuals, areas, etc.) just below and just above this threshold will otherwise be similar in terms of characteristics that may influence outcomes, so that an estimate of treatment effect can be obtained by comparing regression slopes on either side of the cutoff. When the Headstart programme to improve the health of disadvantaged children was first implemented in the US, help with applying for funding was targeted on the 300 poorest counties, and a higher proportion of those counties received funding. Ludwig and Miller[28] compared regressions of child mortality on poverty for counties either side of the cutoff, and found lower than expected mortality in those that qualified for assistance.
Table 1.
Population | Intervention | Comparator | Outcome | Analysis method | Results | Reference |
---|---|---|---|---|---|---|
Sri Lanka - whole population |
Legal restriction on pesticide imports |
Suicide rates pre-ban; method specific suicide rates; rates of non- fatal self- poisoning |
Overall suicide mortality and mortality mortality from suicide by self- poisoning with pesticide |
Graphical analysis of trends |
Import bans reduced method-specific and overall suicide mortality |
Gunnell et al 2007[16] |
UK - 12-17 year olds |
Restriction on prescribing SSRIs to people aged <18 |
Suicidal behaviour pre- restriction |
Hospitalisations for self-harm; suicide mortality |
Joinpoint regression |
Restriction on prescribing of SSRIs was not associated with changes in suicidal behaviour |
Wheeler et al 2008[18] |
Finland - men and women aged >15 |
Reduction in alcohol taxes |
Alcohol related mortality before the tax change |
Alcohol-related mortality |
Poisson regression to obtain relative rates; timeseries analysis |
Alcohol-related mortality increased, especially among the unemployed |
Herttua et al 2008,[40] Herttua et al 2011[41] |
Hong Kong - whole population |
Legislation to restrict sulphur content of fuel |
Mortality pre- restriction, overall and by district |
All-cause and cause specific mortality |
Poisson regression of change in seasonally adjusted death rates |
Cardiovascular, respiratory and overall mortality fell post-restriction; decline greater in districts with larger falls in SO2 concentrations |
Hedley et al 2002[26] |
Dublin – whole population |
Ban on coal sales | Mortality pre- ban, and in the rest of Ireland |
Non-trauma, respiratory and cardiovascular mortality |
Interrupted time-series |
Non-trauma, respiratory and cardiovascular death rates fell post-ban |
Clancy et al 2002[25] |
Scotland – patients admitted to 9 hospitals |
Legislative ban on smoking in public places |
Hospitalisations pre-ban and in England |
Hospitalisations for acute coronary syndrome |
Comparison of numbers of admissions pre and post ban |
Admissions for acute coronary syndrome fell among both smokers and nonsmokers post-ban |
Pell et al 2008[23] |
England – patients aged >17 |
Legislative ban on smoking in public places |
Hospitalisations pre-ban |
Emergency admissions for myocardial infarction |
Interrupted time-series |
Small but significant fall in emergency admissions in the first year post-ban |
Sims et al 2010[24] |
India – pregnant women |
Cash incentives to use a health facility to give birth |
Districts with low rates of take up; births to women not receiving payments |
Use of health facilities; infant and maternal mortality |
Matched and unmatched comparisons of recipient and non-recipient births; difference-indifferences analysis of district level usage and mortality rates |
Higher rates of take-up of the incentives were associated with higher proportions of births within health care facilities; use of health care facilities to give birth was associated with fewer perinatal and neonatal deaths. There was a non- significant reduction in maternal deaths |
Lim et al 2010[30] |
England – general practitioners |
Abolition of GP fundholding |
Non- fundholding practices; pre- abolition admission rates |
Referral for elective and emergency admissions |
Difference-indifference analysis of referrals from fundholders and non-fundholders |
Fundholders had lower rates of elective referral while fundholding was in operation and their rates of referral increased more than those of non-fundholders following abolition. There was no difference in emergency admissions pre or post abolition. |
Dusheiko et al 2003[31] |
USA – low income families with children aged 3-5 |
Headstart - help with parenting, nutrition, health and social services and schooling |
US counties with poverty levels above the cutoff used to allocate help with accessing Headstart funding |
Mortality from causes of death amenable to Headstart services |
Regression discontinuity design, comparing regressions of mortality on poverty for counties above and below the cutoff |
For Headstart-related causes, there was a discontinuity in mortality rates at the cutoff, but no difference in deaths from other causes or among children too old to qualify for Headstart services |
Ludwig and Miller 2007[28] |
USA – patients admitted to hospital with acute myocardial infarction |
Invasive cardiac treatment (catherisation followed by revascularisation) |
Non-invasive cardiac treatment |
Long term (i.e. 7-year) mortality |
Instrumental variable analysis using regional catherisation rates as the instrument |
Cardiac catherisation was associated with lower mortality; instrumental variable analyses produced smaller estimates of effect than analyses (multivariate risk adjustment, propensity score-based methods) that only adjust for observed prognostic factors |
Stukel et al 2007[39] |
Analysis
The defining feature of a natural experiment is that manipulating exposure to the intervention is impossible. There are a few examples where assignment is by a ‘real life’ lottery, but selection is the rule and a range of methods is available for dealing with the resulting bias.
Where the factors that determine exposure can be measured accurately and comprehensively, matching, regression and propensity scores can be used to reduce confounding (Box 1). Bias will remain if there are unobserved or imperfectly measured factors that influence both exposure and outcomes. Given the difficulty of measuring accurately all of the characteristics associated with exposure to an intervention, methods such as difference in differences, instrumental variables and regression discontinuity designs that deal with unobserved factors are a potentially valuable advance on those that only deal with observed factors (Box 2).
Box 1. Selection on observables.
Matching
This involves finding unexposed individuals (or clusters of individuals) which are similar to those receiving the intervention, and comparing outcomes in the two groups.
Regression adjustment
Measured characteristics that differ between those receiving the intervention and others can be taken into account in multiple regression analyses.
Propensity scores
The likelihood of being exposed to an intervention given a set of covariates can be estimated by logistic regression[35] and used to match exposed with unexposed cases, or for covariate adjustment.
In practice, none of these approaches provides a comprehensive solution to the central problem of selective exposure to the intervention.[20] Methods of controlling for observed factors associated with exposure are vulnerable to selection on unobservables. Methods for dealing with selection on unobservables require strong and untestable assumptions[29] and their use is restricted by the often very limited availability of variables that can be used to model exposure. They are therefore best used in conjunction with additional tests for the plausibility of any causal inferences.
Combining methods that address different sources of bias and comparing the results is one such approach and there are several examples in Table 1. In their evaluation of a conditional cash transfer scheme to encourage women to use health facilities to give birth, Lim et al[30] combined methods for dealing with selection on both observable and non-observable characteristics. Another useful technique is to analyse outcomes that are not expected to change. Dusheiko et al[31] used trends in emergency admissions as a non-equivalent dependent variable to test whether changes in elective admissions could plausibly be attributed to GP fundholding, while Ludwig and Miller[28] compared mortality from causes that were likely or unlikely to respond to Headstart services.
Given the difficulty of eliminating bias, single studies are unlikely to be definitive. Replication and careful synthesis of evidence across studies will be needed to support confident inferences about effectiveness. Exact replication of a natural experiment is unlikely, but partial replication is often possible and may be more informative. Consistent findings from studies using varying designs makes it less likely that common biases are present, and consistent findings across settings or populations increase confidence in the generalisability of causal inferences. For example, a number of studies in different countries have shown that legal restrictions on smoking in public places reduce hospital admissions for heart attacks. Although the size of the effect varies widely, as might be expected given variation in smoking rates and the extent of partial restrictions prior to outright bans, the balance of evidence suggests a real effect.[22]
Reporting
Transparent reporting of natural experimental studies is vital. Established guidelines such as STROBE[32] should be followed, with particular attention to: clearly identifying the approach as a study of a natural experiment; providing a clear description of the intervention and the assignment process; and explicitly stating the methods used to estimate impact. Procedures used to reduce bias should be discussed in a detailed and balanced way. Ideally, qualitative judgements about the risk of bias, and how well it has been dealt with, should be supplemented by a quantitative assessment.[33-34] If a study has used multiple methods, variation in the estimates should be highlighted. The context within which the intervention was implemented should be described as this may affect interpretation and help users assess the generalisability of the findings. Wherever possible, the results should be compared with those of other evaluations of similar interventions, paying attention to any associations between effect sizes and variations in evaluation methods and intervention design, content and context.
Conclusion
There are important areas of public health policy – such as suicide prevention, air pollution control, public smoking bans and alcohol taxation – where natural experimental studies have already contributed a convincing body of evidence. Such approaches are most readily applied where an intervention is implemented on a large scale, the effects are substantial and good population data on exposure and outcome are available. But they can also be used to detect more subtle effects where there is a suitable source of variation in exposure.
Even so, it would be unwise to assume that a particular policy or intervention could be evaluated as a natural experiment without very detailed consideration of the methodological challenges. Optimism about the use of a natural experimental approach should not be a pretext for discounting the option of conducting a planned experiment, where this would be possible and more robust.
Research effort should focus on addressing important and answerable questions, taking a pragmatic approach based on combinations of research methods and the explicit recognition and careful testing of assumptions. Priorities for the future are to build up experience of promising but lesser used methods, and to improve the infrastructure that enables opportunities presented by natural experiments to be seized, including good routine data from population surveys and administrative sources, good working relationships between researchers and policy makers, and flexible forms of research funding.
Supplementary Material
What is already known.
Natural experimental approaches widen the range of interventions that can usefully be evaluated, but they are also more prone to bias than randomised controlled trials. It is important to understand when and how to use natural experiments and when planned experiments are preferable
What this study adds
The UK Medical Research Council has published new guidance on the use of natural experimental approaches to evaluate public health policies and other interventions that affect health.
Natural experimental approaches work best when the effects of the intervention are large and rapid, and good quality data on exposure and outcomes in a large population are available.
They can be also used to study more subtle effects, so long as a suitable source of variation in exposure can be found, but the design and analysis of such studies is more demanding.
Priorities for the future are to build up experience of promising but lesser used methods, and to improve the infrastructure that enables research opportunities presented by natural experiments to be seized.
Acknowledgments
Funding Preparation of this paper was supported by the MRC Population Health Sciences Research Network, and the MRC Methodology Research Panel.
Footnotes
The final published version of this article can be found at http://jech.bmj.com/content/66/12/1182.long doi: 10.1136/jech-2011-200375
Competing interests None declared.
Contributor Information
Peter Craig, MRC Population Health Sciences Research Network and Chief Scientist Office, Scottish Government Health Directorates.
Cyrus Cooper, MRC Lifecourse Epidemiology Unit, University of Southampton, UK.
David Gunnell, School of Social and Community Medicine, University of Bristol, UK.
Sally Haw, Centre for Public Health and Population Health Research, University of Stirling, UK.
Kenny Lawson, Institute of Health and Wellbeing, University of Glasgow, UK.
Sally Macintyre, MRClCSO Social and Public Health Sciences Unit, University of Glasgow, UK.
David Ogilvie, MRC Epidemiology Unit and UKCRC Centre for Diet and Activity Research (CEDAR), Cambridge, UK.
Mark Petticrew, Department of Social and Environmental Medicine, London School of Hygiene and Tropical Medicine, UK.
Barney Reeves, Clinical Trials and Evaluation Unit, University of Bristol, UK.
Matt Sutton, Health Methodology Research Group, School of Community-based Medicine, University of Manchester, UK.
Simon Thompson, Department of Public Health and Primary Care, University of Cambridge, UK.
Bibliography
- 1.Wanless D. Securing good health for the whole population. HM Treasury; London: 2004. [Google Scholar]
- 2.Foresight . Tackling Obesities: Future Choices. Department for Innovation, Universities and Skills; London: 2007. [Google Scholar]
- 3.Department of Health . Healthy Weight, Healthy Lives: a research and surveillance plan for England. Department of Health; London: 2008. [Google Scholar]
- 4.Kelly MP, Morhan A, Bonnefoy J, et al. Final report to World Health Organisation Commission on the Social Determinants of Health. National Institute for Health and Clinical Excellence, UK/Universidad de Desarollo; Chile: 2007. The social determinants of health: developing an evidence base for political action. [Google Scholar]
- 5.Davey Smith G. Commentary: Behind the Broad Street pump: aetiology, epidemiology and prevention of cholera in mid 19th century Britain. Int J Epidemiol. 2002;31:920–32. doi: 10.1093/ije/31.5.920. [DOI] [PubMed] [Google Scholar]
- 6.West SG. Alternatives to randomized experiments. Curr Dir Psych Sci. 2009;18:299–304. [Google Scholar]
- 7.Craig P, Dieppe P, Macintyre S, et al. Developing and evaluating complex interventions: the new Medical Research Council guidance. British Medical Journal. 2008;337:a1655. doi: 10.1136/bmj.a1655. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Academy of Medical Sciences . Identifying the environmental causes of disease: how should we decide what to believe and when to take action. Academy of Medical Sciences; London: 2007. [Google Scholar]
- 9.Meyer BD. Natural and quasi-experiments in economics. Journal of Business and Economic Studies. 1995;13:151–61. [Google Scholar]
- 10.Shadish WR, Cook TD, Campbell DT. Experimental and quasi-experimental designs for generalised causal inference. Houghton Mifflin Company; Boston, Mass: 2002. [Google Scholar]
- 11.West SG, Duan N, Pequegnat W, et al. Alternatives to the randomised controlled trial. Am J Public Health. 2008;98:1359–66. doi: 10.2105/AJPH.2007.124446. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Ogilvie D, Mitchell R, Mutrie N, et al. Evaluating health effects of transport interventions: methodologic case study. Am J Prev Med. 2006;31:118–26. doi: 10.1016/j.amepre.2006.03.030. [DOI] [PubMed] [Google Scholar]
- 13.House of Commons Health Committee . Health Inequalities. Third Report of Session 2008-9. Volume 1: Report, together with formal minutes. The Stationery Office; London: 2009. [Google Scholar]
- 14.Bonell CP, Hargreaves J, Cousens S, et al. Alternatives to randomisation in the evaluation of public health interventions. J Epidemiol Community Health. 2010 doi: 10.1136/jech.2008.082602. doi:10.1136/jech.2008.082602. [DOI] [PubMed] [Google Scholar]
- 15.Melhuish E, Belsky J, Leyland AH, et al. Effects of fully established Sure Start Local Programmes on 3 year old children and their families living in England: a quasi-experimental observational study. Lancet. 2008;372:1641–7. doi: 10.1016/S0140-6736(08)61687-6. [DOI] [PubMed] [Google Scholar]
- 16.Gunnell D, Fernando R, Hewagama M, et al. The impact of pesticide regulations on suicide in Sri Lanka. Int J Epidemiol. 2007;36:1235–42. doi: 10.1093/ije/dym164. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Kreitman N. The coal gas story. United Kingdom suicide rates, 1960-71. British Journal of Preventive Medicine. 1976;30:86–93. doi: 10.1136/jech.30.2.86. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Wheeler B, Gunnell D, Metcalfe C, et al. The population impact on incidence of suicide and non-fatal self harm of regulatory action against the use of selective serotonin reuptake inhibitors in under 18s in the United Kingdom: ecological study. BMJ. 2008;336:542–5. doi: 10.1136/bmj.39462.375613.BE. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Taylor S, Kingdom D, Jenkins R. How are nations trying to prevent suicide? An analysis of national suicide prevention strategies. Acta Psychiatr Scand. 1997;95:457–63. doi: 10.1111/j.1600-0447.1997.tb10132.x. [DOI] [PubMed] [Google Scholar]
- 20.Deeks J, Dinnes J, D’Amico R, et al. Evaluating non-randomised intervention studies. Health Technol Assess. 2003;7 doi: 10.3310/hta7270. [DOI] [PubMed] [Google Scholar]
- 21.Rubin DR. For objective causal inference, design trumps analysis. Annals of Applied Statistics. 2008;2:808–40. [Google Scholar]
- 22.Mackay D, Pell J, Irfan M, et al. Meta-analysis of the influence of smoke-free legislation on acute coronary events. Heart. 2010 doi: 10.1136/hrt.2010.199026. (in press) [DOI] [PubMed] [Google Scholar]
- 23.Pell J, Haw SJ, Cobbe S, et al. Smokefree legislation and hospitalisations for acute coronary syndrome. JAMA. 2008;359:482–91. doi: 10.1056/NEJMsa0706740. [DOI] [PubMed] [Google Scholar]
- 24.Sims M, Maxwell R, Bauld L, et al. Short term impact of smoke-free legislation in England: retrospective analysis of hospital admissions for myocardial infarction. BMJ. 2010;340:c2161. doi: 10.1136/bmj.c2161. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Clancy L, Goodman P, Sinclair H, et al. Effect of air pollution control on death rates in Dublin, Ireland: an intervention study. Lancet. 2002;360:1210–14. doi: 10.1016/S0140-6736(02)11281-5. [DOI] [PubMed] [Google Scholar]
- 26.Hedley A, Wong C, Thach T, et al. Cardiorespiratory and all-cause mortality after restrictions on sulphur content of fuel in Hong Kong: an intervention study. Lancet. 2002;360:1646–52. doi: 10.1016/s0140-6736(02)11612-6. [DOI] [PubMed] [Google Scholar]
- 27.Campostrini S, Holtzman D, McQueen D, et al. Evaluating the effectiveness of health promotion policy: changes in the law on drinking and driving in California. Health Promot Int. 2006;21:130–5. doi: 10.1093/heapro/dak005. [DOI] [PubMed] [Google Scholar]
- 28.Ludwig J, Miller DL. Does Head Start improve children’s life chances? Evidence from a regression discontinuity design. Q J Econ. 2007:159–208. [Google Scholar]
- 29.Hernan M, Robins JA. Instruments for causal inference. An epidemiologist’s dream?. Epidemiology. 2006;17:360–72. doi: 10.1097/01.ede.0000222409.00878.37. [DOI] [PubMed] [Google Scholar]
- 30.Lim SS, Dandona L, Hoisington JA, et al. India’s Janani Suraksha Yonana: a conditional cash transfer scheme to increase births in health facilities: an impact evaluation. Lancet. 2010;375:2009–23. doi: 10.1016/S0140-6736(10)60744-1. [DOI] [PubMed] [Google Scholar]
- 31.Dusheiko M, Gravelle H, Jacobs R, et al. Discussion papers in economics. University of York; York: 2003. The effect of budgets on doctor behaviour: evidence from a natural experiment. [Google Scholar]
- 32.von Elm E, Altman D, Egger M, et al. Strengthening the reporting of observational studies in epidemiology (STROBE) statement: guidelines for reporting observational studies. BMJ. 2007;335:335–8. doi: 10.1136/bmj.39335.541782.AD. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Fox MP. Creating a demand for bias analysis in epidemiological research. J Epidemiol Community Health. 2009;63:91. doi: 10.1136/jech.2008.082420. [DOI] [PubMed] [Google Scholar]
- 34.Turner R, Spiegelhalter D, Smith G, et al. Bias modelling in evidence synthesis. J R Stat Soc Ser A Stat Soc. 2009;172:21–47. doi: 10.1111/j.1467-985X.2008.00547.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.D’Agostino R. Propensity score methods for bias reduction in the comparison of a treatment to a non-randomized control group. Stat Med. 1998;17:2265–81. doi: 10.1002/(sici)1097-0258(19981015)17:19<2265::aid-sim918>3.0.co;2-b. [DOI] [PubMed] [Google Scholar]
- 36.Angrist J, Pischke J-S. Mostly harmless econometrics: an empiricist’s companion. Princeton University Press; Princeton: 2009. [Google Scholar]
- 37.Brookhart MA, Rassen JA, Schneeweiss S. Instrumental variable methods in comparative safety and effectiveness research. Pharmacoepidemiol Drug Saf. 2010;19:537–54. doi: 10.1002/pds.1908. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.McClellan M, McNeil BJ, Newhouse JP. Does more intensive treatment of acute myocardial infarction in the elderly reduce mortality? Analysis using instrumental variables. JAMA. 1994;272:859–66. [PubMed] [Google Scholar]
- 39.Stukel TA, Fisher ES, Wennberg DE, et al. Analysis of observational studies in the presence of treatment selection bias. Effects of invasive cardiac management on AMI survival using propensity score and instrumental variable methods. JAMA. 2007;297:278–85. doi: 10.1001/jama.297.3.278. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Herttua K, Makela P, Martikainen P. Changes in alcohol-related mortality and its socioeconomic differences after a large reduction in alcohol prices: a natural experiment based on register data. Am J Epidemiol. 2008;168:1110–8. doi: 10.1093/aje/kwn216. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Herttua K, Makela P, Martikainen P. An evaluation of the impact of a large reduction in alcohol prices on alcohol-related and all-cause mortality: time series analysis of a population-based natural experiment. Int J Epidemiol. 2011;40:441–54. doi: 10.1093/ije/dyp336. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.