Abstract
Objectives
Indirect comparisons via a common comparator (anchored comparisons) are commonly used in Health Technology Assessment. However, common comparators may not be available, or the comparison may be biased due to differences in effect modifiers between the included studies. Recently proposed population adjustment methods aim to adjust for differences between study populations in the situation where individual patient data are available from at least one study, but not all studies. They can also be used when there is no common comparator or for single-arm studies (unanchored comparisons). We aim to characterise the use of population adjustment methods in technology appraisals (TAs) submitted to the UK National Institute for Health and Care Excellence (NICE).
Methods
We reviewed NICE TAs published between 01/01/2010 and 20/04/2018.
Results
Population adjustment methods were used in 7% (18/268) of TAs. Most applications used unanchored comparisons (89%, 16/18), and were in oncology (83%, 15/18). Methods used included Matching-Adjusted Indirect Comparisons (89%, 16/18) and Simulated Treatment Comparisons (17%, 3/18). Covariates were included based on: availability, expert opinion, effective sample size, statistical significance, or cross validation. Larger treatment networks were commonplace (56%, 10/18), but current methods cannot account for this. Appraisal committees received results of population-adjusted analyses with caution and typically looked for greater cost effectiveness to minimise decision risk.
Conclusions
Population adjustment methods are becoming increasingly common in NICE TAs, though their impact on decisions has been limited to date. Further research is needed to improve upon current methods, and to investigate their properties in simulation studies.
Keywords: network meta-analysis, technology assessment, effect modifier, bias, comparative effectiveness research
Introduction
Health technology assessments and appraisals require reliable estimates of relative treatment effects to inform reimbursement decisions. When head-to-head evidence is not available but the two treatments of interest have each been studied against a common comparator (e.g. placebo or standard care), a standard indirect comparison may be performed using published aggregate data from each study (1). With larger numbers of treatments and studies a network meta-analysis is the standard approach, of which indirect comparison is a simple special case (2, 3). Standard indirect comparisons and network meta-analyses assume that the distributions of any effect modifying variables are similar in each study population and the decision target population, so that relative effects are constant across populations (the constancy of relative effects assumption). These methods are widely used in health technology appraisals, such as those undertaken by the National Institute for Health and Care Excellence (NICE) in the United Kingdom. As part of the NICE technology appraisal process, a company submits evidence on the clinical effectiveness of their treatment compared to other relevant treatments, frequently informed by indirect comparisons or network meta-analysis. The submitting company will have individual patient data (IPD) available from their own trial or trials, but very often only published aggregate data (AgD) from those of their competitors. More recently, methods have been proposed which use the available IPD to adjust for differences between IPD and AgD study populations. These include Matching-Adjusted Indirect Comparison (MAIC) (4) and Simulated Treatment Comparison (STC) (5), both first introduced in 2010, which use re-weighting or regression respectively to adjust the IPD estimates to the AgD study population. MAIC is a re-weighting method, where weights are derived for the individuals in the IPD study such that the moments of the weighted covariate distribution match the covariate summaries reported in the AgD study (typically using mean and variance for continuous covariates, and proportions for categorical covariates). The weights are then used to obtain predicted outcomes in the AgD study population, for example by taking weighted mean outcomes on each treatment. STC is a regression adjustment method, where a regression model is fitted in the IPD study. This model is then used to predict average outcomes in the AgD study population. Whichever method is used, once the predicted outcomes are obtained for the AgD study population these are compared with the outcomes reported by the AgD study. The development and use of these methods is motivated by one of two reasons: either i) there is evidence for effect modification, and these variables are distributed differently in each study population; or ii) there is no common comparator or the relevant studies are single arm, and so adjustment is required for all prognostic and effect modifying variables.
Phillippo et al. (6, 7) reviewed the properties and assumptions of population adjustment methods and provided recommendations for their use in submissions to NICE. As summarised in Table 1, population adjustment in anchored scenarios (where a common comparator is available) relaxes the constancy of relative effects assumption (to conditional constancy of relative effects) by adjusting for effect modifiers. In unanchored scenarios a much stronger assumption is required (conditional constancy of absolute effects), since it is necessary to reliably predict absolute outcomes. This requires all effect modifiers and prognostic variables to be adjusted for, and is very difficult to achieve or justify—either empirically or otherwise. As such, unanchored comparisons are subject to unknown amounts of residual bias due to unobserved prognostic variables and effect modifiers (6, 7).
Table 1.
All indirect comparisons and meta-analyses require some form of constancy assumption. Unanchored comparisons require a much stronger assumption, which is widely considered impossible to meet.
| Method | Anchored comparisons | Unanchored comparisons | |
|---|---|---|---|
| Standard indirect comparison or meta-analysis | Anchored population-adjusted indirect comparison | Unanchored population-adjusted indirect comparison | |
| Constancy assumption | Constancy of relative effects | Conditional constancy of relative effects | Conditional constancy of absolute effects |
| ⇒ Relative effects are the same across populations | ⇒ Reliable predictions of relative effects | ⇒ Reliable predictions of absolute effects | |
| Valid only if | No effect modifiers in imbalance | All effect modifiers known and adjusted for | All effect modifiers and prognostic variables known and adjusted for |
| Data requirements | Aggregate data | IPD on at least one trial | IPD on at least one trial |
Although statistical theory is clear on which variables must be adjusted for to obtain an unbiased indirect comparison, in practice variable selection requires judgement and justification (6). For anchored comparisons, evidence of effect modifier status on the chosen scale should be provided prior to analysis. Unanchored comparisons are very difficult to justify, but a principled approach to selecting variables prior to analysis should be taken in order to avoid “gaming”.
The level of overlap in the covariate distributions between the IPD and AgD study populations is another key property of population adjusted indirect comparisons. For regression-based methods such as STC, the lesser the overlap the greater the amount of extrapolation required, which requires additional assumptions to be valid. For re-weighting methods such as MAIC, extrapolation is simply not possible; sufficient overlap is therefore crucial for re-weighting methods. As well as checking the distributions of the covariates in each study, a simple indicator of the amount of overlap is the effective sample size (ESS), which may be calculated from the weights (6). Large reductions in ESS may indicate poor overlap between the IPD and AgD studies; small absolute ESS shows that the comparison is dependent on a small number of individuals in the IPD study and may be unstable.
MAIC and STC were both designed with simple indirect comparisons in mind and do not generalise naturally to larger networks of studies or treatments, where there may be multiple comparators of interest and/or multiple aggregate study populations. We are therefore interested in the prevalence of such scenarios in NICE TAs, to determine how larger network structures have been handled using current methods and to motivate the development of more appropriate methods. In this paper we undertake a comprehensive review of technology appraisals (TAs) published by NICE (8), aiming to characterise the use of population adjustment methods. As well as investigating the uptake of population adjustment in different clinical areas, we are interested in the ways in which these methods are used and whether the key assumptions are likely to hold, to assess the adequacy of current practice for decision making. We discuss how these methods have been received by appraisal committees and how they have impacted decision making. We conclude with a discussion, and suggest several key improvements to current practice, towards providing better evidence for decision makers and greater impact for submitting companies.
Methods
We reviewed all NICE TAs published between 1st January 2010 and 20th April 2018 for the use of population adjustment methods. We excluded appraisals that had access to IPD from all included studies, and focussed on those with only partial availability of IPD. From those appraisals using one or more forms of population adjustment, we extracted the following information from company submissions:
Population adjustment method used;
Whether the comparison was anchored or unanchored;
Outcome type;
Clinical area;
Number of covariates adjusted for;
How the covariates were chosen;
For appraisals using MAIC, effective sample sizes after weighting;
Whether a larger network structure was present (e.g. multiple comparators and/or aggregate studies), and how this was dealt with.
Screening and data extraction were carried out by a single primary reviewer (DMP).
Results
A total of 268 technology appraisals have been published by NICE since 2010—when MAIC and STC were first suggested in the literature (4, 5)—up until 20th April 2018. Of these, 21 appraisals used a form of population adjustment; three of these had IPD available from all included studies, so we focus on the remaining 18 appraisals with only partial IPD. Figure 1 shows the selection process. The included appraisals are tabulated in Supplementary Table 1.
Figure 1.
Flow chart showing the process of selecting technology appraisals, and the numbers excluded and remaining at each stage.
The first use of population adjustment in a TA was TA311 in 2014. Since then, the use of population adjustment in TAs has increased rapidly, in terms of both the absolute number and the relative proportion of appraisals using population adjustment methods (Fig. 2). In 2017, a total of nine appraisals used population adjustment, accounting for 14.5% of all appraisals that year.
Figure 2.
The number and percentage of NICE technology appraisals using population adjustment methodology has increased greatly since the introduction of these methods in the literature in 2010.
* Two TAs used population adjustment out of 25 up to 20th April 2018.
Usage by clinical area
Since 2010, almost half of all published TAs have been in oncology (127 of 268, 47.4%). Of these, 15 (11.8%) used population adjustment, accounting for over 80 percent of all applications of population adjustment in appraisals to date. Only two other clinical areas saw any applications of population adjustment: two out of 12 (16.7%) appraisals in hepatology (both for hepatitis C), and one out of 28 (3.6%) appraisals in rheumatology. The usage of population adjustment methods in oncology TAs has increased since 2010, both in terms of the number and proportion of TAs using these methods. In 2017, a total of 9 appraisals in oncology (25.7%) used population adjustment methods, up from one appraisal (9.1%) in 2014 (Fig. 3). The increasing use of population adjustment in oncology appraisals, which themselves make up the largest proportion of all appraisals, is the main driver behind the overall results in Fig. 2.
Figure 3.
For technology appraisals in oncology, the number and percentage using population adjustment methodology has increased greatly since the introduction of these methods in the literature in 2010.
* Two TAs used population adjustment out of 12 up to 20th April 2018.
Outcome types
Unsurprisingly, due to the majority of applications of population adjustment being in oncology appraisals, survival outcomes (e.g. progression-free survival, overall survival) were the most common outcome type used in population-adjusted analyses—13 of 18 appraisals (72.2%) included a population-adjusted survival outcome. Rate outcomes such as response rates were used in 5 appraisals, and duration and change from baseline outcomes in one appraisal each. Two appraisals (TA462, TA451) used population adjustment for more than one type of outcome (survival and response rate, and response rate and duration respectively).
Population adjustment method
The large majority of appraisals using some form of population adjustment used MAIC (16 out of 18, 88.9%). STC was less popular, used in only 3 appraisals (16.7%). Two appraisals used both MAIC and STC and compared the results, which were reported to be similar in each case (TA383, TA492).
One appraisal (TA410) used neither MAIC nor STC. In this appraisal, a published prediction model (developed for a previous appraisal (9)) was used to adjust the survival curves from the AgD trials to the population of the IPD trial.
Of the 16 appraisals performing MAIC, only 9 (56.3%) reported an effective sample size. Of these, the median effective sample size was 80.0 (range: 4.0 to 335.5, IQR: 15.4 to 52.0), with a median reduction in effective sample size from the original sample size of 74.2% (range: 7.9% to 94.1%, IQR: 48.0% to 84.6%). Such large reductions in ESS indicate that in many cases there may be poor overlap between the IPD and AgD studies. A substantial proportion of TAs reported small absolute ESS, and the resulting comparisons are therefore dependent on a small number of individuals in the IPD study and may be unstable.
Anchored and unanchored comparisons
Only 2/18 appraisals (11.1%) formed anchored comparisons (TA383, TA449). The remaining 16 appraisals (88.9%) instead formed unanchored comparisons without a common comparator, relying on strong assumptions that are very difficult to justify and are thus subject to unknown amounts of residual bias. No appraisals attempted to quantify residual bias, although this is challenging to achieve (6). Appraisal committees and review groups treated estimates from unanchored comparisons with strong caution.
Covariates adjusted for
For appraisals reporting unanchored comparisons, the median number of covariates adjusted for was 6, and ranged from 1 to 13 covariates. Only one of the two appraisals reporting anchored comparisons presented any information on the choice of covariates; in this appraisal (TA383) 10 covariates were adjusted for.
Common covariates adjusted for in oncology appraisals were age, Eastern Cooperative Oncology Group (ECOG) performance status, gender, and the number and/or type of previous therapies. Many appraisals also adjusted for other clinical factors such as biomarker levels or disease subtypes.
Both hepatitis C appraisals (TA364, TA331) adjusted for age, body mass index (BMI), gender, fibrosis staging, and viral load. One appraisal (TA364) further adjusted for race, genotype, and several biomarker levels in two MAIC analyses for different genotypes and comparator treatments, but in a third MAIC analysis only had sufficient sample size to adjust for viral load.
The single rheumatology appraisal (TA383) adjusted for 10 covariates including age, gender, race, concomitant treatments, two biomarkers and three functional/activity scores.
The most common justification for covariate selection amongst appraisals reporting unanchored comparisons was simply to adjust for all baseline characteristics reported in both studies. This was also true for appraisal TA383 which used an anchored comparison, despite the fact that adjustment is only required for covariates which were effect modifiers in anchored comparisons. (The other appraisal with an anchored comparison, TA449, did not report any information on variable selection.) Unnecessary adjustment will not introduce bias but may increase uncertainty, particularly with MAIC (6) (although we note that TA383 took place before the advice in (6) was published). Two appraisals (TA429, TA457) justified the selection of covariates using expert clinical opinion. One appraisal using MAIC (TA510) asked experts to rank covariates by importance, then added covariates into the model one-by-one in decreasing order of importance; the final model choice was determined by consideration of effective sample size. Unanchored MAICs in particular have to make trade-offs between effective sample size and the number of adjustment variables, since the number of potential prognostic factors is likely to be large. However, unless all prognostic factors and effect modifiers are included in the adjustment, the estimates will remain biased (6). Moreover, the covariates for which the effective sample size reduction is greatest are those which are most imbalanced between populations, and are therefore more important to adjust for amongst the covariates with similar prognostic or effect modifying strength. Two appraisals using STC used statistical techniques to choose covariates. One (TA333) selected covariates that were “significant” in the regression model, which is again likely to incur residual bias—particularly in small samples (10). Another (TA492) selected covariates to maximise cross-validated predictive performance, which is more appropriate given that STC relies on accurate predictions into the aggregate population, but is still subject to the limitations of in-sample validation (6).
Larger networks
As originally proposed, MAIC and STC cannot be extended to larger network structures with multiple comparators of interest and/or multiple aggregate studies. However, these scenarios frequently arise in practice: a total of 10 out of 18 TAs (55.6%) involved larger networks of treatments and studies.
In five of these (71.4%; TA331, TA383, TA429, TA500, TA510), multiple population adjusted indirect comparisons were performed and then simply left as stand-alone estimates. Each of these estimates will be valid for different target populations, and so cannot be interpreted together coherently unless additional assumptions are met—namely that all the target populations are in fact identical (in terms of effect modifiers for anchored comparisons, and also in terms of prognostic variables for unanchored comparisons).
One appraisal (TA492) used STC (and MAIC as a sensitivity analysis) to predict active treatment arms for each single-arm study in an unconnected network, and then analysed this newly-connected network using network meta-analysis (NMA). This results in a coherent set of relative effect estimates (11). However, aside from the very strong assumptions required for the unanchored comparisons, this analysis must also assume that there are no imbalances in effect modifiers between the single-arm studies included in the NMA. Another serious concern is the repeated use of the predicted active treatment arms, which are all based on the same data set and so are not independent.
Two appraisals (TA311, TA380) had wider networks of treatments and studies including the two treatments of primary interest, but that were not fully connected. These networks were analysed using NMA (without any population adjustment) using an equivalency assumption for two treatments (TA311) and a matched pairs analysis (TA380) to connect the networks. Separate unanchored MAICs were then used to create population-adjusted comparisons as sensitivity analyses.
One appraisal (TA427) had additional single-arm IPD sources which were used to provide additional stand-alone comparisons (in this case using Cox regression for survival outcomes).
Lastly, the method of analysis was unclear for one appraisal (TA364) which had multiple comparators of interest, some with several aggregate studies available. However, given that unanchored MAIC was used, this analysis is susceptible to the same sets of pitfalls described above depending on whether the estimates were left as stand-alone estimates or synthesised as a network.
Discussion
In this review we have focussed on the use of population adjustment methods in NICE Technology Appraisals. Different practices may be found in submissions to other reimbursement agencies, who may also receive and interpret such analyses differently, and outside of the technology appraisal context. A general review of applications in the literature has previously been published by Phillippo et al. (6) and found similar issues to those discussed here, although with even greater variation in analysis practices. This review of TAs also spans a limited time period (8 years) since these methods were first published. Practice is likely to continue to evolve, for example as methodological guidance is published (6). A further limitation of this review is that the data extraction was carried out by a single reviewer only.
Population adjustment methods (and in particular unanchored MAIC) have been used in NICE TAs either as the main source of comparative clinical effectiveness, or as supportive evidence alongside the company’s base-case analysis. Whilst these methods may account for some differences between study populations that conventional indirect comparison methods cannot, appraisal committees were often concerned by the quality of the estimates they produced. This was not necessarily due to inherent methodological limitations; rather, the methods were used in situations where the data underpinning the analyses were often weak (for example immature follow-up data or small single arm studies).
Furthermore, population-adjusted comparisons were often associated with uncertainty regarding the covariates that were adjusted for—specifically, which ones were selected for adjustment, how they were selected, and whether and to what extent any unobserved characteristics biased the analysis. A key challenge to appraisal committees, especially with unanchored comparisons, was where to draw the line between the number of variables to adjust for and the precision of the resulting estimates. This was particularly apparent for MAIC, where the effective sample size decreases with each additional covariate adjusted for.
In NICE TAs, decisions are not based solely on clinical effectiveness; cost considerations are also taken into account in a cost-effectiveness analysis, summarised by an Incremental Cost Effectiveness Ratio (ICER). The impact of evidence from population-adjusted indirect comparisons is therefore understood within this context. In some instances, appraisal committees could not make a positive recommendation for the technology because the uncertainty in the population-adjusted estimates were not offset by a sufficiently low ICER to manage the decision risk. Where the appraisal committee made a positive recommendation, the committee typically compared the most plausible ICER against the lower end of the acceptable range (requiring the technology to be more cost effective) to minimise the risks associated with uncertainty. Where appraisal committees judged a technology to have plausible potential to be cost effective, they often recommended the use of the technology with interim funding as part of a data collection arrangement. In general, appraisal committees tended to use population adjustment methods for decision-making when they were presented alongside an alternative, confirmatory analysis, and when the uncertainty in the method was acknowledged, described, and explored as far as possible (for example using sensitivity analyses). Appraisal committees have previously suggested that companies should also consider validating the results of their analyses (e.g. TA510), for example by estimating the effect of the technology using population adjustment methods in an external cohort (such as registry data) and comparing that estimate with the observed effect of the technology in that cohort.
Population adjustment methods are becoming ever more prevalent in NICE TAs. The majority were unanchored with no common comparator, and hence rely on very strong assumptions as outlined in Table 1. The proliferation of unanchored analyses is likely to escalate, in large part due to the rise of single-arm studies for accelerated or conditional approval with regulators such as the US Food and Drug Administration or the European Medicines Agency (12). However, the evidential requirements for demonstrating clinical efficacy (to obtain licensing) can be less stringent than those for demonstrating cost effectiveness (to obtain reimbursement). NICE appraisal committees and evidence review groups have been justifiably wary of the use of unanchored population adjustment methods to bridge this evidence gap, with many commenting that the results should be interpreted with caution and may contain an unknown amount of bias. As such, committees typically looked for greater cost effectiveness (lower ICER) to minimise the decision risk resulting from clinical effectiveness evidence perceived to be uncertain and poor quality. Increased dialogue between regulators and reimbursement agencies may help bridge this gap in evidence requirements.
All current population adjustment methods assume that there are no unmeasured effect modifiers when making anchored comparisons. For unanchored comparisons, it is further assumed that there are no unmeasured prognostic factors. This latter assumption is particularly strong and difficult to justify. Some suggestions for quantifying residual bias due to unmeasured confounding are made by Phillippo et al. (6), and this is an area for further work.
Several technology appraisals had multiple comparators and/or AgD study populations for which comparisons were required. Current MAIC and STC methodology cannot handle larger network structures: multiple analyses were performed in each case, and then either left as stand-alone comparisons or themselves synthesised using network meta-analysis, requiring further assumptions in the process. Furthermore, current MAIC and STC methods produce estimates which are valid only for the aggregate study population (typically that of a competitor) without additional assumptions, which may not match the target population for the decision (6). This fact has been largely overlooked in appraisals to date, although one appraisal (TA451) did note that the MAIC analysis that was performed took the results of an IPD trial deemed to be relevant to the decision population and adjusted them into a non-representative aggregate trial population. Clearly if effect modification is present then it is not enough to simply produce “unbiased” estimates: the estimates produced must be specific to the decision population, otherwise they are of little use to decision-makers. This motivates the need to develop new methods which can extend naturally to larger networks of treatments, and can produce estimates for a given target decision population. Furthermore, if all trials are a subset of the decision target population with respect to one or more effect modifiers, then any adjustment must rely on extrapolation; if these effect modifiers are discrete, adjustment may be impossible.
The large majority of technology appraisals used MAIC to obtain population-adjusted indirect comparisons. Effective sample sizes were typically small and often substantially reduced compared to the original sample sizes, indicating potential lack of overlap between the IPD and AgD populations. Lack of overlap is of particular concern with re-weighting methods such as MAIC since they cannot extrapolate to account for covariate values beyond those observed in the IPD, and thus may produce estimates that remain biased even when all necessary covariates are included in the model (6). This motivates the need for simulation studies to explore the robustness of MAIC (and other population adjustment methods) in scenarios where there is a lack of overlap between populations.
Three appraisals were excluded from our review, as IPD were available from all included studies (13–15). These appraisals were all unanchored comparisons of survival outcomes in oncology, and used a selection of propensity score, covariate matching, and regression methods. Having IPD available from all studies is the gold-standard and is preferable if at all possible. This is because IPD allows for analyses that have more statistical power and may rely on less stringent assumptions, and allows assumptions to be tested. Separate methodological guidance is available for analyses with full IPD (16).
For population-adjusted analyses to have the desired impact on decision making in technology appraisals, several key improvements are needed to current practice in line with recent guidance (6). Firstly, a target population relevant to decision makers should be defined, and estimates must be produced for this population in order to be relevant. Current population adjustment methods can only produce estimates valid for the population represented by the aggregate study unless further assumptions are made, which may not represent the decision at hand; this has been largely overlooked in appraisals to date (although note that several of the TAs we identified pre-date published guidance). For anchored comparisons there should be clear prior justification for effect modification, based on empirical evidence from previous studies and/or clinical expertise. Appraisals reporting anchored comparisons to date did not provide any such justification. Unanchored comparisons require reliable predictions of absolute effects via adjustment for both prognostic and effect-modifying covariates, and are highly susceptible to unobserved confounding due to a lack of randomisation. Simply adjusting for all available covariates, as is currently common practice, is not sufficient. For unanchored comparisons to be impactful, covariates should be selected with predictive performance in mind and estimates of the potential range of residual bias are required; otherwise, the amount of bias in the estimates is unknown and may even be larger than for an unadjusted comparison. This is not easy to achieve (some suggestions are made in (6)), but without such reassurance appraisal committees are likely to remain justifiably wary of unanchored analyses. Many of the above issues can be mitigated—at least in part—by the availability of IPD from all studies in an analysis, and thus the increased sharing of IPD is greatly encouraged.
Supplementary Material
Funding
This work was supported by the UK Medical Research Council grants MR/P015298/1 (DMP, SD, AEA, NJW) and MR/M005232/1 (SD), and the ConDuCT-II Hub for Trials Methodology Research grant MR/K025643/1 (NJW).
Footnotes
Conflict of interest
NJW is PI on a methodology grant jointly funded by the MRC and Pfizer Ltd. Pfizer part fund a junior researcher. The project is purely methodological and unrelated to this work.
DMP, SD, AE, and AEA declare that they have no conflicts of interest.
References
- 1.Bucher HC, Guyatt GH, Griffith LE, Walter SD. The results of direct and indirect treatment comparisons in meta-analysis of randomized controlled trials. Journal of Clinical Epidemiology. 1997;50:683–691. doi: 10.1016/s0895-4356(97)00049-8. [DOI] [PubMed] [Google Scholar]
- 2.Dias S, Sutton AJ, Ades AE, Welton NJ. Evidence Synthesis for Decision Making 2: A Generalized Linear Modelling Framework for Pairwise and Network Meta-analysis of Randomized Controlled Trials. Medical Decision Making. 2013;33:607–617. doi: 10.1177/0272989X12458724. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Dias S, Ades AE, Welton NJ, Jansen JP, Sutton AJ. Network Meta-Analysis for Decision Making. John Wiley & Sons Inc; 2018. p. 488. [DOI] [Google Scholar]
- 4.Signorovitch JE, Wu EQ, Yu AP, et al. Comparative Effectiveness Without Head-to-Head Trials A Method for Matching-Adjusted Indirect Comparisons Applied to Psoriasis Treatment with Adalimumab or Etanercept. Pharmacoeconomics. 2010;28:935–945. doi: 10.2165/11538370-000000000-00000. [DOI] [PubMed] [Google Scholar]
- 5.Caro JJ, Ishak KJ. No Head-to-Head Trial? Simulate the Missing Arms. Pharmacoeconomics. 2010;28:957–967. doi: 10.2165/11537420-000000000-00000. [DOI] [PubMed] [Google Scholar]
- 6.Phillippo DM, Ades AE, Dias S, Palmer S, Abrams KR, Welton NJ. NICE DSU Technical Support Document 18: Methods for population-adjusted indirect comparisons in submission to NICE. Tech. rep. National Institute for Health and Care Excellence. 2016 url: www.nicedsu.org.uk. [Google Scholar]
- 7.Phillippo DM, Ades AE, Dias S, Palmer S, Abrams KR, Welton NJ. Methods for Population-Adjusted Indirect Comparisons in Health Technology Appraisal. Medical Decision Making. 2018;38:200–211. doi: 10.1177/0272989X17725740. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.National Institute for Health and Care Excellence. Technology appraisals guidance. url: www.nice.org.uk/guidance/published?type=ta (visited on 2018)
- 9.National Institute for Health and Care Excellence. Ipilimumab for previously untreated advanced (unresectable or metastatic) melanoma. NICE guideline (TA319) 2014 url: www.nice.org.uk/guidance/TA319. [Google Scholar]
- 10.Steyerberg EW, Eijkemans MJC, Habbema JDF. Stepwise Selection in Small Data Sets: A Simulation Study of Bias in Logistic Regression Analysis. Journal of Clinical Epidemiology. 1999;52:935–942. doi: 10.1016/s0895-4356(99)00103-1. [DOI] [PubMed] [Google Scholar]
- 11.Caldwell DM, Ades AE, Higgins JPT. Simultaneous comparison of multiple treatments: combining direct and indirect evidence. BMJ. 2005;331:897–900. doi: 10.1136/bmj.331.7521.897. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Hatswell AJ, Baio G, Berlin JA, Irs A, Freemantle N. Regulatory approval of pharmaceuticals without a randomised controlled study: analysis of EMA and FDA approvals 1999–2014. BMJ Open. 2016;6:e011666. doi: 10.1136/bmjopen-2016-011666. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.National Institute for Health and Care Excellence. Osimertinib for treating locally advanced or metastatic EGFR T790M mutation-positive non-small-cell lung cancer. NICE guideline (TA416) 2016 url: www.nice.org.uk/guidance/TA416. [Google Scholar]
- 14.National Institute for Health and Care Excellence. Ibrutinib for treating Waldenstrom’s macroglobulinaemia. NICE guideline (TA491) 2017 url: www.nice.org.uk/guidance/TA491. [Google Scholar]
- 15.National Institute for Health and Care Excellence. Avelumab for treating metastatic Merkel cell carcinoma. NICE guideline (TA517) 2018 url: www.nice.org.uk/guidance/TA517. [Google Scholar]
- 16.Faria R, Hernandez Alava M, Manca A, Wailoo AJ. NICE DSU Technical Support Document 17: The use of observational data to inform estimates of treatment effectiveness for Technology Appraisal: Methods for comparative individual patient data. Tech. rep. National Institute for Health and Care Excellence. 2015 url: http://www.nicedsu.org.uk. [Google Scholar]
- 17.National Institute for Health and Care Excellence. Daratumumab monotherapy for treating relapsed and refractory multiple myeloma. NICE guideline (TA510) 2018 url: www.nice.org.uk/guidance/TA510. [Google Scholar]
- 18.National Institute for Health and Care Excellence. Ceritinib for untreated ALK-positive nonsmall-cell lung cancer. NICE guideline (TA500) 2018 url: www.nice.org.uk/guidance/TA500. [Google Scholar]
- 19.National Institute for Health and Care Excellence. Atezolizumab for untreated locally advanced or metastatic urothelial cancer when cisplatin is unsuitable. NICE guideline (TA492) 2017 url: www.nice.org.uk/guidance/TA492. [Google Scholar]
- 20.National Institute for Health and Care Excellence. Nivolumab for treating relapsed or refractory classical Hodgkin lymphoma. NICE guideline (TA462) 2017 url: www.nice.org.uk/guidance/TA462. [Google Scholar]
- 21.National Institute for Health and Care Excellence. Brentuximab vedotin for treating relapsed or refractory systemic anaplastic large cell lymphoma. NICE guideline (TA478) 2017 url: www.nice.org.uk/guidance/TA478. [Google Scholar]
- 22.National Institute for Health and Care Excellence. Carfilzomib for previously treated multiple myeloma. NICE guideline (TA457) 2017 url: www.nice.org.uk/guidance/TA457. [Google Scholar]
- 23.National Institute for Health and Care Excellence. Everolimus and sunitinib for treating unresectable or metastatic neuroendocrine tumours in people with progressive disease. NICE guideline (TA449) 2017 url: www.nice.org.uk/guidance/TA449. [Google Scholar]
- 24.National Institute for Health and Care Excellence. Ponatinib for treating chronic myeloid leukaemia and acute lymphoblastic leukaemia. NICE guideline (TA451) 2017 url: www.nice.org.uk/guidance/TA451. [Google Scholar]
- 25.National Institute for Health and Care Excellence. Everolimus for advanced renal cell carcinoma after previous treatment. NICE guideline (TA432) 2017 url: www.nice.org.uk/guidance/TA432. [Google Scholar]
- 26.National Institute for Health and Care Excellence. Ibrutinib for previously treated chronic lymphocytic leukaemia and untreated chronic lymphocytic leukaemia with 17p deletion or TP53 mutation. NICE guideline (TA429) 2017 doi: 10.1016/S1470-2045(17)30062-1. url: www.nice.org.uk/guidance/TA429. [DOI] [PubMed] [Google Scholar]
- 27.National Institute for Health and Care Excellence. Pomalidomide for multiple myeloma previously treated with lenalidomide and bortezomib. NICE guideline (TA427) 2017 url: www.nice.org.uk/guidance/TA427. [Google Scholar]
- 28.National Institute for Health and Care Excellence. Talimogene laherparepvec for treating unresectable metastatic melanoma. NICE guideline (TA410) 2016 url: www.nice.org.uk/guidance/TA410. [Google Scholar]
- 29.National Institute for Health and Care Excellence. TNF-alpha inhibitors for ankylosing spondylitis and non-radiographic axial spondyloarthritis. NICE guideline (TA383) 2016 url: www.nice.org.uk/guidance/TA383. [Google Scholar]
- 30.National Institute for Health and Care Excellence. Panobinostat for treating multiple myeloma after at least 2 previous treatments. NICE guideline (TA380) 2016 url: www.nice.org.uk/guidance/TA380. [Google Scholar]
- 31.National Institute for Health and Care Excellence. Daclatasvir for treating chronic hepatitis C. NICE guideline (TA364) 2015 url: www.nice.org.uk/guidance/TA364. [Google Scholar]
- 32.National Institute for Health and Care Excellence. Simeprevir in combination with peginterferon alfa and ribavirin for treating genotypes 1 and 4 chronic hepatitis C. NICE guideline (TA331) 2015 url: www.nice.org.uk/guidance/TA331. [Google Scholar]
- 33.National Institute for Health and Care Excellence. Axitinib for treating advanced renal cell carcinoma after failure of prior systemic treatment. NICE guideline (TA333) 2015 doi: 10.1016/S1470-2045(14)70484-X. url: www.nice.org.uk/guidance/TA333. [DOI] [PubMed] [Google Scholar]
- 34.National Institute for Health and Care Excellence. Bortezomib for induction therapy in multiple myeloma before high-dose chemotherapy and autologous stem cell transplantation. NICE guideline (TA311) 2014 url: www.nice.org.uk/guidance/TA311. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.



