INTRODUCTION
Critical care research is characterized by many opportunities due to the data-rich environment of the intensive care unit (ICU): physiological monitoring derangements, patients’ prior characteristics, acute organ dysfunction, metabolic derangements, and many laboratory and organ support measurements. Many ICUs integrate, in real-time, monitors, laboratory results and charts, allowing high time-resolution and completeness, providing large, open-sourced, databases.(1,2) While this leads to a scientifically rich environment, it may cloud scientific thinking and lead to data-driven studies without clear hypotheses. Here, we will discuss common pitfalls researchers should be aware of when conducting critical care research.
Lack of a clear, relevant research question
A clear research question is a tenet for the planning, execution, and interpretation of the results, and it will pave the way to avoid many pitfalls. The underlying theory must guide the research, not the statistical results. Although it seems trivial, even experienced researchers may dive into a loophole due to an unclear research question. Research question clarity involves two main issues. First, the researcher must be sure of their primary aim: is it mainly descriptive, causal, or predictive? Second, the researcher should have a clear theoretical model of how the intervention works, in case of causal questions, or how a new diagnostic or predictive marker will fit in clinical practice, in case of diagnostic and prognostic research. Mixing concepts is very common and clouds scientific thinking.
Definition of inclusion and exclusion criteria
In clinical trials, much thought is put into inclusion and exclusion criteria to adequately sample the target population. This rationale must be followed for other study designs too and it is dependent on the study aim: for causal questions, exclusion criteria should mirror those of a pragmatic clinical trial. For prognosis and diagnosis questions, the study population should mirror the one that would have the test applied in practice. Authors should avoid excluding patients due to missing data, which should be appropriately dealt with.(3)
Over reliance on statistical significance instead of estimation
Traditional teaching in healthcare research emphasizes null hypothesis significance testing and p value interpretation. Researchers must focus, however, on estimation.(4) Quantitative research is about estimating a given quantity of interest. It could be the prevalence of a condition for descriptive research, an estimate of risk ratios for a causal question, or the risk of an outcome for prognosis studies. This does not mean statistical significance is doomed, but that it must be presented with point estimates and confidence intervals and, sometimes, not even conducted depending on the scenario.(5)
Association of physiological derangements with death
Although critical care practice is intimately related to avoiding death, authors should avoid the common shortcut of finding an association between physiological derangements and mortality. Essentially, all physiological derangements will be to some extent associated with mortality (i.e., those who are most seriously ill are more likely to die), but these are commonly only prognostic associations. Well-thought research questions may need a demonstration of this association at the outstart of a line of research, but authors should consider potential causal paths amenable to treatment when moving forward.
Variable selection for multivariable models
Variable selection must be aligned with the research question. For causal questions, variable selection should be based on subject-matter knowledge by drawing conceptual models, such as causal directed acyclic graphs (DAGs).(6,7) For predictive studies, as the aim is to estimate a probability, the causal relationship is not essential, but the outcome explainability.(8) Nevertheless, for predictive questions, the authors must consider what variables are available at the point of care. These variables should not be a surrogate of what is meant to be predicted, which would lead to overly optimistic estimates of the model and poor applicability. Mixing concepts of prediction variable selection with DAG-based variable selection for causal questions may lead to low-value data presentation.
Table 2 Fallacy
The Table 2 Fallacy is an underrecognized issue in observational studies with multivariable models.(9) Except in particular circumstances of prognostic research in which researchers aim to find an association of a combination of variables with clinical outcomes, not all coefficients of a multivariable model should be presented. In causal research, variables included in a multivariable model for adjustment should not have their results presented since their "effect estimate" is not estimated by that model. Each set of variables included in a model was so because they were necessary to address that research question. Furthermore, even in prognostic research, the association of a new prognostic variable with outcome should be evaluated by accounting for known, observed prognostic variables in the model, regardless of hypothesis testing in that given dataset.
No pre-registration
Registration of clinical trials is a standard practice to avoid publication bias and undescribed modifications to study procedures during study roll-out. Systematic reviews must also have a registered protocol to allow a-priori hypothesis specification.(10) Protocol deviations must be justified when needed. Absent pre-registration might preclude their publication. While observational studies still do not need pre-registration, authors should consider doing so or making the analysis protocol publicly available.
Abstracts without estimates
Abstracts are the first impression of a manuscript. In an abstract, a single main clear objective should be answered with the conclusion of the manuscript, but the focus should be on a brief description of methods and the main results’ presentation. The results need to be presented with estimates and their confidence intervals. Authors should avoid presenting only p values.
New methods are not always better methods
While as scientists we desire better methods to answer our research questions, new methods are not necessarily better methods. For example, the net reclassification improvement arose a decade ago as a new method for biomarker research, but it did not pass the test of time.(11) While Bayesian analysis,(12) machine learning, and artificial intelligence(13,14) are in the spotlight, they should not drive the research question.
CONCLUSION
From this viewpoint, we presented a few issues that we believe are important for critical care researchers. Figure 1 provides a contrast of good versus not-so-good practices in an observational study for further reference, which can be applied to a recently published systematic review.(15)
Figure 1. Examples of a contrast of a good (and not so good) hypothetical observational study about the effects of awake prone positioning in hospitalized COVID-19 patients with respiratory failure.
DAG - directed acyclic graph; NHST - null hypothesis significance testing.
Footnotes
Publisher's note
REFERENCES
- 1.Johnson AE, Bulgarelli L, Shen L, Gayles A, Shammout A, Horng S, et al. MIMIC-IV, a freely accessible electronic health record dataset. Sci Data. 2023;10(1):1–1. doi: 10.1038/s41597-022-01899-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Pollard TJ, Johnson AE, Raffa JD, Celi LA, Mark RG, Badawi O. The eICU Collaborative Research Database, a freely available multi-center database for critical care research. Sci Data. 2018;5(1):180178–180178. doi: 10.1038/sdata.2018.178. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Vesin A, Azoulay E, Ruckly S, Vignoud L, Rusinovà K, Benoit D, et al. Reporting and handling missing values in clinical studies in intensive care units. Intensive Care Med. 2013;39(8):1396–1404. doi: 10.1007/s00134-013-2949-1. [DOI] [PubMed] [Google Scholar]
- 4.Lederer DJ, Bell SC, Branson RD, Chalmers JD, Marshall R, Maslove DM, et al. Control of confounding and reporting of results in causal inference studies. Guidance for authors from editors of respiratory, sleep, and critical care journals. Ann Am Thorac Soc. 2019;16(1):22–28. doi: 10.1513/AnnalsATS.201808-564PS. [DOI] [PubMed] [Google Scholar]
- 5.Greenland S, Senn SJ, Rothman KJ, Carlin JB, Poole C, Goodman SN, et al. Statistical tests, p values, confidence intervals, and power: a guide to misinterpretations. Eur J Epidemiol. 2016;31(4):337–350. doi: 10.1007/s10654-016-0149-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Digitale JC, Martin JN, Glymour MM. Tutorial on directed acyclic graphs. J Clin Epidemiol. 2022;142:264–267. doi: 10.1016/j.jclinepi.2021.08.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Fumo-dos-Santos C, Ferreira JC. Dealing with confounding in observational studies. J Bras Pneumol. 2023;49(4):e20230281. doi: 10.36416/1806-3756/e20230281. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Leisman DE, Harhay MO, Lederer DJ, Abramson M, Adjei AA, Bakker J, et al. Development and reporting of prediction models: guidance for authors from editors of respiratory, sleep, and critical care journals. Crit Care Med. 2020;48(5):623–633. doi: 10.1097/CCM.0000000000004246. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Westreich D, Greenland S. The table 2 fallacy: presenting and interpreting confounder and modifier coefficients. Am J Epidemiol. 2013;177(4):292–298. doi: 10.1093/aje/kws412. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Adigbli D, Yang L, Hammond N, Annane D, Arabi Y, Bilotta F, et al. Intensive glucose control in critically ill adults: a protocol for a systematic review and individual patient data meta-analysis. Crit Care Sci. 2023;35(4):345–354. doi: 10.5935/2965-2774.20230162-en. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Pepe MS, Fan J, Feng Z, Gerds T, Hilden J. The Net Reclassification Index (NRI): a misleading measure of prediction improvement even with independent test data sets. Stat Biosci. 2015;7(2):282–295. doi: 10.1007/s12561-014-9118-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Zampieri FG, Damiani LP, Biondi RS, Freitas FG, Veiga VC, Figueiredo RC, et al. BRICNet Effects of balanced solution on short-term outcomes in traumatic brain injury patients: a secondary analysis of the BaSICS randomized trial. Rev Bras Ter Intensiva. 2022;34(4):410–417. doi: 10.5935/0103-507X.20220261-en. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Ururahy RD, Gallo CA, Besen BA, Carvalho MT, Ribeiro JM, Zigaib R, et al. Bedside clinical data subphenotypes of critically ill COVID-19 patients: a cohort study. Rev Bras Ter Intensiva. 2021;33(2):196–205. doi: 10.5935/0103-507X.20210027. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Santana A, Prestes GD, Silva MD, Girardi CS, Silva LD, Moreira JC, et al. Identification of distinct phenotypes and improving prognosis using metabolic biomarkers in COVID-19 patients. Crit Care Sci. 2024;36:e20240028en. doi: 10.62675/2965-2774.20240028-en. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Vásquez-Tirado GA, Meregildo-Rodríguez ED, Asmat-Rubio MG, Salazar-Castillo MJ, Quispe-Castañeda CV, Cuadra-Campos MD. Conscious prone positioning in nonintubated COVID-19 patients with acute respiratory distress syndrome: systematic review and meta-analysis. Crit Care Sci. 2024;36:e20240176en. doi: 10.62675/2965-2774.20240176-en. [DOI] [PMC free article] [PubMed] [Google Scholar]

