Skip to main content
Journal of Mass Spectrometry and Advances in the Clinical Lab logoLink to Journal of Mass Spectrometry and Advances in the Clinical Lab
. 2023 Feb 17;28:47–55. doi: 10.1016/j.jmsacl.2023.02.003

Small molecule biomarker discovery: Proposed workflow for LC-MS-based clinical research projects

S Rischke a, L Hahnefeld a,b, B Burla c, F Behrens b,d, R Gurke a,b,, TJ Garrett e,
PMCID: PMC9982001  PMID: 36872952

Graphical abstract

graphic file with name ga1.jpg

Abbreviations: HILIC, Hydrophilic interaction liquid chromatography; HRMS, High resolution mass spectrometry; LC-MS, Liquid chromatography – mass spectrometry; MRM, Multiple reaction monitoring; PCA, Principal component analysis; QA, Quality assurance; QC, Quality control; RF, Random Forest; RP, Reversed phase; SVA, Support vector machine; (U)HPLC (Ultra-), High pressure liquid chromatography

Keywords: LC-MS-Based Clinical Research, Biomarker Discovery Study, Lipidomics, Metabolomics

Highlights

  • LC-MS centered biomarker studies can be displayed in five overlapping phases.

  • Interdisciplinary work is of utmost importance for successful project realization.

  • Complexity of LC-MS based studies poses unique challenges that need to be tackled.

  • Each phase of LC-MS based clinical research has significant impact on study results.

  • Data quality must be monitored and can be enhanced by quality control measures.

Abstract

Mass spectrometry focusing on small endogenous molecules has become an integral part of biomarker discovery in the pursuit of an in-depth understanding of the pathophysiology of various diseases, ultimately enabling the application of personalized medicine. While LC-MS methods allow researchers to gather vast amounts of data from hundreds or thousands of samples, the successful execution of a study as part of clinical research also requires knowledge transfer with clinicians, involvement of data scientists, and interactions with various stakeholders.

The initial planning phase of a clinical research project involves specifying the scope and design, and engaging relevant experts from different fields. Enrolling subjects and designing trials rely largely on the overall objective of the study and epidemiological considerations, while proper pre-analytical sample handling has immediate implications on the quality of analytical data. Subsequent LC-MS measurements may be conducted in a targeted, semi-targeted, or non-targeted manner, resulting in datasets of varying size and accuracy. Data processing further enhances the quality of data and is a prerequisite for in-silico analysis. Nowadays, the evaluation of such complex datasets relies on a mix of classical statistics and machine learning applications, in combination with other tools, such as pathway analysis and gene set enrichment. Finally, results must be validated before biomarkers can be used as prognostic or diagnostic decision-making tools. Throughout the study, quality control measures should be employed to enhance the reliability of data and increase confidence in the results.

The aim of this graphical review is to provide an overview of the steps to be taken when conducting an LC-MS-based clinical research project to search for small molecule biomarkers.

Introduction

Liquid chromatography – mass spectrometry (LC-MS) aided biomarker discovery is a complex and extensive process that requires interdisciplinary expertise, horizontal knowledge transfer between clinicians, analysts, data scientists, and other stakeholders (e.g., patients), and detailed planning. This process can be divided into five interwoven phases, with quality control (QC) measures supervising the other phases and facilitating confidence in findings and reproducibility. The other stages include clinical trial (including sample collection and preprocessing), sample preparation (e.g., extraction), LC-MS analyses, data processing, and evaluation. The following graphical review outlines a general workflow for LC-MS-based clinical research projects that are intended to search for small molecule biomarkers or biomarker profiles.

Study and trial design

Clinical research projects typically originate from observations of pre-clinical research or epidemiological data that suggest a need for a better understanding of pathophysiological processes, with the goal of improved diagnosis, prediction of progression, or selection of therapy in certain diseases [Fig. 1]. A plan is laid out based on literature, clinical relevance, and feasibility, including patient feedback, and funding is secured from either public organizations or industry [1]. It is important to discuss the implementation of potential biomarker screenings in terms of equipment, expertise, and capacity between end users and researchers early on. An ethics council must then evaluate the study protocol and give their approval. Before the first samples are collected, analytical scientists will develop or select the methods to be used in the project and specify the pre-analytical requirements for the compounds or compound classes of interest. Depending on the research goals, different trial designs should be evaluated, taking into account various strategies for cohort randomization (e.g., simple randomization, block randomization or stratified randomization) and blinding (e.g., double/triple blinding) [2], [3]. Meanwhile, subject recruitment and stratification will occur at the clinical site and samples will be taken and processed following predefined standard operating procedures [4]. Comprehensive subject data (e.g., age, gender, known diseases and medication) for later evaluation must be documented, while national and international standards for data privacy must be adhered to. Sampling protocols should carefully regulate sampling to control for pre-analytical variables. Pre-analytical factors are known to distort analyte levels and disrupt meaningful analysis, so emphasis should be put on sample handling that ensures sample integrity [5]. Sample documentation should include unique sample IDs, data on sample collection and the respective participant, using either unique sample-IDs or barcodes. Sample matrices, regulations in shipping biological specimens, sample stability, as well as distance between the clinical and analytical site must all be taken into account when planning the clinical research project [6], [7].

Fig. 1.

Fig. 1

Before generating any sample, a clinical study for discovering biomarkers requires careful planning, including obtaining ethics approval, defining a research scope involving stakeholders such as clinicians, analysts, data scientists, investors, and patients, and obtaining sufficient funding. Subject recruitment and stratification are instrumental to subsequent evaluations. Documenting meta data such as clinical (e.g., BMI), individual (e.g., smoking habits), and demographic information (e.g., age) on study subjects is highly relevant for trial design, as well as subsequent data analyses. The trial design depends on a multitude of factors, including modes of randomization and blinding, as well as more practical considerations such as sampling plans and study sites.

Pre-analytics - sampling and sample preprocessing

Sample preprocessing, including the generation of aliquots, should take place at the clinical site(s). Aliquot volumes should be selected appropriately to avoid unwanted freeze–thaw cycles. Further processing may also occur in the analytical laboratory before transferring samples to their final storage or starting the measurement. Special consideration should be placed on preserving sample stability in this phase as well. Analyte concentrations can be altered by simple chemical reactions, such as hydrolysis, enzymatic metabolism, or mass exchange between cells, or even the intrinsic instability of the compound of interest [8]. Therefore, the mode and temperature of intermediate storage, as well as the acceptable amount of processing delay should be predetermined. Sample handling steps for every sample need to be documented for tracking potential quality defects. For instance, in the case of anticoagulated whole blood samples, the time and temperature until centrifugation and separation of plasma, as well as centrifugation parameters and final storage temperature must be documented [Fig. 2] [9], [10], [11], [12]. In cases where the protocol of sample collection is not within the scope of the research (e.g., samples derived from biobanks), sample quality can be assessed retrospectively. Especially for blood specimens, various markers for sample quality defects have been described. For example, hemolysis in blood samples can be assessed by determining the hemolytic index [13], while other metabolomics biomarkers have been proposed for the identification of other pre-analytical risk factors, such as prolonged storage at room temperature [9], [14]. Without consensus on the best mode of detection of stability issues, documentation of sample collection remains an important tool for tracking potential quality defects. Aliquoting preprocessed matrices allows for obtaining standardized volumes from individual samples, as well as preparing technical replicates for different measurements and diluting matrices for the process of analyte extraction, which might be especially necessary for highly abundant compounds. The most important reason for generating small-volume aliquots is the prevention of additional freeze/thaw cycles, which can have strong effects on analyte concentration [15], [16]. Sample quality impairments, such as evaporation or oxidation, should be considered when generating very small aliquots. The effect of aliquot volume on the duration of thawing and subsequent stability implications should be considered as well [17]. As measurements are not necessarily performed immediately after sample preprocessing, biospecimens should be frozen at storage temperatures below −70 °C. When freeze–thaw cycles are kept to a minimum, deep freezing facilitates long-term storage and collection of sufficiently high sample numbers for performing LC-MS analytics [6]. Sample stability concerns related to the measurement should be pre-conceived during LC-MS method development and validation.

Fig. 2.

Fig. 2

Sampling should follow a predefined protocol aligned with quality standards for the analytes of focus in the study. The pre-analytical phase in the life cycle of samples has the potential to significantly alter analyte concentrations. To ensure accurate results, well-defined procedures and relevant documentation on sample processing should be in place. Matrix preprocessing should be conducted at the sampling site, as whole blood cannot be stored or transported effectively without significant ex-vivo changes; it must be processed promptly into a plasma sample. When multiple measurements or technical replicates are planned for individual samples, aliquoting is required and should be conducted directly after sampling and before dry-ice shipping and final storage, usually at < -70 °C.

Planning of LC-MS-analysis

LC-MS is able to quantify a wide variety of small molecules, including lipids or polar metabolites. The choice of methods depends largely on research requirements and the accessibility of compounds for analysis. Depending on whether the clinical study aims to answer highly specific or more general research questions, measurement can be conducted in targeted or non-targeted modes. Targeted LC-MS analysis involves the analysis of a predefined set of substances, often related by substance class or biological pathway, whereas non-targeted analysis records a broad set of mass traces that are identified and quantified later and usually result in comprehensive data sets [18]. Targeted analysis is often focused on low-concentration compounds, like lipid mediators, which require specific method optimization. LC-MS/MS is most commonly used for targeted analysis, due to its high sensitivity, selectivity, and robustness. Targeted analysis can also include absolute quantification, which requires the acquisition of reference substances and internal standards that are preferably isotopically labeled and allow for external calibration on an individual analyte level [19]. Absolute quantitative data allows for drawing conclusions regarding physiological reference intervals and simplifies data integration from multiple sites or studies. In the absence of substance-specific reference standards, relative quantification can be applied, either comparing the results to internal standards or to other samples (or reference materials) like a control group [20]. Non-targeted LC-HRMS analysis typically generates relative quantitative datasets of an extensive number of analytes, but focuses on species of higher abundance, due to a limited dynamic range [21]. Unlike targeted analysis, no predefinition of analytes is required for non-targeted approaches. To identify analytes, inferences are made from retention time and mass traces using previous experimental data and/or MS/MS libraries. To accelerate data evaluation, the analytes reviewed in a non-targeted analysis can be restricted post-acquisition to previously identified compounds, which is often referred to as semi-targeted analysis [22]. While the classical workflow of multiple reaction monitoring (MRM) allows for the adjustment of collision energy and de-clustering potential to optimize ion fragmentation and enhance sensitivity for a specific set of analytes [23], non-targeted approaches rely on a fixed combination of instrument parameters for a, generally, much broader analyte coverage [24]. Yet certain metabolites might still be elusive to the dynamic range of non-targeted setups. MS/MS approaches with a wide range of (scheduled) MRM transitions can serve as a compromise between the more reliable quantification of analytes provided by targeted analyses and the more extensive analyte coverage of non-targeted approaches [25].

Special care should be taken when planning the batch sequence for LC-MS analysis, regardless of whether it is targeted, semi-targeted, or non-targeted [Fig. 3]. Randomizing samples can mitigate the risk of bias due to signal drift [26]. Further considerations should be given to the placement of blank samples, calibration standards, QC samples, long-term control samples, and bridging samples (if applicable). Bridging samples are technical replicates of study samples between batches, which can be used to track the robustness of measurements over multiple batches. The batch size, among other factors, depends on the overall sample size, duration of measurement, and reasonable assumptions of sample stability [27].

Fig. 3.

Fig. 3

Meticulous planning of LC-MS measurements avoids pitfalls, like selection biases, and can be instrumental in quality control of the measurement. Initial batch design refers to dividing samples into different analytical runs. Different kinds of randomization can come into play, when aiming to mitigate the risk of individual samples being subjected to signal drift. While simple randomization is suitable for independent samples, a stratified cluster randomization might be more suitable for samples with nested dependence (e.g. time series data for individual subjects). Bridging refers to deliberately measuring technical replicates of samples, in order to connect batches by some proportion of shared measurements, allowing for comparability and assessment of robustness. The batch sequence (e.g., order of sample types) is subject to the mode of measurement. Depending on whether the measurement is absolute quantitative or relative quantitative, different combinations of blank samples, inter- and intra-measurement quality control samples, calibration standards and subject related unknown samples need to be considered. (Used abbreviations: U = “Sample of unknown concentration”; DB = “Double blank (without IS)”; B = “Blank (with IS)”; CS = “Calibration standard”; QC = “Quality control”; C = “Control plasma”).

The extraction of analytes from sample matrices refers to the isolation of compounds of interest while excluding or minimizing those that could interfere with the measurement. Analyte extraction can take different forms, ranging from protein precipitation, liquid–liquid extraction, to solid-phase extraction [21]. An internal or surrogate standard, which is distinguishable from the analyte yet behaves similarly in measurement (e.g., an isotopically labeled variant), is processed within the same sample. Since the amount of internal standards is held constant across samples, it is suitable for assessing analytical variability or matrix effects on a sample-by-sample basis [28]. In addition to samples, calibration standards (in the case of quantitative analysis), quality control (QC) samples, and long-term control samples are also prepared.

Targeted LC-MS-analysis

In the case of targeted and absolute quantitative LC-MS analytics, calibration standards and QC samples at different concentration levels are necessary [29]. Quantitative analysis is used to compare compound-specific calibration standards to an internal standard in order to enable the calculation of absolute analyte concentrations and to ensure confidence in compound identification. Retention time, fragmentation patterns, and peak shape of reference standards allow for drawing conclusions on analyte behavior. It is ideal to use isotopically labeled internal standards as analogues of the compound of interest. Calibration standards are derived by diluting and extracting reference substances in an analyte-free matrix to create a reference frame of concentrations (calibration curve), which can then be compared to samples of unknown concentration in order to calculate their relative concentrations [Fig. 3] [30]. Reference standard-based QC samples are processed with known concentrations spanning the range of measurable concentrations defined by the calibration curve, allowing for an assessment of accuracy and measurement system stability. In targeted analysis, these QC samples are usually placed at the beginning, intermittently at regular intervals, and at the end of the batch sequence so their accuracy and coefficient of variance can be used to determine measurement robustness and identify (and possibly correct for) drift effects within the batch [31]. Long-term control samples are similarly used as other QC samples, but concentration levels of technical replicates can be tracked in control charts over multiple measurements of the same or different studies in order to assess consistency of the LC-MS setup [32]. Thus, while regular QC samples serve as a control within the measurement, long-term control samples serve as a quality control method for the whole analytical process. Bridging samples are also technical replicates of individual samples that must be planned and prepared during sample aliquoting; one sample from each pair is measured in the batch following the first one. This is especially useful for assessing technical variations (i.e., batch effects) across multiple batches within the study. The selection for bridging samples should be done preemptively and randomized [33]. After completing the planning phase of the measuring sequence, extraction of analytes can begin. The measurement method, LC-column, and solvents should be specific to the chemical properties of the investigated analytes [Fig. 4]. Reversed phase (RP) columns are considered standard for most separation methods and are especially suitable for separation of lipophilic substances; however, polar metabolites may require Hydrophilic Interaction Liquid Chromatography (HILIC) phases in order to achieve suitable selectivity [35]. For targeted analysis, tandem mass spectrometers are often used due to their high selectivity (and wide linear range) and sensitivity, which is well-suited for analyzing a predefined set of analytes. While many targeted methods use absolute quantification, targeted screening methods also allow for analyzing hundreds of analytes without reference standards using relative quantification and a measurement setup similar to non-targeted methods [36].

Fig. 4.

Fig. 4

The process of LC-MS analysis begins with the extraction of analytes that are specific to the class. Each extraction method includes one or more purification steps. Internal or surrogate standards are added to the samples, which are processed alongside the analytes in question. This allows the signal of the analytes of interest to be related to the surrogate signal, as they are all subject to the same influences during extraction, such as matrix effects and other measurement-specific variations in sensitivity. In an often used setup, (U)HPLC is used to separate analytes, while mass spectrometry facilitates their detection. Targeted analysis using tandem mass spectrometry is limited to a predefined set of analytes, but, when combined with matching reference standards, it allows for absolute quantification. Non-targeted analysis using high resolution mass spectrometry has fewer restrictions on what analytes can be detected, allowing for simultaneous analysis of hundreds or thousands of features. However, without compound-specific reference standards, quantification is only relative. Semi-targeted analysis is a compromise between these two methods, usually focusing on hundreds of well-characterized analytes while granting broad insights at reduced processing times. As reference standards are limited in availability and economical viability, semi-targeted analysis is mostly only relative quantitative.

Semi- and Non-targeted LC-MS-analysis

In non-targeted analysis, features with unique combinations of mass-to-charge and retention time are aligned, annotated, and compared based on relative quantification. High resolution mass spectrometry is typically used to allow for compound annotation or identification based on exact mass, isotopic distribution, (relative) retention time, and by comparison of MS/MS fragmentation patterns to reference spectra [24]. This expands the range of potentially observable analytes compared to targeted analysis; however, it also increases the chance of false identification. Quantification is limited by the fact that many analytes are related to only few surrogate internal standards or relative sample concentrations are calculated in relation to all other samples. The signal intensity of compounds can vary based on carbon chain length, the number of double bonds, and structural differences [37]. The amount of relatively quantified features can easily reach into the thousands; therefore, internal standards cannot be selected on an individual analyte level. Absolute quantitative analysis with reference standard calibration is also not suitable for the same reasons. Long-term control samples can still be used for tracking system performance. QC samples, used for assessing measuring robustness within an analysis run, also do not require a range of concentrations; rather, they are replicates from a representative sample pool distributed over the batch at regular intervals (e.g., every 8 to 20 samples).

Data integration and processing

Signal interpretation uses specialized software, either specific to the manufacturer or offered by third-party vendors, to identify and integrate signal peaks for each analyte and its respective internal standard. This is instrumental in calculating the concentration, as it relies on the area under the curve. The evaluation software usually automates significant proportions of peak annotation and integration when referencing a database of earlier measurements [38]. Exceptions include non-targeted analysis, where it might not be possible to find a suitable internal standard for all features, yet surrogates are still used to verify technical variation. Absolute concentrations are calculated using a calibration curve [30] and values that deviate from the line should be deemed indistinct and may be imputed in later steps. QC samples are utilized to evaluate measurement robustness, a key parameter in biomarker discovery [39]. These concentrations should align with other QC samples (e.g., pooled QC samples), and their accuracy and coefficient of variance serves as a measure of quality for the analysis run [40], [41]. Despite this, the resulting dataset might still be skewed and can be transformed further depending on its planned application [Fig. 5]. Missing values can be imputed, although the choice of method usually requires knowledge as to why the values are missing [42]. Erratically missing values may be imputed by more sophisticated methods like kNN or Random Forest Imputation [43], while values that fall below the lower limit of quantification can often be imputed with half of said limit, thus accounting for their approximate scale. Additionally, sample data can be normalized and scaled by various methods in order to remove systematic and statistical biases. Transformations towards normal distribution of measurements can be performed to satisfy the assumptions of statistical and machine learning models [34]. Normalization techniques include approaches of general applicability that aim for comparability of samples (e.g., Probabilistic Quotient Normalization [44] or Quantile Normalization [45]), as well as techniques that target specific biases like batch effect correction methods (e.g., LOESS normalization [46], SVA [47] or ComBat [48], [49]). HRMS datasets usually take up large amounts of hard drive storage (e.g., dozens of gigabytes per 100 samples), and have very high complexity (e.g., randomization, data hierarchy, time series). Therefore, data management and archiving in LC-MS, as well as other study data, is crucial for traceability of the study.

Fig. 5.

Fig. 5

Peak identification depends on the mass traces and retention time of the respective analytes. Peak integration enables relative quantification. Relative quantification (non-targeted analysis) compares the peak areas of multiple analytes with one or multiple internal standards or a reference sample, and does not use a calibration curve of reference standards, which is necessary for targeted measurements. Further data processing might be needed after obtaining the results. Missing values must be imputed, outliers potentially removed, samples may need to be scaled for comparability, normalized to meet the assumptions of statistical models, and batch effects must be identified and corrected retrospectively.

Data analysis and Clinical translation

Data analysis specific to the clinical scope of the study requires domain knowledge when pursuing specific questions. Before cohorts are evaluated, data sets may need to be adjusted for covariates (e.g., age distribution among groups) to avoid confounding [50], [51]. A mix of classical statistics and machine learning applications is employed for data investigation [Fig. 6] [52], with classical statistics focusing on explaining phenomena, while machine learning methods focus on prediction [53]. Unsupervised machine learning algorithms offer a general way of learning underlying structures of datasets; principal component analysis (PCA) and hierarchical clustering are instrumental in data exploration. Besides information on patterns within the data, they are able to offer techniques of variable selection (i.e., dimensionality reduction), which is crucial in identifying potential biomarkers from hundreds or thousands of measured analytes. Supervised machine learning algorithms can put selected variables to the test and rely on powerful regression and classification algorithms, like the Random Forest (RF) or Support Vector Machines (SVM) [54]. Model validation, such as Cross-validation, must be facilitated to assess supervised machine learning model performance and exclude model overfitting [55]. Optimization of supervised machine learning models is usually performed through a process of model parameter tuning and model assessment [54]. Pathway analysis helps to embed findings in a broader biological context and form hypotheses and make causal assumptions [56].

Fig. 6.

Fig. 6

Data analysis can take place between the fields of classical statistics and machine learning, which often overlap. Statistics obtain p-values, usually in uni- or bivariate analysis, to claim significance, while machine learning approaches strive for the best possible test metrics, such as optimal prediction accuracy. Machine learning can be further divided into unsupervised and supervised approaches. Unsupervised approaches act independently of pre-defined sample groups to develop hypotheses and explore data (although inferences are also possible). In contrast, supervised models usually aim for optimal regression or classification in the context of the research scope. Therefore, both supervised models and classical statistics can be used to confirm or reject hypotheses. Pathway analyses can integrate the obtained data into a broader biological context. Finally, resulting findings and potential biomarkers or biomarker panels need to be clinically validated, demonstrate clinical benefit, and be practical in order to be translated into clinical routine.

Besides analytical and clinical validity, biomarkers or biomarker panels need to prove their utility by being practical and broadly applicable in a clinical context. However, validating biomarkers poses methodological and general challenges. A common problem in initial biomarker research is the lack of power due to small sample sizes, which increases the risk for confounders and reduces the applicability of machine learning methods. A high number of potential biomarkers, generated from a small sample size can thus be a result of overfitting [57]. Awareness of this is of particular interest, as there is no consensus on the best way of feature selection and machine learning methods are often selected for isolating a set of variables [58]. Furthermore, metabolites are especially prone to vary due to various covariates, such as age, gender, and diet. Therefore, if feature selection steps result in a high number of analytes, special caution needs to be taken to avoid such biases (e.g., by stratifying the cohort properly). Additionally, due to the complexity of many diseases and subsequent systemic changes, individual biomarkers are unlikely to emerge for diagnostic or prognostic purposes [59]. To overcome this issue, validation calls for increasing the sample size with an independent cohort to confirm selectivity, stability, and repeatability of the measurement and raise statistical power [60]. Moreover, there needs to be an emphasis on the utility of potential biomarkers for their end users throughout the research process. Insufficient interdisciplinary exchange can obscure the lack of clinical relevance, which is not only influenced by predictive power, but also by cost-effectiveness in the clinical laboratory (in terms of instruments, work space, expertise, and staff allocation) and benefit for physicians and their patients in making therapy decisions [61]. While many biomarkers fail at this stage, those that are successful can improve clinical decision making and/or elucidate novel therapeutic targets [46], [47].

Quality control & quality assurance

Published potential biomarkers found in LC-MS studies routinely fail in clinical translation, due to inadequate reporting, low quality control, low reproducibility and overfitting of models to available data sets [62]. Other issues that hinder the implementation of valuable biomarkers include biases in cohort selection and evaluation, as well as insufficient independent validation of research results. Quality management measures should be put in place to provide quality control (QC) and quality assurance (QA) throughout the entire research process, so as to improve the reliability and reproducibility of the data and experiments [Fig. 7]. This can be achieved by providing specific instructions for each step of the analytical process and requiring frequent documentation, which will enable traceability [36]. Documentation for quality control purposes should also include recording metadata that describe the study results. To enhance reproducibility across studies, national and international guidelines should be developed to harmonize quality control protocols [63].

Fig. 7.

Fig. 7

Quality control and quality assurance accompany the entire process of clinical research projects. Standard operating procedures, as well as good documentation practices, are essential for quality assurance in all stages of research. Data management of clinical data, metadata, data on (pre-)processing, analysis, and evaluation of results allows for traceability across the entire study. Clinical data, documentation on sample processing, analytical phases, and detailed reporting on the data analysis facilitate comprehension of the results and are necessary for reproducibility and validity of the study.

Funding

This work was supported by the Deutsche Forschungsgemeinschaft Sonderforschungsbereich SFB 1039/Z01 “Krankheitsrelevante Signaltransduktion durch Fettsäurederivate und Sphingolipide”, by Fraunhofer Cluster of Excellence for Immune Mediated diseases CIMD and the HIPPOCRATES project. HIPPOCRATES has received funding from the Innovative Medicines Initiative 2 Joint Undertaking (JU) under grant agreement no. 101007757. The JU receives support from the European Union’s Horizon 2020 research and innovation program and EFPIA. BB is supported by grants from the National University of Singapore via the Life Sciences Institute.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Footnotes

Appendix A

Supplementary data to this article can be found online at https://doi.org/10.1016/j.jmsacl.2023.02.003.

Contributor Information

R. Gurke, Email: gurke@med.uni-frankfurt.de.

T.J. Garrett, Email: tgarrett@ufl.edu.

Appendix A. Supplementary data

The following are the Supplementary data to this article:

Supplementary Data 1
mmc1.pptx (2.1MB, pptx)

References

  • 1.Bhandari, Mohit M.D., M.SC., F.R.C.S.C.*, Schemitsch, Emil H. M.D., F.R.C.S.C.†, Planning a Randomized Clinical Trial - An Overview, Techniques in Orthopaedics 19 (2004) 66–71.
  • 2.Broglio K. Randomization in Clinical Trials: Permuted Blocks and Stratification. J. Am. Med. Assoc. 2018;219:2223–2224. doi: 10.1001/jama.2018.6360. [DOI] [PubMed] [Google Scholar]
  • 3.Hróbjartsson A., Boutron I. Blinding in randomized clinical trials: imposed impartiality. Clin. Pharmacol. Ther. 2011;90:732–736. doi: 10.1038/clpt.2011.207. [DOI] [PubMed] [Google Scholar]
  • 4.Schmidt R.L., Factor R.E. Understanding sources of bias in diagnostic accuracy studies. Arch. Pathol. Lab. Med. 2013;137:558–565. doi: 10.5858/arpa.2012-0198-RA. [DOI] [PubMed] [Google Scholar]
  • 5.Lehmann R. From bedside to bench-practical considerations to avoid pre-analytical pitfalls and assess sample quality for high-resolution metabolomics and lipidomics analyses of body fluids. Anal. Bioanal. Chem. 2021 doi: 10.1007/s00216-021-03450-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Lippi G., Chance J.J., Church S., Dazzi P., Fontana R., Giavarina D., Grankvist K., Huisman W., Kouri T., Palicka V., Plebani M., Puro V., Salvagno G.L., Sandberg S., Sikaris K., Watson I., Stankovic A.K., Simundic A.-M. Preanalytical quality improvement: from dream to reality. Clin. Chem. Lab. Med. 2011;49:1113–1126. doi: 10.1515/CCLM.2011.600. [DOI] [PubMed] [Google Scholar]
  • 7.Hyötyläinen T., Orešič M. Bioanalytical techniques in nontargeted clinical lipidomics. Bioanalysis. 2016;8:351–364. doi: 10.4155/bio.15.244. [DOI] [PubMed] [Google Scholar]
  • 8.Nishiumi S., Suzuki M., Kobayashi T., Yoshida M. Differences in metabolite profiles caused by pre-analytical blood processing procedures. J. Biosci. Bioeng. 2018;125:613–618. doi: 10.1016/j.jbiosc.2017.11.011. [DOI] [PubMed] [Google Scholar]
  • 9.Kamlage B., Maldonado S.G., Bethan B., Peter E., Schmitz O., Liebenberg V., Schatz P. Quality markers addressing preanalytical variations of blood and plasma processing identified by broad and targeted metabolite profiling. Clin. Chem. 2014;60:399–412. doi: 10.1373/clinchem.2013.211979. [DOI] [PubMed] [Google Scholar]
  • 10.Vuckovic D. Current trends and challenges in sample preparation for global metabolomics using liquid chromatography-mass spectrometry. Anal. Bioanal. Chem. 2012;403:1523–1548. doi: 10.1007/s00216-012-6039-y. [DOI] [PubMed] [Google Scholar]
  • 11.Liebisch G. Biochimica et biophysica acta. BBA. 2017;62:636. doi: 10.1016/0006-3002(62)90265-2. [DOI] [Google Scholar]
  • 12.Naz S., Moreira dos Santos D.C., García A., Barbas C. Analytical protocols based on LC-MS, GC-MS and CE-MS for nontargeted metabolomics of biological tissues. Bioanalysis. 2014;6:1657–1677. doi: 10.4155/bio.14.119. [DOI] [PubMed] [Google Scholar]
  • 13.Lippi G., Favaloro E.J., Franchini M. Haemolysis index for the screening of intravascular haemolysis: a novel diagnostic opportunity? Blood Transfus. 2018;16:433–437. doi: 10.2450/2018.0045-18. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Liu X., Hoene M., Yin P., Fritsche L., Plomgaard P., Hansen J.S., Nakas C.T., Niess A.M., Hudemann J., Haap M., Mendy M., Weigert C., Wang X., Fritsche A., Peter A., Häring H.-U., Xu G., Lehmann R. Quality Control of Serum and Plasma by Quantification of (4E,14Z)-Sphingadienine-C18-1-Phosphate Uncovers Common Preanalytical Errors During Handling of Whole Blood. Clin. Chem. 2018;64:810–819. doi: 10.1373/clinchem.2017.277905. [DOI] [PubMed] [Google Scholar]
  • 15.Fliniaux O., Gaillard G., Lion A., Cailleu D., Mesnard F., Betsou F. Influence of common preanalytical variations on the metabolic profile of serum samples in biobanks. J. Biomol. NMR. 2011;51:457–465. doi: 10.1007/s10858-011-9574-5. [DOI] [PubMed] [Google Scholar]
  • 16.S. Cuhadar, M. Koseoglu, A. Atay, A. Dirican, The effect of storage time and freeze-thaw cycles on the stability of serum samples, Biochem. Med. (Zagreb) 23 (2013) 70–77. 10.11613/bm.2013.009. [DOI] [PMC free article] [PubMed]
  • 17.Pizarro C., Arenzana-Rámila I., Pérez-del-Notario N., Pérez-Matute P., González-Sáiz J.M. Thawing as a critical pre-analytical step in the lipidomic profiling of plasma samples: New standardized protocol. Anal. Chim. Acta. 2016;912:1–9. doi: 10.1016/j.aca.2016.01.058. [DOI] [PubMed] [Google Scholar]
  • 18.Reveglia P., Paolillo C., Ferretti G., de Carlo A., Angiolillo A., Nasso R., Caputo M., Matrone C., Di Costanzo A., Corso G. Challenges in LC-MS-based metabolomics for Alzheimer's disease early detection: targeted approaches versus untargeted approaches. Metabolomics. 2021;17:78. doi: 10.1007/s11306-021-01828-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Köfeler H.C., Eichmann T.O., Ahrends R., Bowden J.A., Danne-Rasche N., Dennis E.A., Fedorova M., Griffiths W.J., Han X., Hartler J., Holčapek M., Jirásko R., Koelmel J.P., Ejsing C.S., Liebisch G., Ni Z., O'Donnell V.B., Quehenberger O., Schwudke D., Shevchenko A., Wakelam M.J.O., Wenk M.R., Wolrab D., Ekroos K. Quality control requirements for the correct annotation of lipidomics data. Nat. Commun. 2021;12:4771. doi: 10.1038/s41467-021-24984-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Triebl A., Burla B., Selvalatchmanan J., Oh J., Tan S.H., Chan M.Y., Mellet N.A., Meikle P.J., Torta F., Wenk M.R. Shared reference materials harmonize lipidomics across MS-based detection platforms and laboratories. J. Lipid Res. 2020;61:105–115. doi: 10.1194/jlr.D119000393. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Petras D., Koester I., Da Silva R., Stephens B.M., Haas A.F., Nelson C.E., Kelly L.W., Aluwihare L.I., Dorrestein P.C. High-Resolution Liquid Chromatography Tandem Mass Spectrometry Enables Large Scale Molecular Characterization of Dissolved Organic Matter. Front. Mar. Sci. 2017;4:405. doi: 10.3389/fmars.2017.00405. [DOI] [Google Scholar]
  • 22.Onorato J.M., Langish R., Bellamine A., Shipkova P. Applications of HILIC for targeted and non-targeted LC/MS analyses in drug discovery. J. Sep. Sci. 2010;33:923–929. doi: 10.1002/jssc.200900659. [DOI] [PubMed] [Google Scholar]
  • 23.Zhou J., Yin Y. Strategies for large-scale targeted metabolomics quantification by liquid chromatography-mass spectrometry. Analyst. 2016;141:6362–6373. doi: 10.1039/c6an01753c. [DOI] [PubMed] [Google Scholar]
  • 24.Ribbenstedt A., Ziarrusta H., Benskin J.P. Development, characterization and comparisons of targeted and non-targeted metabolomics methods. PLoS One. 2018;13:e0207082. doi: 10.1371/journal.pone.0207082. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Munjoma N., Isaac G., Muazzam A., Cexus O., Azhar F., Pandha H., Whetton A.D., Townsend P.A., Wilson I.D., Gethings L.A., Plumb R.S. High Throughput LC-MS Platform for Large Scale Screening of Bioactive Polar Lipids in Human Plasma and Serum. J. Proteome Res. 2022;21:2596–2608. doi: 10.1021/acs.jproteome.2c00297. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Jonsson P., Wuolikainen A., Thysell E., Chorell E., Stattin P., Wikström P., Antti H. Constrained randomization and multivariate effect projections improve information extraction and biomarker pattern discovery in metabolomics studies involving dependent samples. Metabolomics. 2015;11:1667–1678. doi: 10.1007/s11306-015-0818-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Evans A.M., O'Donovan C., Playdon M., Beecher C., Beger R.D., Bowden J.A., Broadhurst D., Clish C.B., Dasari S., Dunn W.B., Griffin J.L., Hartung T., Hsu P.-C., Huan T., Jans J., Jones C.M., Kachman M., Kleensang A., Lewis M.R., Monge M.E., Mosley J.D., Taylor E., Tayyari F., Theodoridis G., Torta F., Ubhi B.K., Vuckovic D. Dissemination and analysis of the quality assurance (QA) and quality control (QC) practices of LC-MS based untargeted metabolomics practitioners. Metabolomics. 2020;16:113. doi: 10.1007/s11306-020-01728-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Wilkes E.H., Whitlock M.J., Williams E.L. A data-driven approach for the detection of internal standard outliers in targeted LC-MS/MS assays. J. Mass Spectrom. Adv. Clin. Lab. 2021;20:42–47. doi: 10.1016/j.jmsacl.2021.06.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.America A.H.P., Cordewener J.H.G. Comparative LC-MS: a landscape of peaks and valleys. Proteomics. 2008;8:731–749. doi: 10.1002/pmic.200700694. [DOI] [PubMed] [Google Scholar]
  • 30.Zhao Y., Liu G., Shen J.X., Aubry A.-F. Reasons for calibration standard curve slope variation in LC-MS assays and how to address it. Bioanalysis. 2014;6:1439–1443. doi: 10.4155/bio.14.71. [DOI] [PubMed] [Google Scholar]
  • 31.Briscoe C.J., Stiles M.R., Hage D.S. System suitability in bioanalytical LC/MS/MS. J. Pharm. Biomed. Anal. 2007;44:484–491. doi: 10.1016/j.jpba.2007.03.003. [DOI] [PubMed] [Google Scholar]
  • 32.Broadhurst D., Goodacre R., Reinke S.N., Kuligowski J., Wilson I.D., Lewis M.R., Dunn W.B. Guidelines and considerations for the use of system suitability and quality control samples in mass spectrometry assays applied in untargeted clinical metabolomic studies. Metabolomics. 2018;14:72. doi: 10.1007/s11306-018-1367-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Buhule O.D., Minster R.L., Hawley N.L., Medvedovic M., Sun G., Viali S., Deka R., McGarvey S.T., Weeks D.E. Stratified randomization controls better for batch effects in 450K methylation analysis: a cautionary tale. Front. Genet. 2014;5:354. doi: 10.3389/fgene.2014.00354. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Ortmayr K., Causon T.J., Hann S., Koellensperger G. Increasing selectivity and coverage in LC-MS based metabolome analysis. TrAC Trends Anal. Chem. 2016;82:358–366. doi: 10.1016/j.trac.2016.06.011. [DOI] [Google Scholar]
  • 35.Harrieder E.-M., Kretschmer F., Böcker S., Witting M. Current state-of-the-art of separation methods used in LC-MS based metabolomics and lipidomics. J. Chromatogr. B Anal. Technol. Biomed. Life Sci. 2022;1188 doi: 10.1016/j.jchromb.2021.123069. [DOI] [PubMed] [Google Scholar]
  • 36.Burla B., Arita M., Arita M., Bendt A.K., Cazenave-Gassiot A., Dennis E.A., Ekroos K., Han X., Ikeda K., Liebisch G., Lin M.K., Loh T.P., Meikle P.J., Orešič M., Quehenberger O., Shevchenko A., Torta F., Wakelam M.J.O., Wheelock C.E., Wenk M.R. MS-based lipidomics of human blood plasma: a community-initiated position paper to develop accepted guidelines. J. Lipid Res. 2018;59:2001–2017. doi: 10.1194/jlr.S087163. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Schymanski E.L., Singer H.P., Slobodnik J., Ipolyi I.M., Oswald P., Krauss M., Schulze T., Haglund P., Letzel T., Grosse S., Thomaidis N.S., Bletsou A., Zwiener C., Ibáñez M., Portolés T., de Boer R., Reid M.J., Onghena M., Kunkel U., Schulz W., Guillon A., Noyon N., Leroy G., Bados P., Bogialli S., Stipaničev D., Rostkowski P., Hollender J. Non-target screening with high-resolution mass spectrometry: critical review using a collaborative trial on water analysis. Anal. Bioanal. Chem. 2015;407:6237–6255. doi: 10.1007/s00216-015-8681-7. [DOI] [PubMed] [Google Scholar]
  • 38.Gorrochategui E., Jaumot J., Lacorte S., Tauler R. Data analysis strategies for targeted and untargeted LC-MS metabolomic studies: Overview and workflow. TrAC Trends Anal. Chem. 2016;82:425–442. doi: 10.1016/j.trac.2016.07.004. [DOI] [Google Scholar]
  • 39.Gika H.G., Theodoridis G.A., Earll M., Wilson I.D. A QC approach to the determination of day-to-day reproducibility and robustness of LC-MS methods for global metabolite profiling in metabonomics/metabolomics. Bioanalysis. 2012;4:2239–2247. doi: 10.4155/bio.12.212. [DOI] [PubMed] [Google Scholar]
  • 40.Kaza M., Karaźniewicz-Łada M., Kosicka K., Siemiątkowska A., Rudzki P.J. Bioanalytical method validation: new FDA guidance vs. EMA guideline. Better or worse? J. Pharm. Biomed. Anal. 2019;165:381–385. doi: 10.1016/j.jpba.2018.12.030. [DOI] [PubMed] [Google Scholar]
  • 41.Ghorasaini M., Mohammed Y., Adamski J., Bettcher L., Bowden J.A., Cabruja M., Contrepois K., Ellenberger M., Gajera B., Haid M., Hornburg D., Hunter C., Jones C.M., Klein T., Mayboroda O., Mirzaian M., Moaddel R., Ferrucci L., Lovett J., Nazir K., Pearson M., Ubhi B.K., Raftery D., Riols F., Sayers R., Sijbrands E.J.G., Snyder M.P., Su B., Velagapudi V., Williams K.J., de Rijke Y.B., Giera M. Cross-Laboratory Standardization of Preclinical Lipidomics Using Differential Mobility Spectrometry and Multiple Reaction Monitoring. Anal. Chem. 2021;93:16369–16378. doi: 10.1021/acs.analchem.1c02826. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Lin W.-C., Tsai C.-F. Missing value imputation: a review and analysis of the literature (2006–2017) Artif. Intell. Rev. 2020;53:1487–1509. doi: 10.1007/s10462-019-09709-4. [DOI] [Google Scholar]
  • 43.Wei R., Wang J., Su M., Jia E., Chen S., Chen T., Ni Y. Missing Value Imputation Approach for Mass Spectrometry-based Metabolomics Data. Sci. Rep. 2018;8:663. doi: 10.1038/s41598-017-19120-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Di Guida R., Engel J., Allwood J.W., Weber R.J.M., Jones M.R., Sommer U., Viant M.R., Dunn W.B. Non-targeted UHPLC-MS metabolomic data processing methods: a comparative investigation of normalisation, missing value imputation, transformation and scaling. Metabolomics. 2016;12:93. doi: 10.1007/s11306-016-1030-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Lee J., Park J., Lim M., Seong S.J., Seo J.J., Park S.M., Lee H.W., Yoon Y.-R. Quantile normalization approach for liquid chromatography-mass spectrometry-based metabolomic data from healthy human volunteers. Anal. Sci. 2012;28:801–805. doi: 10.2116/analsci.28.801. [DOI] [PubMed] [Google Scholar]
  • 46.Rusilowicz M., Dickinson M., Charlton A., O'Keefe S., Wilson J. A batch correction method for liquid chromatography-mass spectrometry data that does not depend on quality control samples. Metabolomics. 2016;12:56. doi: 10.1007/s11306-016-0972-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Leek J.T., Johnson W.E., Parker H.S., Jaffe A.E., Storey J.D. The sva package for removing batch effects and other unwanted variation in high-throughput experiments. Bioinformatics. 2012;28:882–883. doi: 10.1093/bioinformatics/bts034. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Johnson W.E., Li C., Rabinovic A. Adjusting batch effects in microarray expression data using empirical Bayes methods. Biostatistics. 2007;8:118–127. doi: 10.1093/biostatistics/kxj037. [DOI] [PubMed] [Google Scholar]
  • 49.Zhao Y., Wong L., Goh W.W.B. How to do quantile normalization correctly for gene expression data analyses. Sci. Rep. 2020;10:15534. doi: 10.1038/s41598-020-72664-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Witte J., Didelez V. Covariate selection strategies for causal inference: Classification and comparison. Biom. J. 2019;61:1270–1289. doi: 10.1002/bimj.201700294. [DOI] [PubMed] [Google Scholar]
  • 51.Srivastava A., Creek D.J. Discovery and Validation of Clinical Biomarkers of Cancer: A Review Combining Metabolomics and Proteomics. Proteomics. 2019;19:e1700448. doi: 10.1002/pmic.201700448. [DOI] [PubMed] [Google Scholar]
  • 52.Breimann L., Statistical modelin: The two cultures (with comments and a rejoinder by the author), Statistical Science 16 199–231.
  • 53.Boulesteix A.-L., Schmid M. Machine learning versus statistical modeling. Biom. J. 2014;56(4):588–593. doi: 10.1002/bimj.201300226. [DOI] [PubMed] [Google Scholar]
  • 54.Kingsmore K.M., Puglisi C.E., Grammer A.C., Lipsky P.E. An introduction to machine learning and analysis of its use in rheumatic diseases. Nat. Rev. Rheumatol. 2021;17:710–730. doi: 10.1038/s41584-021-00708-w. [DOI] [PubMed] [Google Scholar]
  • 55.Maleki F., Muthukrishnan N., Ovens K., Reinhold C., Forghani R. Machine Learning Algorithm Validation: From Essentials to Advanced Applications and Implications for Regulatory Certification and Deployment. Neuroimaging Clin. N. Am. 2020;30:433–445. doi: 10.1016/j.nic.2020.08.004. [DOI] [PubMed] [Google Scholar]
  • 56.Wörheide M.A., Krumsiek J., Kastenmüller G., Arnold M. Multi-omics integration in biomedical research - A metabolomics-centric review. Anal. Chim. Acta. 2021;1141:144–162. doi: 10.1016/j.aca.2020.10.038. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Long N.P., Yoon S.J., Anh N.H., Nghi T.D., Lim D.K., Hong Y.J., Hong S.-S., Kwon S.W. A systematic review on metabolomics-based diagnostic biomarker discovery and validation in pancreatic cancer. Metabolomics. 2018;14:109. doi: 10.1007/s11306-018-1404-2. [DOI] [PubMed] [Google Scholar]
  • 58.Vladimir Tolstikov, A. James Moser, Rangaprasad Sarangarajan, Niven R. Narain and Michael A. Kiebish, Current Status of Metabolomic Biomarker Discovery: Impact of Study Design and Demographic Characteristics. [DOI] [PMC free article] [PubMed]
  • 59.Smith B.J., Silva-Costa L.C., Martins-de-Souza D. Human disease biomarker panels through systems biology. Biophys. Rev. 2021;13:1179–1190. doi: 10.1007/s12551-021-00849-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Nakayasu E.S., Gritsenko M., Piehowski P.D., Gao Y., Orton D.J., Schepmoes A.A., Fillmore T.L., Frohnert B.I., Rewers M., Krischer J.P., Ansong C., Suchy-Dicey A.M., Evans-Molina C., Qian W.-J., Webb-Robertson B.-J.-M., Metz T.O. Tutorial: best practices and considerations for mass-spectrometry-based protein biomarker discovery and validation. Nat. Protoc. 2021;16:3737–3760. doi: 10.1038/s41596-021-00566-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Marchand C.R., Farshidfar F., Rattner J., Bathe O.F. A Framework for Development of Useful Metabolomic Biomarkers and Their Effective Knowledge Translation. Metabolites. 2018;8 doi: 10.3390/metabo8040059. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Ioannidis J.P.A., Bossuyt P.M.M. Waste, Leaks, and Failures in the Biomarker Pipeline. Clin. Chem. 2017;63:963–972. doi: 10.1373/clinchem.2016.254649. [DOI] [PubMed] [Google Scholar]
  • 63.Köfeler H.C., Ahrends R., Baker E.S., Ekroos K., Han X., Hoffmann N., Holčapek M., Wenk M.R., Liebisch G. Recommendations for good practice in MS-based lipidomics. J. Lipid Res. 2021;62 doi: 10.1016/j.jlr.2021.100138. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary Data 1
mmc1.pptx (2.1MB, pptx)

Articles from Journal of Mass Spectrometry and Advances in the Clinical Lab are provided here courtesy of Elsevier

RESOURCES