Working with human samples |
-
•
Recruitment: enrolling patients with diverse backgrounds from multiple transplant centres remains a challenge
-
•
Small sample size: acquiring a large number of study participants is necessary for significant findings
-
•
Clinical data organisation: collecting, storing, and sorting a significant volume of longitudinal patient data can be difficult to manage
-
•
Study timeline: retention of patients during follow-up is costly, time-consuming, and difficult
-
•
Variable post-transplant drug regimens: immunosuppressive treatment, dosage, and management may change over time, presenting a confounding variable that is difficult to adjust for
|
Study Design |
-
•
Lack of standardised study endpoints: patient/graft survival, acute rejection episodes, biochemical data (i.e., liver enzymes), disease recurrence and development, and non-immunological injury (nephrotoxicity) must all be assessed in relation to post-transplant outcomes
-
•
Lack of control group: post-transplant control groups are difficult to define
-
•
Identification of best markers: there is no standardised method for narrowing down the top candidate markers when using high-throughput approaches
|
Technical challenges |
-
•
Isolating nucleic acids: circulating nucleic acids are characterised by low concentration and high degradation rates
-
•
Experimental methods: utilising high throughput technologies can be costly and time-consuming
-
•
Normalisation methods: multiple approaches can be used to normalise cell-free nucleic acids
|
Building predictive models |
-
•
Team science: biostatisticians are needed to aid in the development and implementation of statistical and mathematical methods
-
•
Overfitting: over-testing the training data can result in a model that appears very accurate but has memorised the key points in the data set rather than learned how to generalise, requiring an independent dataset to validate findings.
|