Skip to main content
PLOS Neglected Tropical Diseases logoLink to PLOS Neglected Tropical Diseases
. 2024 Oct 14;18(10):e0012300. doi: 10.1371/journal.pntd.0012300

Diagnostic serology test comparison for Q fever and Rift Valley fever in humans and livestock from pastoral communities

Valerie Hungerbühler 1,*, Ranya Özcelik 1, Mahamat Fayiz Abakar 2, Fatima Abdelrazak Zakaria 2, Martin Eiden 3, Sonja Hartnack 4, Pidou Kimala 2, Sonja Kittl 5, Janine Michel 6, Franziska Suter-Riniker 7, Salome Dürr 1
Editor: Philip Mshelbwala8
PMCID: PMC11501034  PMID: 39401261

Abstract

Q fever (QF) and Rift Valley fever (RVF) are endemic zoonotic diseases in African countries, causing significant health and economic burdens. Accurate prevalence estimates, crucial for disease control, rely on robust diagnostic tests. While enzyme-linked immunosorbent assays (ELISA) are not the gold standard, they offer rapid, cost-effective, and practical alternatives. However, varying results from different tests and laboratories can complicate comparing epidemiological studies. This study aimed to assess the agreement of test results for QF and RVF in humans and livestock across different laboratory conditions and, for humans, different types of diagnostic tests. We measured inter-laboratory agreement using concordance, Cohen’s kappa, and prevalence and bias-adjusted kappa (PABAK) on 91 human and 102 livestock samples collected from rural regions in Chad. The serum aliquots were tested using ELISA in Chad, and indirect immunofluorescence assay (IFA) (for human QF and RVF) and ELISA (for livestock QF and RVF) in Switzerland and Germany. Additionally, we examined demographic factors influencing test agreement, including district, setting (village vs. camp), sex, age, and livestock species of the sampled individuals. The inter-laboratory agreement ranged from fair to moderate. For humans, QF concordance was 62.5%, Cohen’s kappa was 0.31, RVF concordance was 81.1%, and Cohen’s kappa was 0.52. For livestock, QF concordance was 92.3%, Cohen’s kappa was 0.59, RVF concordance was 94.0%, and Cohen’s kappa was 0.59. Multivariable analysis revealed that QF test agreement is significantly higher in younger humans and people living in villages compared to camps and tends to be higher in livestock from Danamadji compared to Yao, and in small ruminants compared to cattle. Additionally, RVF agreement was found to be higher in younger humans. Our findings emphasize the need to consider sample conditions, test performance, and influencing factors when conducting and interpreting epidemiological seroprevalence studies.

Author summary

Q fever (QF) and Rift Valley fever (RVF) are zoonotic diseases that can be transmitted from animals to humans, causing health problems and economic losses in African countries. While various diagnostic tests for these diseases are available, they can be impractical, especially in resource-limited settings.

For this study, human and livestock samples from Chad were first tested in a local laboratory using a routine test (enzyme-linked immunosorbent assay). Serum aliquots were then sent to laboratories in Germany or Switzerland for retesting, using the same test type for livestock and a different test type for human samples.

We analysed the agreement between the test results and investigated the influence of the demographic characteristics of the sampled individual on this agreement. Our findings are crucial as they reveal discrepancies in test results, even though the samples originated from the same individuals. Additionally, we found that factors such as the age of the sampled individual influenced test agreement.

This study underscores the importance of considering sample conditions, test performance, and influencing factors when conducting and interpreting disease prevalence studies. Enhancing diagnostic procedures will aid in more effective disease control management, benefiting local communities and global health efforts.

1. Introduction

Q Fever (QF) and Rift Valley fever (RVF) are zoonotic diseases prevalent in several African countries [1]. Reported prevalence rates range from 7.8% to 39% for QF and 9.5% to 44.2% for RVF in livestock, and from 27% to 49.2% for QF and 13.2% to 28.4% for RVF in humans [14]. QF and RVF impact human health by causing a flu-like syndrome that can lead to a range of severe clinical manifestations. QF and RVF also result in significant production losses in animals due to abortions [5,6].

High-quality samples and robust diagnostic tests are essential for obtaining accurate prevalence estimates. Epidemiological studies play a critical role in generating the necessary data, subsequently influencing government prioritization of health interventions [7,8]. This prioritization is fundamental to effective disease control.

For QF diagnostics, the indirect immunofluorescence assay (IFA) can differentiate between acute and chronic infections and is regarded as the gold standard test for humans [9,10]. However, commercial IFA kits are not available for veterinary use [11]. The enzyme-linked immunosorbent assay (ELISA) is the most widely used test and is recommended by the WOAH (World Organisation for Animal Health) for rapid routine screening and large-scale epidemiological studies in ruminant populations [7]. For RVF, the virus neutralization test (VNT) is the most specific serological diagnostic test [8]. However, it requires living cell cultures, titered virus stocks, and highly trained personnel, making it labour-intensive, costly, and time-consuming (minimum 48 hours) [12]. Additionally, working with viable viruses poses a biohazard risk, making VNT unsuitable for use in laboratories without appropriate biosecurity facilities and vaccinated personnel [6,8]. For both diseases, ELISAs offer a rapid, cost-effective, and practical alternative with less stringent biosafety requirements, making them suitable for routine use in low- and middle-income countries (LMIC) [7,8].

The use of ELISA in some studies and the gold standard test, which may differ between countries, in others can lead to discrepancies in estimated prevalence, making comparisons challenging; thus, harmonized monitoring and reporting schemes for QF and RVF have been proposed to enable consistent comparisons over time and across countries [1315].

Several studies have assessed the inter-test agreement of ELISA for QF and RVF compared to other diagnostic tests, reporting variable agreement ranging from poor to good for QF and from good to excellent for RVF [1622]. Diagnostic test validation can be achieved through various methods, including assessing the agreement between different tests without assuming one as the gold standard [23]. Concordance, the proportion of test results in agreement over the number of all tests performed, is a straightforward measure but does not account for agreement beyond chance. Therefore, Cohen’s kappa statistic, which adjusts for random matches, is often used to measure the agreement between two test results [23]. Cohen’s kappa values range from zero (agreement is equal to that expected by chance) to one (complete agreement beyond chance), with benchmarks between agreement categories varying among authors [2426]. Although Cohen’s kappa is a standard measure, it has limitations such as prevalence and bias effects. Prevalence effects arise when the proportion of positive results deviates significantly from 50% [27]. The effect of prevalence depends on the method of modeling agreement and can substantially reduce kappa values [27]. Bias effects occur when there is a disparity in the proportion of positive results between the two tests, which leads to reduced kappa values [27]. To address these effects, the prevalence- and bias-adjusted kappa (PABAK) can be calculated [28,29].

The reasons for disagreement between diagnostic tests have rarely been thoroughly investigated. Potential factors include poor sample quality, variability in tests used, and discrepancies arising from the same test being conducted in different laboratories [18,22,30]. Additionally, biological factors such as age, sex, other diseases, and species may influence the consistency of test results for the same sample. Previous studies have suggested associations between test performance and variables such as region, age, and livestock species [3133]. However, these studies have not provided conclusive evidence or statistical significance.

The objective of this study was to assess the inter-laboratory agreement of standard diagnostic tests utilized by the participating laboratories, measured by the concordance, Cohen’s kappa, and PABAK. The comparison involved results obtained from commercial ELISA tests conducted in a laboratory in Chad and results obtained from ELISA and indirect IFA tests performed on livestock and human serum aliquots, respectively, in laboratories in Germany and Switzerland. Additionally, we evaluated the influence of demographic factors on the agreement between the two test results. The study enhances our understanding of the inter-laboratory agreement of diagnostic test results across laboratory conditions and, for humans, test types, which is crucial for accurately interpreting results from epidemiological seroprevalence studies.

2. Material and Methods

2.1. Ethics statement

The study was approved by the Ethics Committee of Northwest and Central Switzerland (EKNZ) (project id 2017–00884) and by the Comité National de Bioéthique du Tchad (CNB-Tchad) (project id 134/PR/MESRS/CNBT/2018). Formal written consent was obtained from study participants and animal owners after we presented our study to the community and before data collection occurred.

2.2. Sample collection and laboratory analysis in Chad

The samples analysed in this study were collected between January and February 2018, as reported by Özcelik, R., et al [4]. In brief, a cross-sectional study in livestock (cattle, sheep, goats, horses, and donkeys) and human populations was conducted in the two rural health districts, Yao and Danamadji, in Chad. Multistage cluster sampling was used, with villages and nomadic camps serving as cluster units. The clusters were chosen based on human population size. The sample size was calculated using R software, assuming a 50% prevalence for Q fever, and RVF, with a 95% confidence interval. The sample size for two-stage cluster sampling was determined using a formula incorporating the sample size for simple random sampling, the intra-cluster correlation coefficient of 0.2, and the number of 20 individuals sampled per cluster [4]. In Danamadji and Yao, respectively, blood samples were collected from 571 and 389 humans and 560 and 483 livestock. No individuals showing clear signs of illness were sampled, and similarly, animals that were visibly ill and hence often isolated were also excluded from the sampling [4].

The samples were subsequently analysed at the Institut de Recherche en Élevage pour le Développement (IRED) in N’Djamena, Chad. Livestock and human samples were analysed using different indirect ELISAs: ID Screen Q Fever Indirect Multi-species ELISA from IDvet for livestock and the Panbio Coxiella burnetii IgG ELISA from Abbott for humans. For RVF, a competitive ELISA (ID Screen Rift Valley Fever Competition Multi-species from IDvet) was used for human and livestock samples. The diagnostic test procedure and thresholds were applied according to the manufacturer’s protocols without modification (S1 Table). Equivocal samples were retested once.

2.3. Diagnostic testing in Switzerland and Germany

Following the initial diagnostic analysis at IRED, 10% of the human and livestock samples from each region were randomly selected and sent as serum aliquots to laboratories in Switzerland and Germany in 2021 for repeated diagnostic analysis for QF and RVF, respectively. The decision to use a 10% subsample of the original sample size was based on the resources available. In Switzerland, two indirect ELISAs (IDEXX Q Fever IgG Antibody and ID Screen Q Fever Indirect Multispecies from IDvet) were used at the Center for Zoonoses, Animal Bacterial Diseases, and Antimicrobial Resistance (ZOBA) for QF diagnostics in livestock samples (ruminants and equids, respectively). At the Institute for Infectious Diseases (IFIK) of the University of Bern, an indirect IFA (Q Fever IFA IgG assay from Focus Diagnostics, US) was used for QF diagnostics in human samples. For RVF, livestock samples were analysed using a competitive ELISA (ID Screen Rift Valley Fever Competition Multi-species ELISA from IDvet) at the Friedrich-Loeffler-Institute (FLI), and human samples were analysed using an indirect IFA (Anti-Rift-Tal-Fieber-Viren-IIFT [IgG] from EUROIMMUN) at the Robert Koch Institute (RKI). The diagnostic test procedure and thresholds were applied according to the manufacturer’s protocols without modification (S1 Table). Equivocal samples were not retested.

2.4. Sample processing

Of the 206 serum aliquots from 103 livestock samples and 192 serum aliquots from 96 human samples sent to laboratories in Switzerland and Germany, 15 livestock aliquots and 30 human aliquots were excluded from further analysis (Table 1). Reasons for exclusion included unsuccessful matching of test identities due to labelling errors, missing serum upon arrival, and equivocal results (Table 1). As a result, one livestock sample and five human samples were fully excluded from the study sample. Ultimately, 91 human and 102 livestock samples were tested using one or both assays and were included in the agreement testing (Table 1, Fig 1).

Table 1. Overview of diagnostic tests from Chad and Switzerland (CH) or Germany (GER).

Disease QF1 RVF2
Samples Livestock Humans Livestock Humans
Laboratory, Chad IRED3 IRED3 IRED3 IRED3
Test 1, Chad IDvet ID Screen Q Fever Indirect Multi-Species Abbott Panbio Coxiella burnetii IgG ELISA IDvet ID Screen Rift Valley Fever Competition Multi-species IDvet ID Screen Rift Valley Fever Competition Multi-species
Laboratory, CH/GER ZOBA (CH)4 IFIK (CH)5 FLI (GER)6 RKI (GER)7
Test 2, CH/GER IDEXX Q Fever Ab Test (ruminants)
IDvet ID Screen Q Fever Indirect Multispecies (equids)
Focus Diagnostics Q Fever IFA IgG assay IDvet ID Screen Rift Valley Fever Competition Multi-species EUROIMMUN RVF IFA IgG assay
Initial number of serum aliquots sent to CH/GER 103 96 103 96
Excluded serum aliquots from test 2 due to missing serum 2 8 0 4
Equivocal results during test 2 8 14 2 0
Non-matchable 2 2 1 2
Number of serum aliquots used for agreement testing 91 72 100 90

1Q fever

2Rift Valley fever

3Institut de Recherche en Elevage pour le Développement, Chad

4Center for Zoonoses, Animal Bacterial Diseases and Antimicrobial, Switzerland (CH)

5Institute for Infectious Diseases, Switzerland (CH)

6Friedrich-Loeffler-Institute, Germany (GER)

7Robert Koch Institute, Germany (GER)

Fig 1. Sample collection and laboratory analysis were conducted at the Institut de Recherche en Élevage pour le Développement (IRED) in Chad, with subsequent diagnostic testing performed in Switzerland (CH) and Germany (GER).

Fig 1

The specific tests used in each laboratory are noted, along with the antibodies detected (IgG = Immunoglobulin G). The number of samples processed at each stage is provided. Different colors in the diagnostic tests indicate that assays from different manufacturers were employed. Of the original 103 livestock and 96 human samples sent to Germany and Switzerland, 102 livestock and 91 human samples underwent either both QF and RVF diagnostics or one of the tests. One livestock sample and five human samples were unavailable for any diagnostic procedure and were fully excluded from further analysis. Abbreviations: ZOBA: Center for Zoonoses, Animal Bacterial Diseases, and Antimicrobial Resistance; IFIK: Institute for Infectious Diseases; FLI: Friedrich-Loeffler-Institute; RKI: Robert Koch Institute; ELISA: Enzyme-linked Immunosorbent assay; IFA: Immunofluorescence assay. Figure created with BioRender.com.

2.5. Statistical analysis

The inter-laboratory agreement of the test results from Chad and Switzerland or Germany was evaluated using concordance, Cohen’s Kappa, and PABAK for each of the four datasets: QF results in human samples, QF results in livestock samples, RVF results in human samples, and RVF results in livestock samples. Cohen’s kappa and PABAK values were interpreted according to the standard scale: ’fair’ agreement (kappa = 0.21–0.40), ’moderate’ agreement (kappa = 0.41–0.60), ’substantial’ agreement (kappa = 0.61–0.80), and ’almost perfect’ agreement (kappa > 0.80) [26].

In addition, we investigated factors associated with test agreement by assigning to each sample a value of 0 if there was disagreement between the two test results (i.e., negative in Chad and positive in Switzerland/Germany, or the opposite) and 1 if the test results were consistent (i.e., both positive or both negative). This binary outcome was used as the dependent variable in logistic regression models to identify the statistical association between test agreement and demographic factors, including the district (Yao versus Danamadji) and setting (village versus camp) where the sample was collected, and sex, age, and livestock species (cattle, small ruminants, and equids) of the sampled individual. The variable age was analysed in two ways, as a continuous and as a categorical variable. For the continuous variable, a unit of 1 year was used for livestock and of 10 years for humans. For the categorical variable, samples were stratified as < 2 years (age group 1), 2–3 years (age group 2), 4 years and older (age group 3) for livestock, and < 30 years (age group 1), 30–39 years (age group 2), 40–60 years (age group 3), 61 years and older (age group 4) for humans.

Univariable logistic regressions were initially performed to assess individual predictors. In cases where the univariable model was infeasible due to perfect agreement in one group, Chi-square tests were applied. Odds ratios (OR) and their corresponding 95% confidence intervals were calculated for these analyses. To consider potential interdependencies between the variables, we applied multivariable logistic regressions to estimate adjusted coefficients and OR. We included all variables and selected age as categorical variable.

Statistical calculations, modeling, and data visualization were conducted in R (version 4.2.2). The package “irr” was used to calculate the concordance and Cohen’s kappa. The package “vcd” was used to obtain confidence intervals of Cohen’s kappa, computed using the standard method based on normal approximation [34]. The epi.kappa() function from the “epiR” package was used to calculate PABAK and corresponding confidence intervals. The confidence intervals for the OR were calculated using the output values of the associated regression model, with the upper and lower CIs derived using the function exp(confint.default(model)).

3. Results

3.1. Sample population

Of the 91 human and 102 livestock samples that were used to perform the inter-laboratory test agreement analysis, most (62% of human and 57% of livestock) samples were collected from Danamadji (S2 Table). Fifty-six percent of human samples and 58% of livestock samples were collected from camps. The sex distribution was uneven, with 70% of human sampled being male and 70% of livestock sampled being females. Among humans, age groups 1–3 were evenly represented (30%, 28%, 33%), while only 9% belonged to age group 4. In livestock, 50% of the samples were from age group 2, with 17% and 33% from age groups 1 and 3, respectively. Most livestock samples were from cattle (46%), followed by small ruminants (41%) and equids (13%) (S2 Table). The livestock sampled originated from 24 different herds (one herd per cluster).

3.2. Diagnostic tests agreement

3.2.1 Level of inter-laboratory test agreement

Concordance values ranged from 62.5% to 94% (Table 2). Cohen’s kappa values, which ranged from 0.31 to 0.59, indicated that livestock QF and RVF, and human RVF tests had ’moderate’ agreement, while human QF tests had ’fair’ agreement (Table 2). PABAK values showed that the livestock QF and RVF tests had ’almost perfect’ agreement, the human RVF test had ’substantial’ agreement, and the human QF tests had ’fair’ agreement (Table 2).

Table 2. Analyses of test agreement (concordance and Cohen’s kappa) of samples tested in an inter-laboratory test agreement study.
Switzerland Germany
QF1 livestock QF1 humans RVF2 livestock RVF2 humans
Chad Pos Neg Pos Neg Pos Neg Pos Neg
Pos 6 3 22 3 5 4 16 9
neg 4 78 24 23 2 89 8 57
Concordance (%) 92.3 62.5 94.0 81.1
Cohen’s kappa value (95% CI) 0.59 (0.31, 0.86) 0.31 (0.13, 0.49) 0.59 (0.30, 0.89) 0.52 (0.33, 0.72)
PABAK value (95% CI) 0.846 (0.696, 0.937) 0.25 (0.006, 0.473) 0.88 (0.748, 0.955) 0.622 (0.43, 0.771)

1Q fever

2Rift Valley fever

3.2.2. Influence of factors on inter-laboratory test agreement

For QF in livestock, none of the investigated demographic factors significantly impacted the agreement between the two test results in both univariable and multivariable analyses (Tables S3 and 3). However, some notable trends (p < 0.15) emerged: small ruminants tended to show better agreement than cattle, and samples from Yao showed lower agreement compared to those from Danamadji (Table 3). For QF in humans, samples from villages had significantly higher agreement compared to those from camps, with odds of agreement being 13.4 times higher (Table 3). Additionally, older age groups had significantly lower agreement compared to the youngest age group (Table 3).

Table 3. Results of multivariable logistic regression models investigating the effect of demographic factors as independent variables on the inter-laboratory test agreement of Q fever (QF) and Rift Valley fever (RVF) in humans and livestock.

Goodness of fit of the models are presented as pseudo-R-squared (R2). Bold p-values indicate significance based on a threshold of 0.05.

Switzerland Germany
QF1 livestock QF1 humans RVF2 livestock RVF2 humans
Independent variable (OR and 95% CI)
P-value (P)
District
(Reference: Danamadji)
Yao:
0.14 (0.01, 1.74)
P = 0.13
Yao:
1.38 (0.38, 4.99)
P = 0.63
Yao:
4.99 (0.41, 60.49)
P = 0.21
Yao:
1.37 (0.37, 5.04)
P = 0.64
Setting
(Reference: camp)
Village:
0.34 (0.05, 2.06)
P = 0.24
Village:
13.35 (3.20, 55.86)
P = 0.0004
Village:
0.32 (0.04, 2.43)
P = 0.27
Village:
1.85 (0.58, 5.93)
P = 0.30
Species
(Reference: cattle)
Equids:
5.51 (0.29, 103.21)
P = 0.25
Small ruminants:
6.26 (0.59, 65.76)
P = 0.13
- Equids:
Inf (perfect agreement)
Small ruminants:
0.65 (0.09, 4.78)
P = 0.67
-
Age group (Reference: age group 1/group 2*)
Group 2:
1.30 (0.10, 17.39)
P = 0.84
Group 3:
0.93 (0.06, 14.38)
P = 0.96
Group 2:
0.67 (0.15, 3.07)
P = 0.61
Group 3:
0.14 (0.03, 0.82)
P = 0.03
Group 4:
0.09 (0.01, 0.83)
P = 0.03
Group 1*:
Inf (perfect agreement)
Group 3*:
0.30 (0.04, 2.41)
P = 0.26
Group 2:
0.62 (0.12, 3.16)
P = 0.57
Group 3:
0.32 (0.07, 1.47)
P = 0.14
Group 4:
0.16 (0.02, 1.14)
P = 0.07
Sex
(Reference: male)
Female:
0.38 (0.03, 4.61)
P = 0.45
Female:
0.82 (0.20, 3.38)
P = 0.78
Female:
0 (perfect agreement in males)
Female:
1.01 (0.27, 3.86)
P = 0.99
R2 0.23 0.35 0.28 0.09

1Q fever

2Rift Valley fever

* In the multivariable logistic regression model, age group 2 was used as the reference for both age group 1 and age group 2 for RVF livestock, due to perfect agreement observed in age group 1.

For RVF in humans and livestock, none of the demographic factors significantly influenced test agreement. However, some trends were observed for RVF test agreement in humans, with older age groups showing lower agreement compared to the youngest age group (Table 3).

Across diseases and populations, there was a consistent trend of lower agreement with increasing age, which was significant for human QF and almost reached significance for human RVF (Table 3 and Fig 2). For livestock RVF tests, the odds ratio for agreement was also lower in older age groups, although the p-value was 0.26 (Table 3).

Fig 2. Mosaic plot illustrating the association between age groups and inter-laboratory test agreement, the widths of the adjacent rectangles represent the proportion of individuals within each age group.

Fig 2

The heights of the rectangles represent the proportion of individuals within each age group who exhibit a specific test agreement outcome (concordant or discrepant). Abbreviations: QF: Q fever; RVF: Rift Valley fever.

4. Discussion

Our study revealed varying levels of test agreement, ranging from fair to moderate (Cohen’s kappa) or almost perfect when considering PABAK. The good inter-laboratory agreement of livestock test results for RVF was in line with other studies on RVF test agreement, although it is important to note that literature on this subject is limited [20,21]. Similarly, the body of research on test agreement for livestock QF is also limited. The inter-laboratory agreement for livestock QF observed in our study was slightly lower than the values reported in previous studies that evaluated agreement between different ELISA tests or between ELISA and IFA [18,19]. However, those previously reported test comparisons were performed within the same laboratories and time frame, which may explain the observed differences in agreement. Notably, the livestock RVF agreement was assessed using the same ELISA from the same manufacturer in both Chad and Germany. In contrast, for livestock QF, only the equid samples were tested using the same ELISA in Switzerland as in Chad, whereas ruminant samples were analyzed with an ELISA from a different manufacturer. The latter ELISA was adapted for automation in ruminants and was therefore the standard assay used by the participating laboratory for this species. This discrepancy may have contributed to the better inter-laboratory agreement observed for livestock RVF compared to livestock QF.

We observed a notably lower agreement for human tests compared to livestock tests for both diseases, which may be attributed to using two different diagnostic methods for human samples, ELISA in Chad and indirect IFA in Switzerland or Germany. Previous studies have shown varying sensitivities and specificities of the commercial diagnostic test for QF used in this study (Panbio from Abbott), ranging from 71% to 100% [10,3537]. In these studies, indirect IFA was used as a reference method to evaluate the ELISA, revealing varying agreement between the results of the two tests. The variability in sensitivity for the same test raises the question of whether it is due to the scope for interpretation in indirect IFA, which is considered a challenge, even though indirect IFA is regarded as the gold standard for human QF diagnostics [5,7]. A study that compared QF indirect IFA results from different reference centres in three countries (United Kingdom, France, and Australia) found a concordance between the indirect IFA results of only 35% [38]. Our results reflect this uncertainty and underline the complexity of QF diagnostics [5,7,39]. A previous study comparing IFA (the same test used in our study) and ELISA (from a different manufacturer than the one used in our study) found that IgG phase II antibody titers decline more slowly when measured by IFA, with significantly higher detection rates after one year [40]. These findings indicate that IFA may offer superior sensitivity and longer-lasting detection of IgG phase II antibodies compared to ELISA, which may account for the lower agreement observed in the human QF tests in our study. Similarly, for RVF, the lower inter-laboratory agreement for human tests compared to livestock tests may be attributed to the use of different testing methods in Chad (ELISA) and Germany (indirect IFA). The reported sensitivity and specificity values for the ELISA kit used in Chad are 98% and 100%, respectively [20]. While the IFA test used in Germany (EUROIMMUN RVF IFA IgG assay) is a certified assay, its performance has not been evaluated by external independent assessment [12]. The existing literature on RVF diagnostic tests, particularly IFA, is limited. Generally, IFA can achieve high specificity and sensitivity; however, these metrics are highly dependent on the operator’s skill in interpreting fluorescent signals under a microscope and the quality of the reagents used [41].

Although quantitative information on the quality of the individual samples in our study are unavailable, laboratory staff in Switzerland and Germany reported concerns regarding haemolysis. Given that haemolysis is a well-documented cause of test result discrepancies in clinical laboratories [42,43], it may have contributed to the observed disagreement between test results in our study. Although serum was sent to the laboratories in Germany and Switzerland, haemolysis may have occurred before or during the process of separating serum from whole blood. Consequently, the serum could contain haemoglobin and other intracellular components from lysed red blood cells, potentially compromising the quality of the sample and interfering with laboratory tests [43]. There are many in vitro causes of haemolysis, mostly pre-analytical problems such as incorrect procedures and/or materials used in blood collection, while transport, processing, and storage account for only a minority of cases [44]. This limitation shows the importance of careful planning and execution of the pre-analytical phase, especially in prevalence studies where the outcome can be influenced by the quality of the sample material. Nevertheless, it is crucial to recognize the challenges associated with collecting samples under difficult field conditions, where access to centrifuges may be limited until several days after blood sampling. In addition, the samples in our study were stored for 2.5 years with repeated freeze-thaw cycles between the performance of the two tests for some of the samples, which probably affected sample quality. Several studies have assessed the effects of repeated freeze-thaw cycles on IgG stability, although not specifically on Anti-QF/RVF IgG, and found minimal impact [45,46]. However, the effects of prolonged storage have not been explored. Future research should investigate whether extended storage, such as the period between processing at the local laboratory and further analysis at secondary laboratories, could compromise sample quality, and consequently test results. We emphasize the importance of considering these challenges when discussing the outcomes of epidemiological prevalence studies or diagnostic test evaluation studies.

Our results demonstrate a statistical relationship between test agreement and age of the sampled individuals, with higher agreement observed in younger individuals for both diseases and in humans and livestock. This finding aligns with a study by de Bronsvoort et al. (2019), which suggested that lower agreement in older individuals may be due to their higher likelihood of previous exposure to other pathogens over their lifetime [31]. This may result in cross-reactivity in serological tests, making serological differentiation between diseases more difficult [39]. Cross-reactions caused by antibodies induced by other pathogens, such as C. burnetii antigens, the agent causing QF, with antibodies produced against Bartonella spp., Legionella spp., and Chlamydiae spp. has been reported [4749], or RVF virus antigens with antibodies produced against Rio Grande virus [5052]. Furthermore, with increasing age, the likelihood of exposure to the causative agent of QF and RVF increases [53,54], as does the chance of having residual antibodies against these diseases in the blood [6,5557]. In addition, in QF, antigen-lipopolysaccharide complexes remaining in the host after infection with C. burnetii can trigger humoral and cell-mediated immune responses, producing interfering antibodies [58,59]. Such antibody titres can potentially produce ambiguous results that are neither clearly positive nor negative, leading to misinterpretation and discordant test results. Additionally, immunosenescence–the gradual deterioration of the immune system associated with aging–may also play a role in the increased disagreement of diagnostic test results among older individuals. Immunosenescence is characterized by reduced immune function, altered antibody responses, and an increased tendency toward chronic inflammation, including the presence of proinflammatory "age-associated B cells," which have been shown to produce autoantibodies [6062]. These autoantibodies may interfere with antigens used in diagnostic tests; a hypothesis that warrants further investigation. Consequently, these findings underscore the need for further research to develop age-specific diagnostic protocols, such as adjusting cut-off values for interpreting results, and to improve test accuracy across diverse populations.

The district-level analysis showed a trend toward higher agreement for livestock QF tests in samples from Danamadji compared to Yao. This regional variation may be influenced by varying local environmental conditions, disease prevalence, or livestock management practices. However, we did not capture such information, so we were not able to identify a potential latent factor that can explain the difference in test agreement detected for the two regions. Continuing with the focus on sampling location, the sample collection setting (village versus camp) was identified as a significant factor in the multivariable model for human QF test agreement. One hypothesis for this finding is that villages generally have shorter distances from the sampling location to the laboratory, allowing for more appropriate storage conditions compared to remote camps. This shorter travel time likely preserves sample integrity, resulting in higher test agreement. The significant role of the setting variable suggests that environmental and logistical factors may be crucial for diagnostic test result interpretation, although we did not observe the influence of the setting for RVF in humans nor for livestock species. Future studies should investigate the impact of environmental and logistical factors on test results and gather detailed data on local conditions and practices that affect sample quality and test agreement.

Small ruminants exhibited a trend towards higher agreement in QF test results compared to cattle. In a different type of QF ELISA, Stellfeld et al. (2020) observed significant differences in the ranges of OD450 values obtained from sera of sheep, goats, and cattle [63]. Sheep exhibited a larger range of OD450 values, whereas cattle showed a smaller range [63]. Greater distribution between OD450 values allows for more precise grading of ELISA results, which may explain the better inter-test agreement for small ruminants compared to cattle. These differences in OD450 ranges are likely due to species-specific immune responses to C. burnetii, suggesting varying immune reactions among ruminant species [63].

Although females showed lower inter-laboratory test agreement compared to males for both diseases and populations, the high p-values and wide confidence intervals of the ORs in our data did not indicate any significant influence of sex on the inter-laboratory test agreement.

Conclusion

Our study highlights the variability in inter-laboratory diagnostic test agreement for QF and RVF serology in humans and livestock based on samples collected in Chad. Despite differences in laboratories, personnel, and test types, test agreements ranged from fair to moderate (Cohen’s kappa) or almost perfect considering PABAK. Given the reliance on serological profiles for QF and RVF epidemiological studies, it is crucial to consider factors that may complicate accurate diagnosis. We identified that human QF test agreement was significantly higher in individuals living in villages and younger individuals, with the latter trend also observed in human RVF tests. Our findings emphasize the need to recognize that diagnostic tests may yield varying results, impacting the outcome and interpretation of disease prevalence studies. If resources permit, it is recommended to confirm positive results by retesting with the gold standard test.

Supporting information

S1 Table. This document (S1A–S1F Tables) provides a summary of the test protocols for all diagnostic assays employed in this study.

(DOCX)

pntd.0012300.s001.docx (20KB, docx)
S2 Table. This table presents the demographic characteristics of the sample population.

(DOCX)

pntd.0012300.s002.docx (17.2KB, docx)
S3 Table. This table presents the findings from the univariable logistic regression models.

(DOCX)

pntd.0012300.s003.docx (18.3KB, docx)
S1 Dataset. This dataset consists of the data required to replicate all study findings reported in the article for the human population.

(XLSX)

pntd.0012300.s004.xlsx (17.2KB, xlsx)
S2 Dataset. This dataset consists of the data required to replicate all study findings reported in the article for the livestock population.

(XLSX)

pntd.0012300.s005.xlsx (19.4KB, xlsx)

Data Availability

All relevant data are within the manuscript and its Supporting Information files. All data are fully anonymized to protect the privacy of the participants involved in the study.

Funding Statement

This study was funded by the Swiss Federal Food Safety and Veterinary Office, grant number 1.17.09 (https://www.blv.admin.ch/blv/de/home.html), and the Wolfermann-Nägeli Foundation, grant number 2016/19 (https://stiftungen.stiftungschweiz.ch/organisation/wolfermann-naegeli-stiftung to S.D.). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

Declaration of generative AI and AI-assisted technologies in the writing process

While preparing this work, the author used ChatGPT to correct the English of revised sentences (using the command: ‘correct the English’). After using this tool, the author reviewed and edited the content as needed and takes full responsibility for the content of the publication.

References

  • 1.Fayiz Abakar M, Naré NB, Schelling E, Hattendorf J, Alfaroukh IO, Zinsstag J. Seroprevalence of Rift Valley fever, Q fever, and brucellosis in ruminants on the southeastern shore of Lake Chad. Vector-Borne and Zoonotic Diseases 2014;14: 757–62. doi: 10.1089/vbz.2014.1585 [DOI] [PubMed] [Google Scholar]
  • 2.Blomström AL, Scharin I, Stenberg H, Figueiredo J, Nhambirre O, Abilio A, et al. Seroprevalence of Rift Valley fever virus in sheep and goats in Zambézia, Mozambique. Infect Ecol Epidemiol 2016;6. doi: 10.3402/IEE.V6.31343 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Ibrahim M, Schelling E, Zinsstag J, Hattendorf J, Andargie E, Tschopp R. Sero-prevalence of brucellosis, Q-fever and Rift Valley fever in humans and livestock in Somali Region, Ethiopia. PLoS Negl Trop Dis 2021;15: 1–18. doi: 10.1371/journal.pntd.0008100 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Özcelik R, Abakar MF, Counotte MJ, Zakaria FA, Kimala P, Issa R, et al. Seroprevalence and associated risk factors of brucellosis, Rift Valley fever and Q fever among settled and mobile agro-pastoralist communities and their livestock in Chad. PLoS Negl Trop Dis 2023;17: e0011395. doi: 10.1371/journal.pntd.0011395 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Eldin C, Mélenotte C, Mediannikov O, Ghigo E, Million M, Edouard S, et al. From Q fever to Coxiella burnetii infection: A paradigm change. Clin Microbiol Rev 2017;30: 115–90. doi: 10.1128/CMR.00045-16 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Wright D, Kortekaas J, Bowden TA, Warimwe GM. Rift Valley fever: biology and epidemiology. J Gen Virol 2019;100: 1187–99. doi: 10.1099/jgv.0.001296 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Q fever—WOAH—World Organisation for Animal Health [cited 18 June 2024]. Available from: https://www.woah.org/en/disease/q-fever/
  • 8.Rift Valley fever—WOAH—World Organisation for Animal Health [cited 18 June 2024]. Available from: https://www.woah.org/en/disease/rift-valley-fever/
  • 9.Imbert G, La Scola B. Diagnosis of Q fever using indirect microimmunofluorescence. Methods Mol Biol 2006;345: 197–202. doi: 10.1385/1-59745-143-6:197 [DOI] [PubMed] [Google Scholar]
  • 10.Dangel L, Koempf D, Fischer SF. Comparison of different commercially available enzyme- linked immunosorbent assays with immunofluorescence test for detection of phase II IgG and IgM antibodies to coxiella burnetii. J Clin Microbiol 2020;58. doi: 10.1128/JCM.00951-19 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Wood C, Muleme M, Tan T, Bosward K, Gibson J, Alawneh J, et al. Validation of an indirect immunofluorescence assay (IFA) for the detection of IgG antibodies against Coxiella burnetii in bovine serum. Prev Vet Med 2019;169: 104698. doi: 10.1016/j.prevetmed.2019.104698 [DOI] [PubMed] [Google Scholar]
  • 12.Lapa D, Pauciullo S, Ricci I, Garbuglia AR, Maggi F, Scicluna MT, et al. Rift Valley Fever Virus: An Overview of the Current Status of Diagnostics. Biomedicines 2024, Vol 12, Page 540 2024;12: 540. doi: 10.3390/biomedicines12030540 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Wilson WC, Weingartl HM, Drolet BS, Davé K, Harpster MH, Johnson PA, et al. Diagnostic Approaches for Rift Valley Fever. Dev Biol (Basel) 2013;135: 73–8. doi: 10.1159/000276828 [DOI] [PubMed] [Google Scholar]
  • 14.Gerken KN, Labeaud AD, Mandi H, Jackson ML, Breugelmans JG, King CH. Paving the way for human vaccination against Rift Valley fever virus: A systematic literature review of RVFV epidemiology from 1999 to 2021. PLoS Negl Trop Dis 2022;16. doi: 10.1371/journal.pntd.0009852 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Sidi-Boumedine K, Rousset E, Henning K, Ziller M, Niemczuck K, Roest H, et al. Development of harmonised schemes for the monitoring and reporting of Q-fever in animals in the European Union. EFSA Supporting Publications 2017;7. doi: 10.2903/SP.EFSA.2010.EN-48 [DOI] [Google Scholar]
  • 16.Sadiki V, Gcebe N, Mangena ML, Ngoshe YB, Adesiyun AA. Prevalence and risk factors of Q fever (Coxiella burnetii) in cattle on farms of Limpopo province, South Africa. Front Vet Sci 2023;10: 1101988. doi: 10.3389/fvets.2023.1101988 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Natale A, Bucci G, Capello K, Barberio A, Tavella A, Nardelli S, et al. Old and new diagnostic approaches for Q fever diagnosis: Correlation among serological (CFT, ELISA) and molecular analyses. Comp Immunol Microbiol Infect Dis 2012;35: 375–9. doi: 10.1016/J.CIMID.2012.03.002 [DOI] [PubMed] [Google Scholar]
  • 18.Kittelberger R, Mars J, Wibberley G, Sting R, Henning K, Horner GW, et al. Comparison of the Q-fever complement fixation test and two commercial enzyme-linked immunosorbent assays for the detection of serum antibodies against Coxiella burnetii (Q-fever) in ruminants: Recommendations for use of serological tests on imported animals in New Zealand. N Z Vet J 2009;57: 262–8. doi: 10.1080/00480169.2009.58619 [DOI] [PubMed] [Google Scholar]
  • 19.Rousset E, Durand B, Berri M, Dufour P, Prigent M, Russo P, et al. Comparative diagnostic potential of three serological tests for abortive Q fever in goat herds. Vet Microbiol 2007;124: 286–97. doi: 10.1016/j.vetmic.2007.04.033 [DOI] [PubMed] [Google Scholar]
  • 20.Kortekaas J, Kant J, Vloet R, Cêtre-Sossah C, Marianneau P, Lacote S, et al. European ring trial to evaluate ELISAs for the diagnosis of infection with Rift Valley fever virus. J Virol Methods 2013;187: 177–81. doi: 10.1016/j.jviromet.2012.09.016 [DOI] [PubMed] [Google Scholar]
  • 21.Pawęska JT, van Vuren PJ, Msimang V, Lô MM, Thiongane Y, Mulumba-Mfumu LK, et al. Large-Scale International Validation of an Indirect ELISA Based on Recombinant Nucleocapsid Protein of Rift Valley Fever Virus for the Detection of IgG Antibody in Domestic Ruminants. Viruses 2021;13. doi: 10.3390/v14010013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Horigan MW, Bell MM, Pollard TR, Sayers AR, Pritchard GC. Q fever diagnosis in domestic ruminants: comparison between complement fixation and commercial enzyme-linked immunosorbent assays. J Vet Diagn Invest 2011;23: 924–31. doi: 10.1177/1040638711416971 [DOI] [PubMed] [Google Scholar]
  • 23.Thrusfield M, Christley R, Brown H, Diggle PJ, French N, Howe K, et al. Veterinary Epidemiology: Fourth Edition. Veterinary Epidemiology: Fourth Edition 2017: 1–861. doi: 10.1002/9781118280249 [DOI] [Google Scholar]
  • 24.Fleiss JL, Levin B, Cho Paik M. Statistical Methods for Rates and Proportions, Third Edition. Statistical Methods for Rates and Proportions, Third Edition 2004: 1–760. doi: 10.1002/0471445428 [DOI] [Google Scholar]
  • 25.Altman DG. Practical Statistics for Medical Research. Practical Statistics for Medical Research 1990. doi: 10.1201/9780429258589 [DOI] [Google Scholar]
  • 26.Landis JR, Koch GG. The Measurement of Observer Agreement for Categorical Data. Biometrics 1977;33: 159. doi: 10.2307/2529310 [DOI] [PubMed] [Google Scholar]
  • 27.Hoehler FK. Bias and prevalence effects on kappa viewed in terms of sensitivity and specificity. J Clin Epidemiol 2000;53: 499–503. doi: 10.1016/s0895-4356(99)00174-2 [DOI] [PubMed] [Google Scholar]
  • 28.Byrt T, Bishop J, Carlin JB. Bias, prevalence and kappa. J Clin Epidemiol 1993;46: 423–9. doi: 10.1016/0895-4356(93)90018-v [DOI] [PubMed] [Google Scholar]
  • 29.Lantz CA, Nebenzahl E. Behavior and interpretation of the κ statistic: Resolution of the two paradoxes. J Clin Epidemiol 1996;49: 431–4. doi: 10.1016/0895-4356(95)00571-4 [DOI] [PubMed] [Google Scholar]
  • 30.Paul S, Toft N, Agerholm JS, Christoffersen AB, Agger JF. Bayesian estimation of sensitivity and specificity of Coxiella burnetii antibody ELISA tests in bovine blood and milk. Prev Vet Med 2013;109: 258–63. doi: 10.1016/J.PREVETMED.2012.10.007 [DOI] [PubMed] [Google Scholar]
  • 31.de Bronsvoort BMC, Bagninbom JM, Ndip L, Kelly RF, Handel I, Tanya VN, et al. Comparison of Two Rift Valley Fever Serological Tests in Cameroonian Cattle Populations Using a Bayesian Latent Class Approach. Front Vet Sci 2019; 6:258. doi: 10.3389/fvets.2019.00258 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Emery MP, Ostlund EN, Schmitt BJ. Comparison of Q fever serology methods in cattle, goats, and sheep. J Vet Diagn Invest 2012;24: 379–82. doi: 10.1177/1040638711434943 [DOI] [PubMed] [Google Scholar]
  • 33.Lurier T, Rousset E, Gasqui P, Sala C, Claustre C, Abrial D, et al. Evaluation using latent class models of the diagnostic performances of three ELISA tests commercialized for the serological diagnosis of Coxiella burnetii infection in domestic ruminants. Vet Res 2021;52. doi: 10.1186/s13567-021-00922-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Package “vcd” Title Visualizing Categorical Data 2023 [cited 18 June 2024]. Available from: https://cran.r-project.org/web/packages/vcd/vcd.pdf
  • 35.Field PR, Mitchell JL, Santiago A, Dickeson DJ, Chan S-W, Ho DWT, et al. Comparison of a Commercial Enzyme-Linked Immunosorbent Assay with Immunofluorescence and Complement Fixation Tests for Detection of Coxiella burnetii (Q Fever) Immunoglobulin M. J Clin Microbiol 2000;38: 1645. doi: 10.1128/JCM.38.4.1645-1647.2000 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Field PR, Santiago A, Chan SW, Patel DB, Dickeson D, Mitchell JL, et al. Evaluation of a novel commercial enzyme-linked immunosorbent assay detecting Coxiella burnetii-specific immunoglobulin G for Q fever prevaccination screening and diagnosis. J Clin Microbiol 2002;40: 3526–9. doi: 10.1128/JCM.40.9.3526-3529.2002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Sanz JC, De Los Ríos R, Martín F, Ángeles Tébar M, Jado I, Anda P. Aplicación de cuatro técnicas de ELISA (dos para IgM y dos para IgG) en el diagnóstico serológico de un brote de fiebre Q. Enferm Infecc Microbiol Clin 2006;24: 178–81. doi: 10.1157/13086551 [DOI] [PubMed] [Google Scholar]
  • 38.Healy B, Van Woerden H, Raoult D, Graves S, Pitman J, Lloyd G, et al. Chronic Q Fever: Different Serological Results in 3 Countries-Results of a Follow-up Study 6 Years After a Point Source Outbreak. Clin Infect Dis. 2011;52: 1013–1019. doi: 10.1093/cid/cir132 [DOI] [PubMed] [Google Scholar]
  • 39.Ferrara G, Colitti B, Pagnini U, Iovane G, Rosati S, Montagnaro S. Characterization of recombinant Ybgf protein for the detection of Coxiella antibodies in ruminants. Journal of Veterinary Diagnostic Investigation 2022;34: 646–53. doi: 10.1177/10406387221093581 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Wegdam-Blans MCA, Wielders CCH, Meekelenkamp J, Korbeeck JM, Herremans T, Tjhie HT, et al. Evaluation of commonly used serological tests for detection of Coxiella burnetii antibodies in well-defined acute and follow-up sera. Clinical and Vaccine Immunology 2012;19: 1110–5. doi: 10.1128/CVI.05581-11 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Immunofluorescence Cimolai N. Serodiagnosis of the Infectious Diseases 1999: 33–8. doi: 10.1007/978-1-4615-5249-9_5 [DOI] [Google Scholar]
  • 42.Lippi G, Von Meyer A, Cadamuro J, Simundic AM. Blood sample quality. Diagnosis (Berl) 2019;6: 25–31. doi: 10.1515/dx-2018-0018 [DOI] [PubMed] [Google Scholar]
  • 43.Simundic AM, Baird G, Cadamuro J, Costelloe SJ, Lippi G. Managing hemolyzed samples in clinical laboratories. Crit Rev Clin Lab Sci 2020;57: 1–21. doi: 10.1080/10408363.2019.1664391 [DOI] [PubMed] [Google Scholar]
  • 44.Lippi G, Plebani M, Di Somma S, Cervellin G. Critical Reviews in Clinical Laboratory Sciences Hemolyzed specimens: a major challenge for emergency departments and clinical laboratories 2011. doi: 10.3109/10408363.2011.600228 [DOI] [PubMed] [Google Scholar]
  • 45.Kanji JN, Bailey A, Fenton J, Robbin Lindsay L, Dibernardo A, Toledo NP, et al. Stability of SARS-CoV-2 IgG in multiple laboratory conditions and blood sample types. Journal of Clinical Virology 2021;142: 104933. doi: 10.1016/j.jcv.2021.104933 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Shurrab FM, Al-Sadeq DW, Amanullah F, Younes SN, Al-Jighefee H, Younes N, et al. Effect of multiple freeze-thaw cycles on the detection of anti-SARS-CoV-2 IgG antibodies. J Med Microbiol 2021;70. doi: 10.1099/jmm.0.001402 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Edouard S, Million M, Casalta JP, Collart F, Amphoux B, Raoult D. Low antibodies titer and serological cross-reaction between Coxiella burnetii and Legionella pneumophila challenge the diagnosis of mediastinitis, an emerging Q fever clinical entity. Infection 2017;45: 911–5. doi: 10.1007/s15010-017-1048-6 [DOI] [PubMed] [Google Scholar]
  • 48.La Scola B, Raoult D. Serological cross-reactions between Bartonella quintana, Bartonella henselae, and Coxiella burnetii. J Clin Microbiol 1996;34: 2270–4. doi: 10.1128/jcm.34.9.2270-2274.1996 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Lukacova M, Melnicakova J, Kazar J. Cross-Reactivity between Coxiella burnetii and Chlamydiae. Folia Microbiol 1999;44: 579–84. doi: 10.1007/BF02816263 [DOI] [PubMed] [Google Scholar]
  • 50.Scott RM, Feinsod FM, Allam IH, Ksiazek TG, Peters CJ, Botros,’ And BAM, et al. Serological Tests for Detecting Rift Valley Fever Viral Antibodies in Sheep from the Nile Delta. J Clin Microbiol 1986;24: 612–4. doi: 10.1128/jcm.24.4.612-614.1986 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Swanepoel R, Struthers JK, Erasmus MJ, Shepherd SP, Mcgillivray GM, Erasmus BJ, et al. Comparison of techniques for demonstrating antibodies to Rift Valley fever virus. J Hyg (Lond) 1986;97: 317–29. doi: 10.1017/s0022172400065414 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Szymczak MS, Reeves WK, Miller MM. Further Characterization of Rio Grande Virus and Potential for Cross Reactivity with Rift Valley Fever Virus Assays. Viruses 2021;13. doi: 10.3390/V13091719 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Edouard S, Mahamat A, Demar M, Abboud P, Djossou F, Raoult D. Comparison between emerging Q fever in French Guiana and endemic Q fever in Marseille, France. Am J Trop Med Hyg 2014;90: 915–9. doi: 10.4269/ajtmh.13-0164 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Tumusiime D, Isingoma E, Tashoroora OB, Ndumu DB, Bahati M, Nantima N, et al. Mapping the risk of Rift Valley fever in Uganda using national seroprevalence data from cattle, sheep and goats. PLoS Negl Trop Dis 2023;17: e0010482. doi: 10.1371/journal.pntd.0010482 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Murphy AM, Field PR. The persistence of complement-fixing antibodies to q–fever (coxiella burneti) after infection. Medical Journal of Australia 1970;1: 1148–50. doi: 10.5694/j.1326-5377.1970.tb84481.x [DOI] [PubMed] [Google Scholar]
  • 56.Pepin M, Bouloy M, Bird BH, Kemp A, Paweska J. Rift Valley fever virus (Bunyaviridae: Phlebovirus): an update on pathogenesis, molecular epidemiology, vectors, diagnostics and prevention. Vet Res 2010;41. doi: 10.1051/vetres/2010013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Brown RD, Scott GR, Dalling T. Persistence of antibodies to rift valley fever in man. The Lancet 1957;270: 345. doi: 10.1016/S0140-6736(57)92243-2 [DOI] [Google Scholar]
  • 58.Sukocheva OA, Marmion BP, Storm PA, Lockhart M, Turra M, Graves S. Long-term persistence after acute Q fever of non-infective Coxiella burnetii cell components, including antigens. QJM 2010;103: 847–63. doi: 10.1093/qjmed/hcq113 [DOI] [PubMed] [Google Scholar]
  • 59.Marmion BP, Sukocheva O, Storm PA, Lockhart M, Turra M, Kok T, et al. Review Q fever: persistence of antigenic non-viable cell residues of Coxiella burnetii in the host-implications for post Q fever infection fatigue syndrome and other chronic sequelae. QJM. 2009;102(10): 673–684. doi: 10.1093/qjmed/hcp077 [DOI] [PubMed] [Google Scholar]
  • 60.Marrella V, Facoetti A, Cassani B. Cellular Senescence in Immunity against Infections. Int J Mol Sci 2022;23. doi: 10.3390/IJMS231911845 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Borgoni S, Kudryashova KS, Burka K, de Magalhães JP. Targeting immune dysfunction in aging. Ageing Res Rev 2021;70: 101410. doi: 10.1016/j.arr.2021.101410 [DOI] [PubMed] [Google Scholar]
  • 62.Li ZY, Cai ML, Qin Y, Chen Z. Age/autoimmunity-associated B cells in inflammatory arthritis: An emerging therapeutic target. Front Immunol 2023;14: 1103307. doi: 10.3389/fimmu.2023.1103307 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.Stellfeld M, Gerlach C, Richter IG, Miethe P, Fahlbusch D, Polley B, et al. Evaluation of the Diagnostic Potential of Recombinant Coxiella burnetii Com1 in an ELISA for the Diagnosis of Q Fever in Sheep, Goats and Cattle. Microorganisms 2020;8: 1–19. doi: 10.3390/MICROORGANISMS8081235 [DOI] [PMC free article] [PubMed] [Google Scholar]
PLoS Negl Trop Dis. doi: 10.1371/journal.pntd.0012300.r001

Decision Letter 0

Elvina Viennet, Philip Mshelbwala

19 Aug 2024

Dear DVM Hungerbühler,

Thank you very much for submitting your manuscript "Diagnostic serology test comparison for Q fever and Rift Valley fever in humans and livestock from pastoral communities" for consideration at PLOS Neglected Tropical Diseases. As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. The reviewers appreciated the attention to an important topic. Based on the reviews, we are likely to accept this manuscript for publication, providing that you modify the manuscript according to the review recommendations.

Reviewer #2: Manuscript with reviewer's comments attached

Please prepare and submit your revised manuscript within 30 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email.

When you are ready to resubmit, please upload the following:

[1] A letter containing a detailed list of your responses to all review comments, and a description of the changes you have made in the manuscript.

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out

[2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file).

Important additional instructions are given below your reviewer comments.

Thank you again for your submission to our journal. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Philip Mshelbwala

Academic Editor

PLOS Neglected Tropical Diseases

Elvina Viennet

Section Editor

PLOS Neglected Tropical Diseases

***********************

Reviewer #2: Manuscript with reviewer's comments attached

Reviewer's Responses to Questions

Key Review Criteria Required for Acceptance?

As you describe the new analyses required for acceptance, please consider the following:

Methods

-Are the objectives of the study clearly articulated with a clear testable hypothesis stated?

-Is the study design appropriate to address the stated objectives?

-Is the population clearly described and appropriate for the hypothesis being tested?

-Is the sample size sufficient to ensure adequate power to address the hypothesis being tested?

-Were correct statistical analysis used to support conclusions?

-Are there concerns about ethical or regulatory requirements being met?

Reviewer #1: -Are the objectives of the study clearly articulated with a clear testable hypothesis stated?

Yes

-Is the study design appropriate to address the stated objectives?

Not entirely. My concerns about this are stated in the "Summary and General Comments" box below

-Is the population clearly described and appropriate for the hypothesis being tested?

Yes

-Is the sample size sufficient to ensure adequate power to address the hypothesis being tested?

There was no sample size justification or posthoc power analyses

-Were correct statistical analysis used to support conclusions?

Yes

-Are there concerns about ethical or regulatory requirements being met?

No

Reviewer #2: The objectives of the study are very clear and the study design was appropriate. The researchers gave a clear description of their population, sampling technique, and used relevant statistical analysis. Standard ethical approval was obtained as evidenced by the name of the ethics committee and the approval number. Details of the sample size determination was however not included.

Reviewer #3: The authors evaluated the agreement between a laboratory in Chad and laboratories in Germany and Switzerland. This study is interesting because it sheds light on the reliability of diagnostic tests performed in developing countries, despite the lack of adequate biosecurity facilities. The sampling, sample size, and statistical analysis are sufficient to support the conclusions.

Reviewer #4: The paper aims to investigate the differences between diagnostic tests of Q fever and RVF in livestock and human in different labs. The study design and the statistical analysis would support the testing of the comparison. However, the sampling design was not clearly introduced. The main concerns are: 1. need more demographic information of livestock from the two areas: by species: age, sex, herd sizes, and all livestock from one herd? 2.for the clustering sampling: within each cluster how to choose the apparently healthy ones? how did the authors determine the sample size then?

--------------------

Results

-Does the analysis presented match the analysis plan?

-Are the results clearly and completely presented?

-Are the figures (Tables, Images) of sufficient quality for clarity?

Reviewer #1: -Does the analysis presented match the analysis plan?

Not entirely. My concerns about this are stated in the "Summary and General Comments" box below

-Are the results clearly and completely presented?

Yes

-Are the figures (Tables, Images) of sufficient quality for clarity?

Not entirely. My concerns about this are stated in the summary box below

Reviewer #2: The results were presented in a clear cut manner. Appropriate data visualizations were used to effectively communicate the research findings. Both presented and supplementary figures were sufficient.

Reviewer #3: (No Response)

Reviewer #4: The results are well presented in general. However, a few places could be improved:

1. lines 202-207: this paragraph should be included in the method section

2. Fig 1: Please add the testing methods, institutes and countries aside the buildings so easier for readers to understand

3. line 222: suggest revising "men" to "males" as the studying group of <30 may had included boys

4. Table 3: the table looks incomplete at the bottom? and the "*" needs to explained in a note under.

--------------------

Conclusions

-Are the conclusions supported by the data presented?

-Are the limitations of analysis clearly described?

-Do the authors discuss how these data can be helpful to advance our understanding of the topic under study?

-Is public health relevance addressed?

Reviewer #1: -Are the conclusions supported by the data presented?

Yes

-Are the limitations of analysis clearly described?

Not entirely

-Do the authors discuss how these data can be helpful to advance our understanding of the topic under study?

Yes

-Is public health relevance addressed?

Indirectly; seroprevalence studies provide data for making policies. This study highlights concerns that need to be considered/addressed whenever using seroprevalence data

Reviewer #2: The authors' conclusion was comprehensible and in accordance with the presented data. In addition, their findings were competently compared with current scientific reports. The study limitations and factors likely to be responsible for the disparities observed in the study were highlighted and areas for further research were indicated. The authors elucidated the public health importance and epidemiological relevance of their study.

Reviewer #3: The conclusion is superficial and requires more detail, as indicated in the attached file.

Reviewer #4: In general, the conclusions were well presented. However, the limitation of the study could have been addressed better by discussing why not using the same methods in different labs to compare. For example, authors mentioned the poor consistent may be due to different tests used in different labs, but why can't use the same IFA in Chad to compare? Need to discuss the challenge for having other tests than ELISA in Chad.

Other comments:

1. Lines 270-273: No need to repeat this as it has been introduced.

2. Line 291: please elaborate on the possible misinterpretations

3. Line 321: Since you mentioned the age had impacts on antibody titres, please advise on the implementation of this finding. For example, what if you adjust the cut-offs of reading results?

--------------------

Editorial and Data Presentation Modifications?

Use this section for editorial suggestions as well as relatively minor modifications of existing data that would enhance clarity. If the only modifications needed are minor and/or editorial, you may wish to recommend “Minor Revision” or “Accept”.

Reviewer #1: Minor Revision

Reviewer #2: Minor editorial modifications required have been highlighted in the reviewed manuscript attached to this submission. The authors should include the details of sample size determination. This information is crucial for research reproducibility.

Reviewer #3: I have attached a file detailing suggested changes and indicating areas that need more details. I recommend minor revisions.

Reviewer #4: Minor Revision

--------------------

Summary and General Comments

Use this section to provide overall comments, discuss strengths/weaknesses of the study, novelty, significance, general execution and scholarship. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. If requesting major revision, please articulate the new experiments that are needed.

Reviewer #1: This study sought to address an often-overlooked concern with seroprevalence studies. The sampling technique used, the diagnostic performance of the laboratory test, the diagnostic quality of the sample type, and the lab-to-lab variation in out-of-kit reagents, personnel, and SOPs are important factors that drive the validity of findings from such studies. Specifically, this study evaluates the agreement of Q fever and RVF serological results in a laboratory in Chad compared to results from laboratories in 2 European countries. The findings of this study are most relevant to epidemiologists.

Overall, the authors were clearly able to justify the study and discuss the results in-depth. I have a few concerns:

Major concerns:

Line 138: I missed the portion of the methods that explains how they came about the sample size; it would be great if the authors included some sample size justification, if there was one, or indicate if it was non-probabilistic.

Lines 140 to 161: In studies to investigate the effect of a variable on an outcome, it is best practice to keep every other variable constant (as much as is within the power of the investigator). The authors used only the Q fever Indirect ELISA by IDvet for livestock sample testing in Chad but used the same test only for equids in Switzerland. A different kit was used for ruminants. The authors were silent about this noteworthy discrepancy in their comparisons and discussions. Would the authors mind sharing why this does not influence their findings? Considering the study's design aimed to investigate the lab-to-lab or place-to-place differences in test results, is there a reason why the same tests were not used in the laboratories (also the RVF testing for Humans that used a different test in Europe)?

Lines 138, 148, 203, 206, 209, 218: It is unclear how the 10% mentioned in line 148 matches the sample numbers in 138. The final number of samples mentioned in lines 206 and 218 does not match the numbers in the last row of Table 1. Would the authors be implying that the total number of livestock samples used was 102 because 1 sample that was tested in the Q fever group could not be tested for RVF? If that is the message being conveyed, the authors may want to adjust line 206 to read, “91 human and 102 livestock samples were tested for either one OR both tests.” More text may be added in Figure 1 and other portions of the manuscript to make this clearer for the readers.

Line 197: the function exp(coefficients(model)) does not give an upper CI. Please revise lines 197 and 198

Minor concerns:

Lines 116-119: Please consider rewording your objective to be easier understood or piecing the objective into more than one sentence.

Figure 1: The authors may consider adding more details to the figure and title to let the readers grasp the full scope of what is being described without referring to the main manuscript text.

Reviewer #2: It is evident that the authors put a lot of effort into writing a well structured and detailed manuscript. The manuscript presents a clear purpose, informative literature review with clearly defined terms that can be easily understood by any reader. The research methods were also comprehensibly described and the results were well explained. The discussion has developed ideas and the authors provided relevant supporting data for their claims. There was a smooth transition between sections, giving a high level of cohesion and an overall high quality manuscript.

The study is of utmost relevance in disease detection and control. The authors identified the need for considering demographics in disease detection which is novel and of great significance in field and laboratory studies, as well as in the development of diagnostic protocols. Hence, the manuscript is of high scientific value.

Reviewer #3: See attached file

Reviewer #4: The manuscript offers great insights on the challenge in disease diagnosis using methods in the context of poor resourced area. This paper could be of interest for local to regional levels readers. Manuscript topicality is excellent.

--------------------

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: Yes: Dr. Ifeoluwapo Omolola Akanbi

Reviewer #3: No

Reviewer #4: No

Figure Files:

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org.

Data Requirements:

Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5.

Reproducibility:

To enhance the reproducibility of your results, we recommend that you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option to publish peer-reviewed clinical study protocols. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

References

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article's retracted status in the References list and also include a citation and full reference for the retraction notice.

Attachment

Submitted filename: Reviewer 2_reviewed MS after first submission.pdf

pntd.0012300.s006.pdf (1.3MB, pdf)
Attachment

Submitted filename: PNTD-D-24-00848_Review.docx

pntd.0012300.s007.docx (16.2KB, docx)
PLoS Negl Trop Dis. doi: 10.1371/journal.pntd.0012300.r003

Decision Letter 1

Elvina Viennet, Philip Mshelbwala

30 Sep 2024

Dear DVM Hungerbühler,

We are pleased to inform you that your manuscript 'Diagnostic serology test comparison for Q fever and Rift Valley fever in humans and livestock from pastoral communities' has been provisionally accepted for publication in PLOS Neglected Tropical Diseases.

Before your manuscript can be formally accepted you will need to complete some formatting changes, which you will receive in a follow up email. A member of our team will be in touch with a set of requests.

Please note that your manuscript will not be scheduled for publication until you have made the required changes, so a swift response is appreciated.

IMPORTANT: The editorial review process is now complete. PLOS will only permit corrections to spelling, formatting or significant scientific errors from this point onwards. Requests for major changes, or any which affect the scientific understanding of your work, will cause delays to the publication date of your manuscript.

Should you, your institution's press office or the journal office choose to press release your paper, you will automatically be opted out of early publication. We ask that you notify us now if you or your institution is planning to press release the article. All press must be co-ordinated with PLOS.

Thank you again for supporting Open Access publishing; we are looking forward to publishing your work in PLOS Neglected Tropical Diseases.

Best regards,

Philip Mshelbwala

Academic Editor

PLOS Neglected Tropical Diseases

Elvina Viennet

Section Editor

PLOS Neglected Tropical Diseases

***********************************************************

Reviewer's Responses to Questions

Key Review Criteria Required for Acceptance?

As you describe the new analyses required for acceptance, please consider the following:

Methods

-Are the objectives of the study clearly articulated with a clear testable hypothesis stated?

-Is the study design appropriate to address the stated objectives?

-Is the population clearly described and appropriate for the hypothesis being tested?

-Is the sample size sufficient to ensure adequate power to address the hypothesis being tested?

-Were correct statistical analysis used to support conclusions?

-Are there concerns about ethical or regulatory requirements being met?

Reviewer #1: Yes

Reviewer #2: The objectives of the study were clearly articulated and the study design well stated. The study population and sample size determination were properly described and appropriate for the tested hypothesis. The sample size was adequate and the statistical analysis were appropriate. Ethical approvals were duly obtained from appropriate institutions.

Reviewer #4: (No Response)

**********

Results

-Does the analysis presented match the analysis plan?

-Are the results clearly and completely presented?

-Are the figures (Tables, Images) of sufficient quality for clarity?

Reviewer #1: Yes

Reviewer #2: The results were comprehensive and well presented. The contents of the tables were adequate and comprehensible. High quality data visualizations were used.

Reviewer #4: (No Response)

**********

Conclusions

-Are the conclusions supported by the data presented?

-Are the limitations of analysis clearly described?

-Do the authors discuss how these data can be helpful to advance our understanding of the topic under study?

-Is public health relevance addressed?

Reviewer #1: Yes

Reviewer #2: The authors' conclusions were in line with the results of the study and their findings were extensively compared with other researchers' submissions.

Reviewer #4: (No Response)

**********

Editorial and Data Presentation Modifications?

Use this section for editorial suggestions as well as relatively minor modifications of existing data that would enhance clarity. If the only modifications needed are minor and/or editorial, you may wish to recommend “Minor Revision” or “Accept”.

Reviewer #1: Accept

Reviewer #2: The manuscript has been improved upon after the first review and it is recommended for acceptance for publication.

Reviewer #4: (No Response)

**********

Summary and General Comments

Use this section to provide overall comments, discuss strengths/weaknesses of the study, novelty, significance, general execution and scholarship. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. If requesting major revision, please articulate the new experiments that are needed.

Reviewer #1: The authors have satisfactorily addressed my concerns

Reviewer #2: The study is very interesting and highly relevant. The authors adopted cutting edge methodology. Hence, the study is reproducible. The manuscript was well written, properly structured and would be easily understood by readers from a wide range of fields.

Reviewer #4: my concerns have been addressed and I recommend that the paper can be accepted.

**********

PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: Yes: Dr. Ifeoluwapo Akanbi

Reviewer #4: No

PLoS Negl Trop Dis. doi: 10.1371/journal.pntd.0012300.r004

Acceptance letter

Elvina Viennet, Philip Mshelbwala

8 Oct 2024

Dear DVM Hungerbühler,

We are delighted to inform you that your manuscript, "Diagnostic serology test comparison for Q fever and Rift Valley fever in humans and livestock from pastoral communities," has been formally accepted for publication in PLOS Neglected Tropical Diseases.

We have now passed your article onto the PLOS Production Department who will complete the rest of the publication process. All authors will receive a confirmation email upon publication.

The corresponding author will soon be receiving a typeset proof for review, to ensure errors have not been introduced during production. Please review the PDF proof of your manuscript carefully, as this is the last chance to correct any scientific or type-setting errors. Please note that major changes, or those which affect the scientific understanding of the work, will likely cause delays to the publication date of your manuscript. Note: Proofs for Front Matter articles (Editorial, Viewpoint, Symposium, Review, etc...) are generated on a different schedule and may not be made available as quickly.

Soon after your final files are uploaded, the early version of your manuscript will be published online unless you opted out of this process. The date of the early version will be your article's publication date. The final article will be published to the same URL, and all versions of the paper will be accessible to readers.

Thank you again for supporting open-access publishing; we are looking forward to publishing your work in PLOS Neglected Tropical Diseases.

Best regards,

Shaden Kamhawi

co-Editor-in-Chief

PLOS Neglected Tropical Diseases

Paul Brindley

co-Editor-in-Chief

PLOS Neglected Tropical Diseases

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Table. This document (S1A–S1F Tables) provides a summary of the test protocols for all diagnostic assays employed in this study.

    (DOCX)

    pntd.0012300.s001.docx (20KB, docx)
    S2 Table. This table presents the demographic characteristics of the sample population.

    (DOCX)

    pntd.0012300.s002.docx (17.2KB, docx)
    S3 Table. This table presents the findings from the univariable logistic regression models.

    (DOCX)

    pntd.0012300.s003.docx (18.3KB, docx)
    S1 Dataset. This dataset consists of the data required to replicate all study findings reported in the article for the human population.

    (XLSX)

    pntd.0012300.s004.xlsx (17.2KB, xlsx)
    S2 Dataset. This dataset consists of the data required to replicate all study findings reported in the article for the livestock population.

    (XLSX)

    pntd.0012300.s005.xlsx (19.4KB, xlsx)
    Attachment

    Submitted filename: Reviewer 2_reviewed MS after first submission.pdf

    pntd.0012300.s006.pdf (1.3MB, pdf)
    Attachment

    Submitted filename: PNTD-D-24-00848_Review.docx

    pntd.0012300.s007.docx (16.2KB, docx)
    Attachment

    Submitted filename: response to reviewers.docx

    pntd.0012300.s008.docx (221.5KB, docx)

    Data Availability Statement

    All relevant data are within the manuscript and its Supporting Information files. All data are fully anonymized to protect the privacy of the participants involved in the study.


    Articles from PLOS Neglected Tropical Diseases are provided here courtesy of PLOS

    RESOURCES