Too many errors in scientific publications, too many in plant sciences, too
Today, there is an astonishing volume of scientific information available for researchers, which can be easily accessed through powerful search tools. Yet, the question now is whether all this vast amount of information is reliable. In this sense, a “bad science” controversy arose recently when many Open Access (OA) journals (more than a half) published a false, error-ridden paper, which had been submitted in order to test the publishing ethics of these journals (Bohannon, 2013). This fake article was published mainly by fraudulent journals, but it was also accepted by a number of OA journals of renowned publishers with peer-review systems. The failure to reject an article full of errors revealed that the system's gearbox is leaking somewhere. The carelessness of peer-reviews in a number of OA journals has opened a Pandora's Box, and what is more disconcerting, nobody can guarantee that it could not also affect regular journals (non OA). Traditionally, it has been assumed that scientific journals should detect and correct all these failings through the peer review before publication. Regrettably, as we show in this communication, the system is far from being perfect (Pulverer, 2010; Székely et al., 2014).
Whilst the detection of laboratory errors is an issue of great attention in medicine (Bonini et al., 2002; Carraro and Plebani, 2007; Hammerling, 2012), in experimental science, it does not seem to be a crucial task. However, we were aware of this concern when we performed a literature compilation with the aim of providing a comprehensive evaluation of the responses of photosynthetic pigment composition to environmental conditions (Esteban et al., 2015). In this survey, we compiled data from 525 papers from the last 21 years (1991–2011). After carrying out a critical analysis of the data, a considerable number of papers, 96 out of 525, were found to have data out of range, errors and inconsistencies in at least one of the parameters reported. In order to detect these errors, we used as an initial screening tool three standard deviations from the mean. Data outside this interval were subsequently examined (Osborne and Overbay, 2004), in order to identify whether errors had arisen from the inherent variability of the data or from mistakes in the data (for a detailed description see Esteban et al., 2015). We decided to carry out an in-depth analysis of these controversial data, establishing a new background for our study. In this sense, data was re-evaluated and classified on the basis of the type of error: I, II, and III. Error I (n = 46) includes those articles containing values out of range (outside ±3 standard deviation from the mean), likely due to pre-analytical and analytical flaws. Error II (n = 37) refers to the presence of wrong values (most of them higher than 1000-fold reference values), most likely caused by post-analytical errors. Error III (n = 13) includes those articles with mistaken units, probably included during the final phase of publishing and editing.
Discarding the non-ethical manipulation of data, errors may occur as a result of experimental (analytical or methodological), mathematical or editing errors. For the case studied here (pigment determinations), we have identified several potential sources of error, which may occur at any stage of the research process: (i) Methodological errors during pre-analytical and analytical phase: inappropriate specimen/sample collection and preservation, labeling errors, wrong biomass/leaf area measurement, incomplete extraction, malfunction of instruments, incorrect compound identification, pipetting errors, etc. (ii) Data analysis errors in the post-analytical phase: handling of mathematics in spreadsheets, mistakes in the preparation of graphics or tables, improper data entry, and failure in reporting; (iii) Publishing and editing errors: confusion in units, typing errors such as Latin instead of Greek letters or errors in graph scales (see Table S1 for complete list of errors, tips and solutions). The data published in regular papers do not allow an assessment to be made of which of these aforementioned error sources is the cause of the mistakes. However, in other fields, such as analytical medicine, in which traceability is easier thanks to the application of quality assurance protocols, most errors occur during the pre- and post-analytical phases (Hammerling, 2012).
How, where and when these errors appear
As an example of the inaccuracy of pigment measurements, we have performed an analysis of data published on pigment composition in the model plant Arabidopsis thaliana ecotype Columbia including all the available literature published in the major journals in this field during the period 1991–2011 (see Esteban et al., 2015 for details). It is assumed that plants cultivated under similar conditions in different laboratories should not differ greatly in their biochemical attributes; otherwise A. thaliana would not be a good “model species.” To reduce the noise we considered only pigment ratios that are supposed to be much more stable than absolute concentrations. However, as is shown in Figure 1A, when pigment composition was plotted against cumulative daily irradiance, a factor that has been described as the main determinant of chlorophyll and carotenoid contents (Niinemets et al., 2003), no pattern of response was observed. Furthermore, it varied greatly (5-fold for some parameters). As it is unlikely that these ranges of variation represent true differences in pigment composition (carotenoids and chlorophylls in green tissues are bound to proteins with their strongly regulated and interdependent proportions and concentrations), they are probably a reflection of the lack of methodological accuracy probably affecting both pigment analysis itself and imprecisions regarding growing procedures or their descriptions throughout the manuscripts.
In general, it seems reasonable to consider that these errors and this lack of accuracy may stem from the careless application of protocols, from sample collection to data calculations and even final publication. In this sense one might expect that the scientific tradition of a country or the quality of the journal in which the research was performed would contribute to filtering out inconsistent results. In order to disentangle the influence of this kind of factors on the frequency of errors we have described the following variables for each discarded reference: (i) The journals' impact factor (IF) as an index of quality, (ii) the year of publication, (iii) the scientific heritage of the authors' country, estimated by the number of Nobel laureates, and (iv) the countries' current investment in Research and Development (R&D) (Figure 1B). Interestingly, our analysis shows that, perhaps just by excluding the scientific tradition factor, the frequency of errors escapes these biases, being similar both in groups from countries that invest a lot in science and have a consolidated research tradition, and in those from emerging economies with low investments in R&D (Figure 1B). Furthermore, even the quality of the journal (measured by the impact factor) does not filter out unreliable results (Figure 1B). Even more striking is the fact that, despite the technical and methodological improvements that have been made during the last two decades, the frequency of errors detected in the present survey shows a clearly upward trend (from 12 to 22% over a 21-year period) (Figure 1B). Perhaps this is the consequence of the pressure exerted by the well-known “publish or perish” dilemma that forces the rapid and careless publication of data. It should be noted here that we have focused our study on numerical errors that affect the quantitative expression of a value. Therefore, we have not considered other types of errors (incorrect use of statistics, experimental design, conceptual error and discursive errors).
The question that arises now is whether this inaccuracy relates only to pigment determinations, or whether this is just the tip of the iceberg, it being indicative of a general malaise in plant science studies. Although some studies have analyzed statistical limitations and biases, as well as methodological and statistical errors in ecology and plant science (Grime and Sinclair, 1999; Ross and Sulev, 2000; Heinemann et al., 2002; Dieleman and Janssens, 2011; Curtis and Queenboroughn, 2012; Hulme et al., 2013; Mudge, 2013) very little has been done to detect and quantify the appearance of quantitative errors. So, without further analyses, it is, of course, impossible to respond to this essential question. However, it is noteworthy that one of the simplest determinations a plant biology lab can perform is pigment content (relatively robust molecules, present at high concentrations, can be directly measured thanks to their absorbance in the visible range, their study does not require investments in expensive equipment). Thus, finding these incongruences regarding the data reported for these parameters should cause a stir among the entire scientific community.
Lessons for the future: “errare humanum est, sed in errare perseverare diabolicum.”
This is a Latin quote attributed to Seneca, which means “Anyone can err, but to persist in error is diabolical.” Two thousand years later, in the “omics” era, when model species such as Arabidopsis are used as genomic reference material, such a degree of inaccuracy in analytical determinations is unacceptable, and the search for solutions is a must for the entire community of plant researchers. What is in question is not the reputation of researchers but the quality, accuracy and reproducibility of the results. All the actors in the scientific chain of activity, from the lab to the publishing house (technicians, researchers, authors, reviewers, editors and even readers), should be involved in the job of detecting and removing errors. Among them, the role of researchers and reviewers is of particular importance for the detection of inconsistent values.
The detection of errors has not been a crucial task in most plant biology research groups. However, in view of the experience gained by the methods employed by analytical laboratories to enhance the quality of results (Smagunova et al., 2006; Su et al., 2011), the reliability of data can be substantially improved by the implementation of inter-laboratory comparison tests. For pigments, these proficiency tests should be based on standard reference plant photosynthetic material (the standards for calibrating HPLC are not in themselves a guarantee) together with the availability of reference values for “normal” ranges (as those reported in Esteban et al., 2015) of pigment contents.
In the publication process, reviewers are considered to be infallible gods, and they must guarantee the quality of the experimental work, adequate data interpretation and the removal of all inaccuracies before publication. However, reviewers are frequently overloaded with too many commitments, and are forced to review manuscripts at odd moments during the day. Paradoxically, in the research publication chain, reviewers are one of the essential links, but at the same time they are the only amateur component (in the sense that their work is not paid), and their contribution is frequently not sufficiently recognized. Aditionally, as Plant Science studies are becoming interdisciplinary, classical review process is very limited. The correction policies of some journals that allow the publication of an amendment note or even the re-publication of a corrected version of the whole article are commendable. In that sense, the role of the whole scientific community as critical readers becomes an inescapable duty. A stronger implication of editors and readers, together with the presentation of results in a traceable form would also contribute to reducing the number of quantitative results in doubt. Also the creation of databases about plant attributes, where it could be easy to check whether a measured parameter falls into the expected range of values or not, could help in the prevention of incorrectness. Thus, this work does not set out to be a Casus Belli, but what it does say is that we must all try harder. Luckily, in the theater of scientific publications, the same actors (we researchers) have to play different roles (authors, readers, reviewers and editors) and this unusual trait of our field of professional activity should warrant the easy implementation of policies of quality assurance. In conclusion, improvements in data correctness could be done in three directions: (i) careful check of all steps in data obtention, analysis and publication (details in Table S1), (ii) more open review system, with easy postpublication procedures, and (iii) generation of databases of plant attributes that could be used as reference.
Conflict of interest statement
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Acknowledgments
We acknowledge Dr. Daniel Marino for helpful comments and discussion and Brian Webster for English revision. Supporting from BFU 2010-15021, UPV/EHU-GV IT-624-13 grant, JAE-Doc-2011-046 fellow from the Spanish National Research Council (CSIC) received by RE and Marie Curie IEF grant (328370 MELISSA) from the European FP7-PEOPLE received by BF-M, are also acknowledged.
Supplementary material
The Supplementary Material for this article can be found online at: http://www.frontiersin.org/journal/10.3389/fpls.2015.00148/full
References
- Bohannon J. (2013). Who's afraid of peer review? Science 342, 60–65. 10.1126/science.342.6154.60 [DOI] [PubMed] [Google Scholar]
- Bonini P., Plebani M., Ceriotti F., Rubboli F. (2002). Errors in laboratory medicine. Clin. Chem. 48, 691–698. [PubMed] [Google Scholar]
- Carraro P., Plebani M. (2007). Errors in a stat laboratory: types and frequencies 10 years later. Clin. Chem. 53, 1338–1342. 10.1373/clinchem.2007.088344 [DOI] [PubMed] [Google Scholar]
- Curtis P. S., Queenboroughn S. A. (2012). Raising the standards for ecological meta-analyses. New Phytol. 195, 279–281. 10.1111/j.1469-8137.2012.04207.x [DOI] [PubMed] [Google Scholar]
- Dieleman W. I. J., Janssens I. A. (2011). Can publication bias affect ecological research? A case study on soil respiration under elevated CO2. New Phytol. 190, 517–521. 10.1111/j.1469-8137.2010.03499.x [DOI] [PubMed] [Google Scholar]
- Esteban R., Barrutia O., Artetxe U., Fernandez-Marin B., Hernandez A., Garcia-Plazaola J. I. (2015). Internal and external factors affecting photosynthetic pigment composition in plants: a meta-analytical approach. New Phytol. 206, 268–280. 10.1111/nph.13186 [DOI] [PubMed] [Google Scholar]
- Grime V. L., Sinclair F. L. (1999). Sources of error in stem heat balance sap flow measurements. Agric. For. Meteorol. 94, 103–121 10.1016/S0168-1923(99)00011-8 [DOI] [Google Scholar]
- Hammerling J. A. (2012). A review of medical errors in laboratory diagnostics and where we are today. Labmedicine 43, 41–44 10.1309/LM6ER9WJR1IHQAUY [DOI] [Google Scholar]
- Heinemann A. B., Hoogenboom G., Chojnicki B. (2002). The impact of potential errors in rainfall observation on the simulation of crop growth, development and yield. Ecol. Model. 157, 1–21 10.1016/S0304-3800(02)00209-0 [DOI] [Google Scholar]
- Hulme P. E., Pysek P., Jarosík V., Pergl J., Schaffner U., Vilà M. (2013). Bias and error in understanding plant invasion impacts. Trends Ecol. Evol. 28, 212–218. 10.1016/j.tree.2012.10.010 [DOI] [PubMed] [Google Scholar]
- Mudge J. F. (2013). Explicit consideration of critical effect sizes and costs of errors can improve decision-making in plant science. New Phytol. 199, 876–878. 10.1111/nph.12410 [DOI] [PubMed] [Google Scholar]
- Niinemets Ü., Kollist H., García-Plazaola J. I., Hernández A., Becerril J. M. (2003). Do the capacity and kinetics for modification of xanthophyll cycle pool size depend on growth irradiance in temperate trees? Plant Cell Environ. 26, 1787–1801 10.1046/j.1365-3040.2003.01096.x [DOI] [Google Scholar]
- Osborne J. W., Overbay A. (2004). The power of outliers (and why researchers should always check for them). Pract. Assess Res. Eval. 9:e1. [Google Scholar]
- Pulverer B. (2010). Transparency showcases strength of peer review. Nature 46, 30–31. 10.1038/468029a [DOI] [PubMed] [Google Scholar]
- Ross J., Sulev M. (2000). Sources of errors in measurements of PAR. Agr. Forest Meteorol. 100, 103–125 10.1016/S0168-1923(99)00144-6 [DOI] [Google Scholar]
- Smagunova A. N., Pan'kov S. D., Kozlov V. A. (2006). Management of quality control in analytical laboratories. J. Anal. Chem. 61, 312–319 10.1134/S1061934806040022 [DOI] [Google Scholar]
- Su Y., Hung H., Stern G., Sverko E., Lao R., Barresi E., et al. (2011). Bias from two analytical laboratories involved in a long-term air monitoring program measuring organic pollutants in the Arctic: a quality assurance/quality control assessment. J. Environ. Monitor. 13, 3111–3118. 10.1039/c1em10513b [DOI] [PubMed] [Google Scholar]
- Székely T., Krüger O., Krause E. T. (2014). Errors in science: the role of reviewers. Trends Ecol. Evol. 29, 371–373. 10.1016/j.tree.2014.05.001 [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.