Abstract
Academic graphs are essential for communicating complex scientific ideas and results. To ensure that these graphs truthfully reflect underlying data and relationships, visualization researchers have proposed several principles to guide the graph creation process. However, the extent of violations of these principles in academic publications is unknown. In this work, we develop a deep learning-based method to accurately measure violations of the proportional ink principle (AUC = 0.917), which states that the size of shaded areas in graphs should be consistent with their corresponding quantities. We apply our method to analyze a large sample of bar charts contained in 300K figures from open access publications. Our results estimate that 5% of bar charts contain proportional ink violations. Further analysis reveals that these graphical integrity issues are significantly more prevalent in some research fields, such as psychology and computer science, and some regions of the globe. Additionally, we find no temporal and seniority trends in violations. Finally, apart from openly releasing our large annotated dataset and method, we discuss how computational research integrity could be part of peer-review and the publication processes.
Author summary
Scientific figures are one of the most effective ways of conveying complex information to fellow scientists and the general public. While figures have the great power to leave a strong impression, they can also confuse and even mislead readers. Visualization researchers suggest several rules to avoid these issues. One such rule is the Principle of Proportional Ink (PPI) which suggests the size of shaded areas in graphs to be consistent with their corresponding quantities. The extent of violations of this principle in scientific literature is unknown, and methods for detecting it are lacking. In this article, we develop a deep learning-based method to perform such tasks on bar charts. An analysis of hundreds of thousands of figures revealed that around 5% of bar charts have violations of PPI. We found differences in the prevalence of these issues across fields and countries. We discuss the implications of these results and the opportunities and challenges posed by technology such as the one proposed here.
This is a PLOS Computational Biology Methods paper.
1. Introduction
Scientists need to understand previous research so as to build on others’ work. This goal is best achieved when research findings are conveyed accurately and transparently. While inaccuracies can be the result of honest mistakes, some are also research integrity matters [1]. Some studies have estimated that 4% to 9% of scientists have been involved in research misconduct or have observed others’ research misconduct [2,3]. However, assigning intentionality can be highly problematic, requiring expert opinion based on quantitative analysis. Regardless of intentionality, early, accurate, and scalable detection of potential problems during pre- and post-publication are crucial steps to make science more robust.
Publications contain a range of media, such as text, images, data, and graphs. Two of the most commonly studied media for integrity issues are text and images. Potential issues in text have revealed quality problems in English writing [4] and plagiarism [5]. Automated analysis of text is widely available, with several open, commercial, and free software to detect them [6]. Image-based media is more difficult to analyze because of the difficulty in parsing images. Studies driven by human examination have found that common manipulations of graphs and photographic images can have mistakes or integrity violations [7,8]. Some of these manipulations, such as adjusting the size and brightness of images, are common but might be misused and questionable. Several guidelines have been proposed to avoid these problems [9,10]. Still, recent advances in automatically detecting image issues offer auspicious results akin to text analysis [11,12]. Thus, text and images are commonly analyzed and have promising advances in automated analysis.
Scientific graphs tend to be the primary way of showing quantitative information in a visually-appealing manner. Compared to text and image integrity analysis, they are comparability much less studied. Still, graph issues can lead readers to misinterpret information [13], affecting their ability to make decisions [14]. Even if the readers are trained to detect problems, they can still be confused or misled [15]. Several general principles for graph construction have been proposed to avoid these issues [16]. One of these principles is the principle of proportional ink. It is a specialized rule derived from one of the general graphical integrity principles introduced by [16] who stated that "the representation of numbers, as physically measured on the surface of the graphic itself, should be directly proportional to the numerical quantities represented." Further, the derivation of the definition is succinctly presented by [17], stating that "when a shaded region is used to represent a numerical value, the size (i.e., area) of that shaded region should be directly proportional to the corresponding value." The violations of these guidances can be considered as graphical integrity issues but have been much less studied.
Researchers have proposed automated and scalable techniques to detect violations of research integrity and responsible research practices. The present study looks at another kind of scientific imagery that has been somewhat less explored before: graphs. To the best of our knowledge, we do not know the seriousness of graphical integrity issues among scientific graphs. Due to the nature of graphs, they are not shared in a machine-readable format, so it is not possible to apply methods such as those applied to text. Also, scientific graphs, such as line and scatter plots, have much less contrast than photographic figures, making it difficult to apply computer vision techniques to extract information as these methods usually rely on detecting regions of interest defined by contrast differences. With the availability of new datasets about scientific figures and open access publications, we have unprecedented opportunities to understand the extent of issues in graphs. However, we lack techniques and curated datasets to create methods to detect them effectively. While we cannot assume the intentionality behind the authors when a proportional ink principle violation is found, our work can act as a pre-filter for authors, journals, or parties interested in double-checking whether their graphical information presentation passes this basic filter of information visualization.
This study develops a novel method based on deep learning to detect graphical integrity issues automatically. We specifically focus on proportional ink principle violations of bar charts. In our study, a proportional ink violation is a y-axis that does not start from zero or has a truncation in its scale. Importantly, our framework is general enough that it can be adapted to other kinds of graphical integrity issues (e.g., data-ink principle). Armed with our technique (estimated to have an AUC of 0.917), we seek to answer how common these kinds of problems are in academia and whether there are systematic differences among researchers, journals, fields, and countries. Further, we examine whether graphical integrity issues “run in the family”: does being a co-author of an article with issues in the past predict problems in the future? Finally, we discuss applications of these ideas beyond the proportional ink principle and graphical integrity domain.
2. Literature review
2.1. Graphical integrity
In visualization design, graphical integrity requires designers to create graphs reflecting the actual data. With this purpose, researchers have proposed several graphical integrity principles in the literature, and two of the most common are the principle of proportional ink and the principle of data-ink [16]. The principle of proportional ink was proposed by [17] and was inspired by Tufte, who stated that “the representation of numbers, as physically measured on the surface of the graphic itself, should be directly proportional to the numerical quantities represented” [16]. The principle of data-ink states that the graphs should use most of the ink to present the data instead of using it to show the background and aesthetic features. From the perspective of visual perception theory, the principle of proportional ink is based on how humans perceive the size of symbols in graphs [18] and the ratio of entities in graphs [19]. According to empirical visualization studies, violations of these principles can lead viewers to misunderstand the underlying data and results [13].
Researchers in business have extensively investigated graphical integrity in corporate reports and other forms of financial information, uncovering graph distortions and truncations of bar charts and other types of graphs [20,21]. Some researchers have examined the effects of these distortions and found they can greatly mislead viewers [22]. We would expect that these potential confusions would translate into scientific graphs as well.
Studies about graphical integrity issues in science are, in contrast, much less common. These issues, however, can be part of more significant problems such as misinformation [23], and therefore are essential to understand. Moreover, the degradation of information quality from science to the public might exacerbate the problems: researchers have found inadequate reporting of research in the news [24] and exaggerations of research from press releases [25].
2.2. Chart mining
In principle, researchers have the technology to investigate the graphical integrity of science automatically. One major step for evaluating graphic integrity automatically is to examine graphs with chart mining techniques. Several studies have attempted to achieve this step through a broad field called chart mining [26]. For example, one study showed that we could tell the field of a manuscript by automatically analyzing its figures [27]. Another study developed machine learning techniques to extract information from graphs, including bar charts, line plots, and scatter plots [28–30]. For research integrity purposes, the key is to coordinate techniques from the field of chart mining and have large amounts of expertly annotated data. The curated data would allow machine learning models to learn what constitutes a violation of graphic integrity or not. We review some standard techniques in chart mining relevant to our research.
Chart extraction. Chart extraction is a relatively mature field that has enabled researchers to extract figures from scientific publications. Some of the chart extraction techniques exploit metadata available in PDFs or are based on computer vision algorithms that guess the location and size of figures within a document [31,32]. However, sometimes metadata might not be available, and therefore we can only rely on computer vision. With the development of neural networks, computer scientists have human-annotated documents to train deep convolutional neural networks to detect charts in documents without such metadata. The accuracy of these techniques has increased substantially over the years [33].
Subfigure separation. After chart extraction from publications, some figures might be compounded. Subfigure separation is the task that splits these compound figures into their subfigures [34]. Researchers have proposed several methods to achieve this separation, either using traditional computer vision algorithms such as edge-based detection of spaces between panels [35,36]. Interestingly, similar to what has happened with chart extraction, subfigure separation has benefited and improved from new deep learning techniques and large annotated compound figure datasets [37]. Although there is still room to improve, subfigure separation is mature enough to be used in production systems such as Semantic Scholar.
Chart classification. Different types of charts present different kinds of relationships. A bar chart might have a different way of expressing the proportional ink principle compared to a pie plot [16,38]. Therefore, we should ideally understand the kind of chart we are dealing with. The performance of this classification also has a cascading effect: the better we classify them, the more accurate our downstream analyses are. Researchers have used a range of techniques from traditional computer vision algorithms [39] to deep learning techniques [40] to perform this task. Currently, deep learning-based chart classification techniques have promising performance [34].
Data extraction. Data extraction is the most complex step as it attempts to reverse engineer the data used to produce the chart. A bar chart, plot, and pie chart panels might contain valuable information to be extracted from the image, such as point locations, tick values, and statistical information. It usually combines text detection, optical character recognition (OCR), keypoint analysis, and other advanced computer vision methods [41]. Neural networks have shown surprising performance at this task [42,43]. Depending on the type of graph, specialized data extraction techniques exist (e.g., bar charts or pie charts use different methods; see [44]). Data extraction is an active area of research that holds promising value for research integrity.
2.2.1. Datasets to analyze science: open access publications and scientific imagery
Chart mining techniques benefit from large datasets. Only recently, we have access to a much broader set of information pieces about science than ever before. These datasets include citations and other artifacts (e.g. [45,46]). Due to the push for open access publications, we also have unprecedented access to full text, citation contexts, figures, tables, and datasets [47]. These new datasets open up a wealth of opportunities to understand graphical integrity issues with data-driven techniques.
3. Results
3.1. Dataset annotation and method evaluation results
3.1.1. Labeled dataset for evaluation
To evaluate our method, we manually annotated a sample of bar charts from images of the large PubMed Central Open Access Subset (PMOAS) collection (see Materials and Methods). The PMOAS is part of the total collection of articles in PubMed Central. Our annotation pipeline starts with a random sample of 50,000 figures (see S1 Table). We then perform compound figure separation and bar chart detection (i.e., figure classification) (see Fig 1). The process detected 8,001 subfigures or figures as bar charts (Table 1). Then, two annotators (co-authors) manually annotated these bar charts as having proportional ink principle violations following the guidelines of [16,17]—more details in section 4.1.1. We found that 356 figures were classified as having integrity issues by at least one annotator. Our two human annotators had a high inter-rater agreement of 0.74 Cohen’s Kappa [48]—a > 0.7 Cohen’s Kappa is considered substantial agreement [49]. This high agreement among our annotators gives us confidence that there is a high potential to distill the rules that humans are using to classify violations.
Table 1. Summary of Training and Validation for Image Preprocessing.
Steps | Training data | Validation Data | Our Performance | Benchmark |
---|---|---|---|---|
Compound Figure Detection | ImageClef Medical 2016 Compound Figure Task | ImageClef Medical 2016 Compound Figure Task | Accuracy: 92% | Accuracy: 92% (Top one team in ImageClef Medical 2016) |
Subfigure Separation | ImageClef Medical 2016 Compound Figure Task | ImageClef Medical 2016 Subfigure Separation Task | Score: 83% | Score: 84% (Top one team of ImageClef Medical 2016) |
Chart Classification | Our generated charts | Revision data (not fully available) | Accuracy: 100% | Accuracy: 80%1 |
Text Localization | Localization from ArXiv papers through pdffigures | Our generated charts | F1: 76% | F1: 88%2 |
Text Recognition | No training data (Used an open-source fine-tuned model) | Our generated charts | F1: Exact 82% / Edit 90% | F1: Exact 95%/ Edit 98%2 |
Text Role Classification | No training data (Used an open-source fine-tuned model) | Our generated charts | F1: 80% | F1: 100%2 |
Because our deep learning-based method needs to automatically extract features from the bar charts (e.g., tick labels, the start of the y-axis, etc.), we could only produce a reduced set of data from the 8,001 annotated bar charts. In particular, our method could only extract features from 4,834 of them (with integrity issue: 265, without integrity issue: 4569). This data was used for training the machine learning model in our deep learning-based method.
3.1.2. Evaluation of the deep-learning-based method for estimating graphical integrity issues
We used Area under the ROC curve (denoted as AUC), precision, and recall with stratified 5-fold cross-validation to evaluate the performance of our deep learning-based method (for our evaluation protocol and our method, see Materials and Methods section 4.2.3). AUC indicates the ability of our detector to classify two classes. Our average AUC across folds is 0.917, with a standard deviation of 0.02. In addition, the average precision across folds is 0.77 with 0.0209 standard error. This result shows that around 77% of our positive predictions are true positive violations. The average recall across folds is 0.37 with 0.0207 standard error. In sum, these results reveal our method has its strength in making sure positive predictions are true while being relatively less successful with false negatives (see section 3.1.4 for error analysis).
3.1.3. Representation quality of unlabeled data vs. labeled data
Once we have our deep learning-based model trained, we set to analyze a large sample of data that is unlabeled. This is part of the prediction pipeline in Table 1. Before doing so, we want to make sure that the unlabeled data is similar to the labeled data in terms of their features. To do this, we compare the distribution of features described in section 4.1.2. Because we randomly sampled both datasets, we expected them to be similar. Indeed, we found that the unlabeled and labeled distributions are nearly identical (Fig 2).
3.2. Graphical integrity in open access journals
With this deep learning-based method, we can examine graphical integrity issues at scale. In the present study, we apply this method to open access journals to investigate the commonness and patterns of graphical integrity issues in science. In the following subsections, we first present a descriptive analysis of our data. Then, we offer analysis by countries, research fields, years, seniority of researchers, and the journal’s impact. We present our analysis result for 250,000 predictions (prediction pipeline in Table 1) in the main text (S1 Text contains a similar analysis was done for the 50,000 human-labeled figures).
3.2.1. Estimated prevalence of graphical integrity issues
Although our annotated data contained 8,001 bar charts, only 4,834 of them were processable by our pipeline (S1 Table). This process involved computing the features for our classification model (see methods). The reason for this drop is that our automated pipeline was sometimes not able to extract the tick values or axes from the plots. Humans do not have such limitations, even at low resolutions.
In our human-annotated dataset, we found 265 proportional ink violations in the 4,834 annotated bar charts that could be processed (6.5% ± 0.4%). Per publication, this prevalence amounted to 5.5% (± 0.3%) of the publications with a bar chart. In our prediction pipeline, we predict that there are 479 proportional ink violations of the 20,666 bar charts that could be processed (2.3% (± 0.1%) per bar chart, 3.6% (± 0.2%) per publication with bar chart).
We found that the percentage of proportional ink violations is significantly larger in human-annotated than automated prediction bar charts (the p-value is less than 0.05 for both the ratio per publication with a bar chart and the ratio per chart by Fisher’s exact test). Even though our deep learning-based method has achieved high accuracy, this difference suggests that our method does not perform well on the metric of recall: we have a high number of false-negative cases. This is because some violations of the principle of proportional ink are hampered due to inaccurate text localization and recognition (see 4.2.1 section for more details). Also, given our conservative choice of threshold to decide whether a violation has occurred, we have this high false-positive rate in the performance test. We could change this threshold to bring our method closer to the human annotation data, but this inevitably increases the false positive rate. We discuss these issues in the Error Analysis section at the end of the result section.
3.2.2. Differences by seniority
We now investigate if there is a relationship between seniority and the ratio of graphical integrity issues (see Table 1 and Method and Materials section for more detail). To examine this research question, we conducted a correlation test between the likelihood of having graphical integrity issues and the h-index of the first author of publications and the last author of publications. (We acknowledge that the h-index of researchers might not fully reflect the seniority of researchers.) We found no statistically significant correlation for first author (r = 0.0091, N = 9,078, p-value = 0.38) or last author (r = 0.18, N = 10,509, p-value = 0.19). Similarly, we compared the h-index of the first authors with graphical integrity issues to those without and found no significant difference (Welch’s t-test, t(370.9) = 0.44, p-value = 0.66). We did a similar analysis comparing the last authors and found no statistical significance (Welch’s t-test, t(408.58) = 0.76, p-value = 0.45).
3.2.3. Differences across the impact of journals
High-impact journals are generally expected to publish rigorous research articles. Graphs as a part of research articles should also be accurate, but we do not know if the review of high-impact journals weeds out low-quality graphs. We tested the correlation between the rank of journals and journals’ likelihood of having at least one article with graphical integrity issues. To measure this likelihood, we first compute the maximum likelihood that a publication contains a graphical integrity issue. Then, we compute the median of this value per journal. For the journal rank, we use the rank based on the importance of papers in the network of publications from Microsoft Academic Graph [46]. For example, a highly cited paper in a prestigious journal has more importance than a preprint article. A Pearson correlational analysis revealed that there is no correlation that the higher the ranking of the journal the more likely the journal has graphical integrity issues (r = -0.01, N = 1357, p-value = 0.63, Fig 3). This suggests that there is no relationship between journal rank and the likelihood of having integrity issues.
3.2.4. Differences across research fields
Graphical representations of information vary substantially across fields. To understand these differences, we compared the likelihood of having graphical integrity issues across research fields. The research field of journals is selected by researchers from Microsoft Academic Graph, and each publication is classified into research fields by a multi-class classification method with publications’ content and connections in the network of publications. Different research fields have different commonness of graphical integrity issues (F(5, 10137) = 22.18, p-value < 0.001, Fig 4). A posthoc analysis (Tukey’s HSD) reveals that psychology contains proportional ink violations significantly more frequently than chemistry (p-value < 0.001) and biology (p-value < 0.001). Also, computer science contains violations of proportional ink principles significantly more frequently than chemistry (p-value < 0.001) and biology (p-value < 0.001).
3.2.5. Differences across countries
Previous research has shown that there is a significantly different prevalence of research integrity across countries [8]. Here, we examined the likelihood of having articles with graphical integrity issues for all their authors from the country of their affiliations. We found that there is a statistical difference between authors from countries of their affiliation (F(16, 11718) = 2.49, p-value < 0.001). We further aggregate the likelihood of having graphical integrity issues per article for each country by taking the mean of these likelihoods (see Fig 5).
3.2.6. Differences across years
It is unknown whether publishers and research communities have been aware of graphical integrity issues in the past and taken actions to be protected from them. We computed the correlation between the likelihood of having graphical integrity issues and publication year to examine if graphical integrity issues are decreasing or increasing over time (see Fig 6). We did not find a relationship between these quantities suggesting that graphical integrity issues have been consistent (r = -0.46, N = 10, p-value = 0.18).
3.2.7. Differences within author careers
We now examined whether having experienced graphical integrity issues in the past is predictive of graphical integrity issues in the future. First, we examine if the graphs produced by authors who had graphical integrity issues in the past are more likely to have graphical integrity issues than the graphs produced by authors who were not found with graphical integrity issues. We analyzed 301 bar charts in the future and found that only 5 of them had problems. Of the graphs produced by authors who did not likely have graphical integrity issues in the past, we analyzed 568 bar chart features and found that 6 of them had problems. According to a Fisher’s Exact Test, the difference between these two groups is not statistically significant (p-value = 0.53). Second, we then analyzed it at the author level. We found that 7 out of 218 authors with issues in the past had problems in the future; we found that 10 of 423 authors with potential issues had problems in the future. Again, Fisher’s Exact Test’s difference between these groups of authors is not statistically significant (p-value = 0.61). These results suggest that graphical integrity issues do not seem to “run in the family”: having issues in the past might not make an author have problems in the future.
3.2.8. Error analysis
Interpreting the differences between human annotation and our method improves the transparency of our method, but this task is difficult. To interpret this difference, we conducted a systematic evaluation of how the parameters on our model change the precision, recall, and predicted prevalence in the datasets. This is because we want to examine why our human annotators find 5.5% of bar charts having proportional ink violation, but our deep learning-based method finds 2.3% of bar charts having proportional ink violation (see 3.2.2 section for more details). We found this discrepancy between predicted prevalence and the human-annotated prevalence can be the result of the threshold in our classifier for classifying a bar chart into proportional ink violation (see section 4.2.3 for more details). We can reduce the discrepancy between predicted prevalence and human-annotated prevalence by lowering this threshold. Although this comes at the cost of increasing the false positive rate, we can approximate the human and predicted prevalence rate with a reasonable cost. To be concrete, we explore parameters that would approximate human prevalence. When we change the default threshold of classifying bar charts to graphical integrity issues from 0.5 to 0.23, we can improve our predicted prevalence rate of graphical integrity issues from 2.3% to 3.4%, without losing the F1 score (see S3 Fig). Thus, depending on the ultimate goals of the system, we do not necessarily want to approximate the human prevalence rate in this fashion. We believe that the choice of such a threshold should be in the hands of the users of the systems—ultimately, authors, journals, and universities.
We also performed a systematic analysis of the false positive and false negative classifications to determine whether we need more data, more sophisticated methods, or better feature engineering (see S4 and S5 Figs). The overall takeaway from this analysis is that our method has issues detecting bar charts with a partially hidden y-axis. This is because, for our method, we need highly accurate text localization and text recognition on the y-axis to perform well on the detection of these violations. However, methods for text localization are an area of active research [50].
3.3. Regression analysis
All the previous analyses were correlations between the variable of interests (e.g., author’s seniority) and the outcome (i.e., likelihood of proportional ink violations). However, we should also account for the partial correlation among the independent variables before relating it to the outcome. To this end, we run a multivariate regression analysis relating journal ranking, research field, citation count, author seniority (h-index), year of publication, number of affiliations, and country of affiliation to the likelihood of proportional ink violation. The model explains a statistically significant and weak proportion of variance (R squared = 0.02, F(42, 8707) = 4.51, p < .001, adjusted R squared = 0.02). The details of the regression are available in the S2 Table. Our results found that higher relationship between some fields (computer science: standardized coefficient 0.036 (standard error = 0.011), t(8707) = 3.41, p < 0.001 and psychology: standardized coefficient 0.11 (standard error = 0.0084), t(8707) = 9.75, p < 0.001) and lower relationship between some countries (e.g., Japan: standardized coefficient -0.037 (standard error = 0.0070), t(8707) = -3.33, p < 0.001; Korea: standardized coefficient -0.03 (standard error = 0.0073), t(8707) = -2.57, p = 0.010) to produce proportional ink violations. Interestingly our regression shows that journal rank has a significant positive effect on the likelihood of having proportional ink violation (standardized coefficient -0.048 (standard error < 0.001), t(8707) = -2.75, p < 0.001). This journal rank effect implies that higher-ranked journals are more likely to contain violations of proportional ink. This effect is there even after controlling for the size of the journal (i.e., number of articles published) and the interaction between ranking and size of the journal.
4. Materials and methods
4.1. Materials
4.1.1. PubMed open access and annotations
PubMed Open Access is a subset of PubMed Central, consisting of thousands of journals in biomedical research [47]. To train our machine learning-based graphical integrity evaluator, we annotated one subset of collected images mentioned in section 3.1.1 (after compound figure detection, subfigure separation, and chart classification) as graphical integrity issues and others (see Fig 7; also see S1 Table). We consider images as graphical integrity issues if they violate the principle of proportional ink. The specific rules are: a bar chart’s y-axis should start from zero, have one scale, and not be partially hidden [16,17]. We annotate images as others if they follow all three rules, do not have a full y-axis or x-axis, or are not bar charts.
4.1.2. Datasets for image processing techniques and features for statistical analysis
To automate graphical integrity evaluation, we need image processing techniques to obtain information from images. This study has six image processing steps: compound figure classification, subfigure separation, image classification, text localization, text recognition, and text role prediction (see the summary of training data and validation data in Table 1).
Data for testing within-author integrity issues. After labeling images, we analyzed the publications of these authors in the following three years after the first year we annotated. Based on the labeled data, we found 218 authors who produced graphs with integrity issues in their publications and analyzed 301 bar charts in their following publications. Also, there are 4780 authors who have not been found graphical integrity issues in our labeled data. Due to the unbalanced amount between the two groups, we randomly selected 3000 images from the following publications and analyzed 568 bar charts.
We now describe the features used in our statistical analyses:
Seniority of researchers. As the research experience of researchers grows, some researchers might be aware of graphical integrity and want to avoid graphical integrity issues. Thus, graphic integrity issues might be less common among senior researchers, for they have more education and knowledge. If this hypothesis holds, then it might imply that researchers can learn graphical integrity over time by themselves. If not, research communities might need more education about this topic to reduce such problematic practices. To examine this hypothesis, we collected researchers’ h-index to proxy their seniority for the analysis.
Impact of journals. High-impact journals usually have a strict standard for their publications. It is natural to expect high-impact journals to publish more rigorous research findings. However, whether high-impact journals value graphic integrity is unknown. Thorough research should make sure its visualizations convey accurate information to readers. If not, the journal publishers might need to include graphical integrity into their review standard. To examine this research question, we collected the rank of journals of articles in our sample.
Research fields. Different disciplines have their traditions, such as citation format and writing style. Similarly, other disciplines might have additional requirements for graphical integrity. However, this is a new and important research question because the publishers might want to enhance the review of graphical integrity for specific research fields. To address this question, we collected the field of study of journals of articles in our sample.
Countries. Education and culture vary from country to country. But we do not know if some countries have stricter graphic integrity. This research question can help researchers to know which country needs more protection of graphical integrity. Thus, we collected the country of the author affiliation of articles in our sample.
Year of publication. Science is changing with our research policy, education, and society over time. The time-series analysis of science helps researchers to understand the direction of new science. Similarly, graphical integrity over time can help the public to know if the awareness of graphical integrity changed in the past. To answer this research question, we collected the year of publications of articles in our sample.
4.2. Methods
Fig 8 explains the steps of our method, and Fig 9 shows an example of how our method analyze images.
4.2.1. Image preprocessing
The goal of image processing is to obtain information from images for feature engineering and classification of graphical integrity issues. Our image preprocessing includes the following steps:
Compound figure classification. Compound figures are standard in academic papers (e.g. [27]), but usually, graphical integrity issues are defined on non-compound figures. Therefore, we use a high-quality feature extractor based on a convolutional neural network (Resnet-152 v2, pre-trained on ImageNet) to classify figures into compound figures and non-compound figures.
Compound figure separation. In the case of a compound figure, we need to separate compound figures into subfigures because subfigures in compound figures might also have graphical integrity issues. To achieve subfigure separation, we trained a convolutional neural network (YOLO v4, pre-trained on MS COCO dataset) to localize subfigures in compound figures [51].
Image classification. This study focuses on bar charts because the violation of the principle of proportional ink is a widely-accepted graphical integrity issue for bar charts. Thus, we generated charts (bar charts, line charts, scatter charts, heatmap charts, box charts, area charts, radar plots, maps, pie charts, tables, Pareto charts, Venn diagrams, and violin charts) and collected diagnostic figures from ImageCLEF 2016 competition [34]. Then we use a high-quality feature extractor based on a convolutional neural network (Resnet-152 v2, pre-trained on ImageNet) to classify figures into categories.
Text localization. To predict if a bar chart has graphical integrity issues, we also need to know the texts on images for our classification. Therefore, we fine-tuned a convolutional neural network (YOLO v4, pre-trained on MS COCO dataset) to detect or localize texts on academic figures, preprocessed with Stroke Width Transformation [51,52].
Text recognition. More than the location of texts on graphs, we also need their content. Thus, we used Tesseract to recognize the content of texts based on the predicted locations of texts from text localization. We used one fine-tuned Tesseract model of English texts for this task.
Text role classification. Also, we need to know the role of texts on each figure for our analysis and thus use one open-source model to predict the role of texts on academic figures, based on the geometric information of texts on graphs [30].
Image preprocessing validation. To evaluate the performance of previous tasks, we used public datasets and our own generated datasets (when we could not find public datasets). The summary of our performance and benchmarks is summarized in Table 1.
4.2.2. Feature engineering
One way to build an automatic graphic integrity issues classifier is to apply classifiers from machine learning. A machine learning-based classifier needs features from data instead of the raw data. Thus, we conducted a feature engineering step from our image preprocessing (Table 2). Because the principle of proportional ink requires the ratios between the numeric value and the area of each bar to be the same in the same bar chart, the violation of this principle has the following symptoms for academic figures: bar charts with a non-zero y-axis and bar charts with a y-axis of multiple scales (also considering partially hidden y-axis). These two kinds of bar charts all violate the principle of proportional ink. Although bar charts with a log scale also violate the principle of proportional ink, we do not consider these bar charts as graphical integrity issues because log-scale is widely accepted in academic figures [54]. To examine if one bar chart has the above symptoms of graphical integrity issues, we transform the information from our image preprocessing into features for the final classification. The first two features are derived directly from the principle of proportional ink for bar charts. However, based on our experiments, these two features are not enough to detect graphical integrity issues because our image preprocessing might introduce errors to these two features. For example, the y-axis label might not be recognized by our text localization model. Thus, we created other features to consider the errors from our image preprocessing.
Table 2. Summary of features for proportional ink violation detection.
No | Feature Description | Reason |
---|---|---|
1 | The value of the lowest y-axis label on the y-axis (detected or inference from y-axis) | The lowest y-axis label should be zero |
2 | The increasing rate between each pair of y-axis labels | The scales of y-axis should be consistent across each pair of neighbor y-axis labels |
3 | If we need to inference the lowest text on the y-axis | If the lowest label on the y-axis is far from the x-axis, then we might ignore the actual lowest label on the y-axis |
4 | If the y-axis has a mix of integer and float number | Tesseract might not perform well with float number, and thus the increasing rate in the y-axis might not be accurate |
5 | The probability of being texts | We prefer texts with a higher probability of being texts |
6 | The OCR confidences of texts on the y-axis | We prefer predictions of the content of texts with a higher confidence |
7 | The probability of being bar charts | Our classifier only classifies bar charts. Thus, we prefer figures with a high probability of being bar charts |
4.2.3. Proportional ink principle evaluator
For an automatic graphical issue detector, rule-based methods are feasible when the features are highly correct. However, given the errors introduced from image processing, the texts and their roles on figures might not be correct. Thus, we took a probabilistic approach: the random forest model from machine learning. Our random forest model can predict if a figure is misleading or not based on the features of figures from our image processing techniques (see Fig 8).
4.2.4. Performance evaluation method
We used three metrics to evaluate our method: AUC, precision, and recall. Before we introduce how to interpret AUC, we first explain how we define true data points, false data points, positive data points, and negative data points, true positive data points, false-positive data points, true negative data points, and false negative data points. Here, true data points are images, which have an annotation as a proportional ink violation. False data points are images, which have an annotation as no proportional ink violation. Positive data points are images, which have a prediction from our method as a proportional ink violation. Negative data points are images, which have a prediction from our method as no proportional ink violation. True positive data points are images that have a prediction from our method as a proportional ink violation and also have an annotation as a proportional ink violation. False-positive data points have a prediction from our method as a proportional ink violation but have an annotation as no proportional ink violation. True negative data points are images, which have a prediction from our method as no proportional ink violation and annotation as no proportional ink violation. In addition, false-negative data points are images, which have a prediction from our method as no proportional ink violation but have an annotation of proportional ink violation.
AUC is the area under the ROC curve, which shows how true positive rate (the ratio between the number of true positive data points and the number of positive data points) changes with the false positive rate (the ratio between false-positive data points and the number of negative data points) with various thresholds as a curve [55].https://www.sciencedirect.com/science/article/pii/B9781558600362500473 When we have a high AUC for classifiers (close to 1), the true positive rate is high, and the false positive rate is low, and we have a much more accurate classifier than doing classification in a random manner. When we have an AUC of 0.5, the classifier is similar to doing classification in a random manner.
We additionally investigated the precision and recall of our method. Precision is computed as Eq (1), which is the ratio of the number of true positive data points to the number of all positive data points [56]. The recall is the ratio of the number of true positive data points to the number of all true data points (Eq (2)).
(1) |
(2) |
5. Discussion
In this article, we attempt to understand the extent of graphical integrity issues in scientific publications. We focused on violations of the proportional ink principle, one of the most basic visualization rules. We developed an automated method based on deep learning to detect these issues by learning from a large sample of human-annotated bar charts. Our approach achieved a high accuracy (AUC = 0.917). Our results suggest varying degrees of prevalence across research fields, journal rank, and affiliation countries. We now also discuss contributions of our method, implications of our research findings, limitations of our study, and future work.
5.1. Contributions
This study contributes an automated method to graphical integrity investigations and chart mining. Our method can detect the violation of the principle of proportional ink automatically and shows the potential to detect other graphical integrity issues if we follow the same steps. In previous studies, researchers have relied on human experts on graphical integrity investigations [21,22]. Although human experts can detect graphical integrity issues with very high accuracy, the speed of annotation can be slow. For a large sample of graphs, we still need automated techniques for graphical integrity investigations. Trying to fill this gap, we proposed our automated method to examine images at scale and to detect proportional ink violations with relatively high accuracy. Moreover, the components in this method can address compound figure classification and subfigure separation for other graphical integrity investigations, complementing previous research (e.g., [35,37]).
Our work shows the potential of using AI to help publishers check integrity issues. Given the overwhelming volume of new manuscript submissions, editors and reviewers face the challenges of maintaining the quality of publications. One potential way of reducing the workload of editors and reviewers is to apply AI to conduct some basic checks on manuscripts. With such considerations, one utility of this method is to analyze a large collection of graphs to detect the violations of the principle of proportional ink. Also, our method provides compound figure classification, subfigure separation, chart classification for publishers to investigate graphical integrity, image reuse, or image fabrications. Publishers could use these components for displaying and further processing purposes.
5.2. Implications
Our study found a varying prevalence of proportional ink violations across research fields and affiliation countries. Our result implies that graphical integrity issues might be more common in specific groups of researchers and various research communities. From the aspect of research fields and graphical integrity, our result shows that psychology and computer science might have more graphical integrity issues.
Our country and journal differences differ from image integrity studies (e.g., [8]). For example, we found that European countries tend to have a higher likelihood of graphical integrity violations. The difference in proportional ink violation across fields might be the result of the coverage of research regulation across research fields. For example, some fields have better research integrity regulations than others [57].
The high accuracy of our method implies that it should be viable to be applied before or after peer review. Journals and reviewers are already overwhelmed enough that checks like our methods are likely to be overlooked. More broadly, our results show that modern computational research integrity technology has the great potential to scan for potential issues at scale. If our analysis is integrated with other kinds of analysis for text and image plagiarism, the scientific publishing system could offer a good chance to improve the overall state of the quality of scientific communication.
5.3. Limitations and risks
Method limitation: Our proposed method makes mistakes in compound figure classification, subfigure separation, and bar chart classification (see Table 1). Thus, some bar charts might not be separated from compound figures correctly or might not be classified into bar charts at all. The method depends on accurate text location and parsing (see Table 1). Future work will incorporate more recent developments in deep learning technology for the detection, location, and parsing of text in images.
Our analysis focuses on proportional ink violations among bar charts only. However, bar charts are only one of many types of charts used in scientific publications (see Table 1). Future studies will adapt our method to detect these kinds of charts and detect adherence to graphical integrity principles.
Generalizability of research findings: Our method cannot detect all graphical integrity issues, and perhaps we are underestimating their prevalence. Our analysis might not be generalizable to non-open access journals. Non-open access journals tend to have a higher impact than open access journals [58] and can have different review processes. In the future, analyzing PDFs and figures of traditional journals would be an obvious next step.
Risks: Our method might be “gamed” by some users to pass our checks. It is unclear to what extent this happens with text plagiarism. We can think of two extreme potential responses to this. The first response is to simply let both good actors (e.g., research integrity investigators, editors) and bad actors (e.g., unscrupulous parties) use our method freely. We see this approach as similar to the one that search engines take: people can game search engine algorithms, but we will adapt. In this sense, we should constantly be paying attention to the bad actors in a game of cat and mouse. According to our analysis, the extend of the problem is worrying (5% have violations), but releasing the tool openly can help spur search on the topic. We believe this approach might be the right cost-benefit response to this kind of research integrity issue.
A second answer to the issue is to keep our method secret and only give access to it to publishers and reliable parties. This would make the process less widely available, but it would prevent scientists who genuinely want to improve their manuscripts to not being able to access this technology. We think this is a reasonable strategy, but the entity or researcher giving access to this method should be responsive enough to not stifle innovation in this space.
5.4. Future of our work
In the future, we hope to incorporate this analysis with other kinds of graph checks. For example, we can study image reuse and tamper at different units of figures (compound figures or subfigure) and in specific kinds of figures (Radiology figures or Microscope figures) [11,12]. Beyond graphical integrity issues, our results could be parts of other kinds of chart analysis, such as the aspect of distortion [22] and clearness [59]. For example, our method can be a part of studies of the quality of visual communications from the perspective of human perception and learning [60].
Our future work would benefit from higher-quality images. According to our error analysis, a large proportion of our method’s mistakes is due to low-quality images. PubMed Open Access subset makes the article’s figures available as separate images, but they are sometimes of lower quality compared to those inside the PDFs. Therefore, future work will attempt to extract images directly from PDFs. Lastly, improvements in these image analysis steps are an active area of research, and therefore should be improved in the future.
In sum, research communities could enhance their education about graphical integrity to all researchers to reduce the burden of publishers and reviewers on graphical integrity. These ideas are already part of research integrity initiatives around the world, including COPE practices and the Hong Kong Principles [61,62].
6. Conclusion
This study presents a deep learning-based method to detect graphical integrity issues and research findings of graphical integrity issues in science. This method enables researchers to detect proportional ink violation at scale and offers the potential to address other graphical integrity issues. With this method, we found that graphical integrity issues do appear among scientific articles. Even though the fraction of bar charts having proportional ink violations is around 5%, one of these violations might cause some readers to misunderstand the research finding because open access journals are easy to access to a wide range of readers. These misunderstandings might harm readers extensively when the research finding is related to safety or health. With these potential consequences, we suggest publishers and research communities consider taking actions to protect readers from graphical integrity issues in publications. In the future, we hope to incorporate this analysis with other kinds of graph checks and improve the accuracy of the components in our method.
Supporting information
Data Availability
All relevant data and code are available in a Github repository (https://github.com/sciosci/graph_check) and Zenodo dataset (https://zenodo.org/record/5500684#.YTvYclspBhF).
Funding Statement
HZ, TYH, DEA were partially funded by the ORI HHS grants ORIIR180041 and ORIIIR190049. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1.Office of Research Integrity. Definition of Research Misconduct | ORI—The Office of Research Integrity. 2021. [cited 12 May 2021]. Available: https://ori.hhs.gov/definition-research-misconduct [Google Scholar]
- 2.Gopalakrishna G, Riet G ter, Cruyff MJLF, Vink G, Stoop I, Wicherts J, et al. Prevalence of questionable research practices, research misconduct and their potential explanatory factors: a survey among academic researchers in The Netherlands. MetaArXiv; 2021. doi: 10.31222/osf.io/vk9yt [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Titus SL, Wells JA, Rhoades LJ. Repairing research integrity. Nature. 2008;453: 980–982. doi: 10.1038/453980a [DOI] [PubMed] [Google Scholar]
- 4.Powell K. Publishing: Foreign tongues. Nature. 2012;487: 129–131. doi: 10.1038/nj7405-129a [DOI] [PubMed] [Google Scholar]
- 5.Weber-Wulff D. False Feathers: A Perspective on Academic Plagiarism. Berlin Heidelberg: Springer-Verlag; 2014. doi: 10.1007/978-3-642-39961-9 [DOI] [Google Scholar]
- 6.Foltýnek T, Dlabolová D, Anohina-Naumeca A, Razı S, Kravjar J, Kamzola L, et al. Testing of support tools for plagiarism detection. Int J Educ Technol High Educ. 2020;17: 46. doi: 10.1186/s41239-020-00192-4 [DOI] [Google Scholar]
- 7.Beck TS. Shaping Images. Shaping Images. De Gruyter Saur; 2016. Available: https://www.degruyter.com/document/doi/ doi: 10.1515/9783110477092/html [DOI] [Google Scholar]
- 8.Bik EM, Casadevall A, Fang FC. The Prevalence of Inappropriate Image Duplication in Biomedical Research Publications. mBio. 2016;7. doi: 10.1128/mBio.00809-16 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Cromey DW. Avoiding Twisted Pixels: Ethical Guidelines for the Appropriate Use and Manipulation of Scientific Digital Images. Sci Eng Ethics. 2010;16: 639–667. doi: 10.1007/s11948-010-9201-y [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Jordan SR. Research integrity, image manipulation, and anonymizing photographs in visual social science research. Int J Soc Res Methodol. 2014;17: 441–454. doi: 10.1080/13645579.2012.759333 [DOI] [Google Scholar]
- 11.Acuna D, Brooks P, Kording K. Bioscience-scale automated detection of figure element reuse. 2018. [Google Scholar]
- 12.Xiang Z, Acuna DE. Scientific Image Tampering Detection Based On Noise Inconsistencies: A Method And Datasets. ArXiv200107799 Cs Eess. 2020. [cited 4 Jun 2021]. Available: http://arxiv.org/abs/2001.07799 [Google Scholar]
- 13.Yang BW, Vargas Restrepo C, Stanley ML, Marsh EJ. Truncating Bar Graphs Persistently Misleads Viewers. J Appl Res Mem Cogn. 2021. [cited 28 Apr 2021]. doi: 10.1016/j.jarmac.2020.10.002 [DOI] [Google Scholar]
- 14.Larkin AA. Misleading Graphics: Can Decision Makers Be Affected By Their Use. 1990. [Google Scholar]
- 15.Raschke RL, Steinbart PJ. Mitigating the Effects of Misleading Graphs on Decisions by Educating Users about the Principles of Graph Design. J Inf Syst. 2008;22: 23–52. doi: 10.2308/jis.2008.22.2.23 [DOI] [Google Scholar]
- 16.Tufte ER. The visual display of quantitative information. Graphics press Cheshire, CT; 2001. [Google Scholar]
- 17.Bergstrom CT, West JD. Calling Bullshit: The Art of Skepticism in a Data-Driven World. Illustrated Edition. New York: Random House; 2020. [Google Scholar]
- 18.Bertin J. Semiology of Graphics: Diagrams, Networks, Maps. ESRI Press; 2011. [Google Scholar]
- 19.Ware C. Information Visualization: Perception for Design: Second Edition. Information Visualization: Perception for Design: Second Edition. 2004. [Google Scholar]
- 20.Beattie V, Jones MJ. The use and abuse of graphs in annual reports: a theoretical framework and an empirical study. Account Bus Res. 1992;22: 291–303. [Google Scholar]
- 21.Courtis JK. Corporate annual report graphical communication in Hong Kong: effective or misleading? J Bus Commun 1973. 1997;34: 269–284. [Google Scholar]
- 22.Beattie V, Jones MJ. Measurement distortion of graphs in corporate reports: an experimental study. Account Audit Account J. 2002. [Google Scholar]
- 23.West JD, Bergstrom CT. Misinformation in and about science. Proc Natl Acad Sci. 2021;118. doi: 10.1073/pnas.1912444117 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Gonon F, Bezard E, Boraud T. Misrepresentation of Neuroscience Data Might Give Rise to Misleading Conclusions in the Media: The Case of Attention Deficit Hyperactivity Disorder. PLOS ONE. 2011;6: e14618. doi: 10.1371/journal.pone.0014618 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Sumner P, Vivian-Griffiths S, Boivin J, Williams A, Venetis CA, Davies A, et al. The association between exaggeration in health related science news and academic press releases: retrospective observational study. BMJ. 2014;349: g7015. doi: 10.1136/bmj.g7015 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Davila K, Setlur S, Doermann D, Kota BU, Govindaraju V. Chart Mining: A Survey of Methods for Automated Chart Analysis. IEEE Trans Pattern Anal Mach Intell. 2021;43: 3799–3819. doi: 10.1109/TPAMI.2020.2992028 [DOI] [PubMed] [Google Scholar]
- 27.Lee P, West JD, Howe B. Viziometrics: Analyzing Visual Information in the Scientific Literature. IEEE Trans Big Data. 2018;4: 117–129. doi: 10.1109/TBDATA.2017.2689038 [DOI] [Google Scholar]
- 28.Cliche M, Rosenberg D, Madeka D, Yee C. Scatteract: Automated Extraction of Data from Scatter Plots. In: Ceci M, Hollmén J, Todorovski L, Vens C, Džeroski S, editors. Machine Learning and Knowledge Discovery in Databases. Cham: Springer International Publishing; 2017. pp. 135–150. doi: 10.1007/978-3-319-71249-9_9 [DOI] [Google Scholar]
- 29.Luo J, Li Z, Wang J, Lin C-Y. ChartOCR: Data Extraction From Charts Images via a Deep Hybrid Framework. 2021. pp. 1917–1925. Available: https://openaccess.thecvf.com/content/WACV2021/html/Luo_ChartOCR_Data_Extraction_From_Charts_Images_via_a_Deep_Hybrid_WACV_2021_paper.html [Google Scholar]
- 30.Poco J, Heer J. Reverse-engineering visualizations: Recovering visual encodings from chart images. Computer Graphics Forum. Wiley Online Library; 2017. pp. 353–363. [Google Scholar]
- 31.Clark C, Divvala S. PDFFigures 2.0: Mining figures from research papers. 2016 IEEE/ACM Joint Conference on Digital Libraries (JCDL). 2016. pp. 143–152. [Google Scholar]
- 32.Li P, Jiang X, Shatkay H. Figure and caption extraction from biomedical documents. Bioinformatics. 2019;35: 4381–4388. doi: 10.1093/bioinformatics/btz228 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Siegel N, Lourie N, Power R, Ammar W. Extracting Scientific Figures with Distantly Supervised Neural Networks. Proceedings of the 18th ACM/IEEE on Joint Conference on Digital Libraries. New York, NY, USA: Association for Computing Machinery; 2018. pp. 223–232. doi: 10.1145/3197026.3197040 [DOI] [Google Scholar]
- 34.García Seco de Herrera A, Schaer R, Bromuri S, Müller H. Overview of the ImageCLEF 2016 medical task. Working Notes of CLEF 2016 (Cross Language Evaluation Forum). 2016. [Google Scholar]
- 35.Taschwer M, Marques O. Automatic separation of compound figures in scientific articles. Multimed Tools Appl. 2018;77: 519–548. doi: 10.1007/s11042-016-4237-x [DOI] [Google Scholar]
- 36.Taschwer M, Marques O. Compound Figure Separation Combining Edge and Band Separator Detection. In: Tian Q, Sebe N, Qi G-J, Huet B, Hong R, Liu X, editors. MultiMedia Modeling. Cham: Springer International Publishing; 2016. pp. 162–173. doi: 10.1007/978-3-319-27671-7_14 [DOI] [Google Scholar]
- 37.Lee SL, Zare MR. Biomedical compound figure detection using deep learning and fusion techniques. IET Image Process. 2018;12: 1031–1037. doi: 10.1049/iet-ipr.2017.0800 [DOI] [Google Scholar]
- 38.Shneiderman B. The Eyes Have It: A Task by Data Type Taxonomy for Information Visualizations. In: Bederson BB, Shneiderman B, editors. The Craft of Information Visualization. San Francisco: Morgan Kaufmann; 2003. pp. 364–371. doi: 10.1016/j.crvi.2003.09.035 [DOI] [Google Scholar]
- 39.Prasad VSN, Siddiquie B, Golbeck J, Davis LS. Classifying Computer Generated Charts. 2007 International Workshop on Content-Based Multimedia Indexing. 2007. pp. 85–92. doi: 10.1109/CBMI.2007.385396 [DOI] [Google Scholar]
- 40.Chagas P, Akiyama R, Meiguins A, Santos C, Saraiva F, Meiguins B, et al. Evaluation of Convolutional Neural Network Architectures for Chart Image Classification. 2018 International Joint Conference on Neural Networks (IJCNN). 2018. pp. 1–8. doi: 10.1109/IJCNN.2018.8489315 [DOI] [Google Scholar]
- 41.Al-Zaidy RA, Giles CL. A Machine Learning Approach for Semantic Structuring of Scientific Charts in Scholarly Documents. 2017. Available: https://openreview.net/forum?id=r1bL5eWOWr [Google Scholar]
- 42.Choudhury SR, Wang S, Giles CLee. Scalable algorithms for scholarly figure mining and semantics. Proceedings of the International Workshop on Semantic Big Data. New York, NY, USA: Association for Computing Machinery; 2016. pp. 1–6. doi: 10.1145/2928294.2928305 [DOI] [Google Scholar]
- 43.Vassilieva N, Fomina Y. Text detection in chart images. Pattern Recognit Image Anal. 2013;23: 139–144. doi: 10.1134/S1054661813010112 [DOI] [Google Scholar]
- 44.Nair RR, Sankaran N, Nwogu I, Govindaraju V. Automated analysis of line plots in documents. 2015 13th International Conference on Document Analysis and Recognition (ICDAR). 2015. pp. 796–800. doi: 10.1109/ICDAR.2015.7333871 [DOI] [Google Scholar]
- 45.Lo K, Wang LL, Neumann M, Kinney R, Weld DS. S2ORC: The Semantic Scholar Open Research Corpus. ArXiv191102782 Cs. 2020. [cited 2 Jun 2021]. Available: http://arxiv.org/abs/1911.02782 [Google Scholar]
- 46.Sinha A, Shen Z, Song Y, Ma H, Eide D, Hsu B-J (Paul), et al. An Overview of Microsoft Academic Service (MAS) and Applications. Proceedings of the 24th International Conference on World Wide Web. New York, NY, USA: Association for Computing Machinery; 2015. pp. 243–246. doi: 10.1145/2740908.2742839 [DOI] [Google Scholar]
- 47.PMC Overview. 2021 [cited 30 Apr 2021]. Available: https://www.ncbi.nlm.nih.gov/pmc/about/intro/
- 48.Cohen J. A Coefficient of Agreement for Nominal Scales. Educ Psychol Meas. 1960;20: 37–46. doi: 10.1177/001316446002000104 [DOI] [Google Scholar]
- 49.Landis JR, Koch GG. The Measurement of Observer Agreement for Categorical Data. Biometrics. 1977;33: 159–174. doi: 10.2307/2529310 [DOI] [PubMed] [Google Scholar]
- 50.Chen X, Jin L, Zhu Y, Luo C, Wang T. Text Recognition in the Wild: A Survey. ACM Comput Surv. 2021;54: 1–35. doi: 10.1145/3440756 [DOI] [Google Scholar]
- 51.Bochkovskiy A, Wang C-Y, Liao H-YM. YOLOv4: Optimal Speed and Accuracy of Object Detection. ArXiv200410934 Cs Eess. 2020. [cited 30 Nov 2020]. Available: http://arxiv.org/abs/2004.10934 [Google Scholar]
- 52.Epshtein B, Ofek E, Wexler Y. Detecting text in natural scenes with stroke width transform. 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. 2010. pp. 2963–2970. doi: 10.1109/CVPR.2010.5540041 [DOI] [Google Scholar]
- 53.Savva M, Kong N, Chhajta A, Fei-Fei L, Agrawala M, Heer J. ReVision: automated classification, analysis and redesign of chart images. Proceedings of the 24th annual ACM symposium on User interface software and technology—UIST ‘11. Santa Barbara, California, USA: ACM Press; 2011. p. 393. doi: 10.1145/2047196.2047247 [DOI] [Google Scholar]
- 54.Bergstrom C, West J. Tools—Proportional Ink and Logarithmic Scales. 2021. [cited 9 Jun 2021]. Available: https://www.callingbullshit.org/tools/logarithmic_scales.html [Google Scholar]
- 55.Spackman KA. Signal Detection Theory: Valuable Tools For Evaluating Inductive Learning. In: Segre AM, editor. Proceedings of the Sixth International Workshop on Machine Learning. San Francisco (CA): Morgan Kaufmann; 1989. pp. 160–163. doi: [DOI] [Google Scholar]
- 56.Schütze H, Manning CD, Raghavan P. Introduction to information retrieval. Cambridge University Press Cambridge; 2008. [Google Scholar]
- 57.Haven TL, Tijdink JK, Martinson BC, Bouter LM. Perceptions of research integrity climate differ between academic ranks and disciplinary fields: Results from a survey among academic researchers in Amsterdam. PLOS ONE. 2019;14: e0210599. doi: 10.1371/journal.pone.0210599 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Swan A, Brown S. Authors and open access publishing. Learn Publ. 2004;17: 219–224. doi: 10.1087/095315104323159649 [DOI] [Google Scholar]
- 59.Jambor H, Antonietti A, Alicea B, Audisio TL, Auer S, Bhardwaj V, et al. Creating clear and informative image-based figures for scientific publications. PLOS Biol. 2021;19: e3001161. doi: 10.1371/journal.pbio.3001161 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Shukla S, Samal A. Recognition and quality assessment of data charts in mixed-mode documents. Int J Doc Anal Recognit IJDAR. 2008;11: 111. doi: 10.1007/s10032-008-0065-5 [DOI] [Google Scholar]
- 61.Committee on Publication Ethics. Core practices | COPE: Committee on Publication Ethics. 2021 [cited 9 Jun 2021]. Available: https://publicationethics.org/core-practices
- 62.Moher D, Bouter L, Kleinert S, Glasziou P, Sham MH, Barbour V, et al. The Hong Kong Principles for assessing researchers: Fostering research integrity. PLOS Biol. 2020;18: e3000737. doi: 10.1371/journal.pbio.3000737 [DOI] [PMC free article] [PubMed] [Google Scholar]