Skip to main content
PLOS Biology logoLink to PLOS Biology
. 2021 Mar 31;19(3):e3001161. doi: 10.1371/journal.pbio.3001161

Creating clear and informative image-based figures for scientific publications

Helena Jambor 1,, Alberto Antonietti 2,3,, Bradly Alicea 4, Tracy L Audisio 5, Susann Auer 6, Vivek Bhardwaj 7,8, Steven J Burgess 9, Iuliia Ferling 10, Małgorzata Anna Gazda 11,12, Luke H Hoeppner 13,14, Vinodh Ilangovan 15, Hung Lo 16,17, Mischa Olson 18, Salem Yousef Mohamed 19, Sarvenaz Sarabipour 20, Aalok Varma 21, Kaivalya Walavalkar 21, Erin M Wissink 22, Tracey L Weissgerber 23,*
Editor: Jason R Swedlow24
PMCID: PMC8041175  PMID: 33788834

Abstract

Scientists routinely use images to display data. Readers often examine figures first; therefore, it is important that figures are accessible to a broad audience. Many resources discuss fraudulent image manipulation and technical specifications for image acquisition; however, data on the legibility and interpretability of images are scarce. We systematically examined these factors in non-blot images published in the top 15 journals in 3 fields; plant sciences, cell biology, and physiology (n = 580 papers). Common problems included missing scale bars, misplaced or poorly marked insets, images or labels that were not accessible to colorblind readers, and insufficient explanations of colors, labels, annotations, or the species and tissue or object depicted in the image. Papers that met all good practice criteria examined for all image-based figures were uncommon (physiology 16%, cell biology 12%, plant sciences 2%). We present detailed descriptions and visual examples to help scientists avoid common pitfalls when publishing images. Our recommendations address image magnification, scale information, insets, annotation, and color and may encourage discussion about quality standards for bioimage publishing.

Introduction

Images are often used to share scientific data, providing the visual evidence needed to turn concepts and hypotheses into observable findings. An analysis of 8 million images from more than 650,000 papers deposited in PubMed Central revealed that 22.7% of figures were “photographs,” a category that included microscope images, diagnostic images, radiology images, and fluorescence images [1]. Cell biology was one of the most visually intensive fields, with publications containing an average of approximately 0.8 photographs per page [1]. Plant sciences papers included approximately 0.5 photographs per page [1].

While there are many resources on fraudulent image manipulation and technical requirements for image acquisition and publishing [24], data examining the quality of reporting and ease of interpretation for image-based figures are scarce. Recent evidence suggests that important methodological details about image acquisition are often missing [5]. Researchers generally receive little or no training in designing figures; yet many scientists and editors report that figures and tables are one of the first elements that they examine when reading a paper [6,7]. When scientists and journals share papers on social media, posts often include figures to attract interest. The PubMed search engine caters to scientists’ desire to see the data by presenting thumbnail images of all figures in the paper just below the abstract [8]. Readers can click on each image to examine the figure, without ever accessing the paper or seeing the introduction or methods. EMBO’s Source Data tool (RRID:SCR_015018) allows scientists and publishers to share or explore figures, as well as the underlying data, in a findable and machine readable fashion [9].

Image-based figures in publications are generally intended for a wide audience. This may include scientists in the same or related fields, editors, patients, educators, and grants officers. General recommendations emphasize that authors should design figures for their audience rather than themselves and that figures should be self-explanatory [7]. Despite this, figures in papers outside one’s immediate area of expertise are often difficult to interpret, marking a missed opportunity to make the research accessible to a wide audience. Stringent quality standards would also make image data more reproducible. A recent study of fMRI image data, for example, revealed that incomplete documentation and presentation of brain images led to nonreproducible results [10,11].

Here, we examined the quality of reporting and accessibility of image-based figures among papers published in top journals in plant sciences, cell biology, and physiology. Factors assessed include the use of scale bars, explanations of symbols and labels, clear and accurate inset markings, and transparent reporting of the object or species and tissue shown in the figure. We also examined whether images and labels were accessible to readers with the most common form of color blindness [12]. Based on our results, we provide targeted recommendations about how scientists can create informative image-based figures that are accessible to a broad audience. These recommendations may also be used to establish quality standards for images deposited in emerging image data repositories.

Results

Using a science of science approach to investigate current practices

This study was conducted as part of a participant-guided learn-by-doing course, in which eLife Community Ambassadors from around the world worked together to design, complete, and publish a meta-research study [13]. Participants in the 2018 Ambassadors program designed the study, developed screening and abstraction protocols, and screened papers to identify eligible articles (HJ, BA, SJB, VB, LHH, VI, SS, EMW). Participants in the 2019 Ambassadors program refined the data abstraction protocol, completed data abstraction and analysis, and prepared the figures and manuscript (AA, SA, TLA, IF, MAG, HL, SYM, MO, AV, KW, HJ, TLW).

To investigate current practices in image publishing, we selected 3 diverse fields of biology to increase generalizability. For each field, we examined papers published in April 2018 in the top 15 journals, which publish original research (S1S3 Tables). All full-length original research articles that contained at least one photograph, microscope image, electron microscope image, or clinical image (MRI, ultrasound, X-ray, etc.) were included in the analysis (S1 Fig). Blots and computer-generated images were excluded, as some of the criteria assessed do not apply to these types of images. Two independent reviewers assessed each paper, according to the detailed data abstraction protocol (see methods and information deposited on the Open Science Framework (OSF) (RRID:SCR_017419) at https://doi.org/10.17605/OSF.IO/B5296) [14]. The repository also includes data, code, and figures.

Image analysis

First, we confirmed that images are common in the 3 biology subfields analyzed. More than half of the original research articles in the sample contained images (plant science: 68%, cell biology: 72%, physiology: 55%). Among the 580 papers that included images, microscope images were very common in all 3 fields (61% to 88%, Fig 1A). Photographs were very common in plant sciences (86%), but less widespread in cell biology (38%) and physiology (17%). Electron microscope images were less common in all 3 fields (11% to 19%). Clinical images, such as X-rays, MRI or ultrasound, and other types of images were rare (2% to 9%).

Fig 1. Image types and reporting of scale information and insets.

Fig 1

(A) Microscope images and photographs were common, whereas other types of images were used less frequently. (B) Complete scale information was missing in more than half of the papers examined. Partial scale information indicates that scale information was presented in some figures, but not others, or that the authors reported magnification rather than including scale bars on the image. (C) Problems with labeling and describing insets are common. Totals may not be exactly 100% due to rounding.

Scale information is essential to interpret biological images. Approximately half of papers in physiology (49%) and cell biology (55%) and 28% of plant science papers provided scale bars with dimensions (in the figure or legend) for all images in the paper (Fig 1B, S4 Table). Approximately one-third of papers in each field contained incomplete scale information, such as reporting magnification or presenting scale information for a subset of images. Twenty-four percent of physiology papers, 10% of cell biology papers, and 29% of plant sciences papers contained no scale information on any image.

Some publications use insets to show the same image at 2 different scales (cell biology papers: 40%, physiology: 17%, plant sciences: 12%). In this case, the authors should indicate the position of the high-magnification inset in the low-magnification image. The majority of papers in all 3 fields clearly and accurately marked the location of all insets (53% to 70%; Fig 1C, left panel); however, one-fifth of papers appeared to have marked the location of at least one inset incorrectly (17% to 22%). Clearly visible inset markings were missing for some or all insets in 13% to 28% of papers (Fig 1C, left panel). Approximately half of papers (43% to 53%; Fig 1C, right panel) provided legend explanations or markings on the figure to clearly show that an inset was used, whereas this information was missing for some or all insets in the remaining papers.

Many images contain information in color. We sought to determine whether color images were accessible to readers with deuteranopia, the most common form of color blindness, by using the color blindness simulator Color Oracle (https://colororacle.org/, RRID: SCR_018400). We evaluated only images in which the authors selected the image colors (e.g., fluorescence microscopy). Papers without any colorblind accessible figures were uncommon (3% to 6%); however, 45% of cell biology papers and 21% to 24% of physiology and plant science papers contained some images that were inaccessible to readers with deuteranopia (Fig 2A). Seventeen percent to 34% of papers contained color annotations that were not visible to someone with deuteranopia.

Fig 2. Use of color and annotations in image-based figures.

Fig 2

(A) While many authors are using colors and labels that are visible to colorblind readers, the data show that improvement is needed. (B) Most papers explain colors in image-based figures; however, explanations are less common for the species and tissue or object shown, and labels and annotations. Totals may not be exactly 100% due to rounding.

Figure legends and, less often, titles typically provide essential information needed to interpret an image. This text provides information on the specimen and details of the image, while also explaining labels and annotations used to highlight structures or colors. Fifty-seven percent of physiology papers, 48% of cell biology papers, and 20% of plant papers described the species and tissue or object shown completely. Five percent to 17% of papers did not provide any such information (Fig 2B). Approximately half of the papers (47% to 58%; Fig 1C, right panel) also failed or partially failed to adequately explain that insets were used. Annotations of structures were better explained. Two-thirds of papers across all 3 fields clearly stated the meaning of all image labels, while 18% to 24% of papers provided partial explanations. Most papers (73% to 83%) completely explained the image colors by stating what substance each color represented or naming the dyes or staining technique used.

Finally, we examined the number of papers that used optimal image presentation practices for all criteria assessed in the study. Twenty-eight (16%) physiology papers, 19 (12%) cell biology papers, and 6 (2%) plant sciences papers met all criteria for all image-based figures in the paper. In plant sciences and physiology, the most common problems were with scale bars, insets, and specifying in the legend the species and tissue or object shown. In cell biology, the most common problems were with insets, colorblind accessibility, and specifying in the legend the species and tissue or object shown.

Designing image-based figures: How can we improve?

Our results obtained by examining 580 papers from 3 fields provide us with unique insights into the quality of reporting and the accessibility of image-based figures. Our quantitative description of standard practices in image publication highlights opportunities to improve transparency and accessibility to readers from different backgrounds. We have therefore outlined specific actions that scientists can take when creating images, designing multipanel figures, annotating figures, and preparing figure legends.

Throughout the paper, we provide visual examples to illustrate each stage of the figure preparation process. Other elements are often omitted to focus readers’ attention on the step illustrated in the figure. For example, a figure that highlights best practices for displaying scale bars may not include annotations designed to explain key features of the image. When preparing image-based figures in scientific publications, readers should address all relevant steps in each figure. All steps described below (image cropping and insets, adding scale bars and annotation, choosing color channel appearances, figure panel layout) can be implemented with standard image processing software such as FIJI [15] (RRID:SCR_002285) and ImageJ2 [16] (RRID:SCR_003070), which are open source, free programs for bioimage analysis. A quick guide on how to do basic image processing for publications with FIJI is available in a recent cheat sheet publication [17], and a discussion forum and wiki are available for FIJI and ImageJ (https://imagej.net/).

1. Choose a scale or magnification that fits your research question

Scientists should select an image scale or magnification that allows readers to clearly see features needed to answer the research question. Fig 3A [18] shows Drosophila melanogaster at 3 different microscopic scales. The first focuses on the ovary tissue and might be used to illustrate the appearance of the tissue or show stages of development. The second focuses on a group of cells. In this example, the “egg chamber” cells show different nucleic acid distributions. The third example focuses on subcellular details in one cell, for example, to show finer detail of RNA granules or organelle shape.

Fig 3. Selecting magnification and using insets.

Fig 3

(A) Magnification and display detail of images should permit readers to see features related to the main message that the image is intended to convey. This may be the organism, tissue, cell, or a subcellular level. Microscope images [18] show D. melanogaster ovary (A1), ovarian egg chamber cells (A2), and a detail in egg chamber cell nuclei (A3). (B) Insets or zoomed-in areas are useful when 2 different scales are needed to allow readers to see essential features. It is critical to indicate the origin of the inset in the full-scale image. Poor and clear examples are shown. Example images were created based on problems observed by reviewers. Images show B1, B2, B3, B5: Protostelium aurantium amoeba fed on germlings of Aspergillus fumigatus D141-GFP (green) fungal hyphae, dead fungal material stained with propidium iodide (red), and acidic compartments of amoeba marked with LysoTracker Blue DND-22 dye (blue); B4: Lendrum-stained human lung tissue (Haraszti, Public Health Image Library); B6: fossilized Orobates pabsti [19].

When both low and high magnifications are necessary for one image, insets are used to show a small portion of the image at higher magnification (Fig 3B, [19]). The inset location must be accurately marked in the low-magnification image. We observed that the inset position in the low-magnification image was missing, unclear, or incorrectly placed in approximately one-third of papers. Inset positions should be clearly marked by lines or regions of interest in a high-contrast color, usually black or white. Insets may also be explained in the figure legend. Care must be taken when preparing figures outside vector graphics suits, as insert positions may move during file saving or export.

2. Include a clearly labeled scale bar

Scale information allows audiences to quickly understand the size of features shown in images. This is especially important for microscopic images where we have no intuitive understanding of scale. Scale information for photographs should be considered when capturing images as rulers are often placed into the frame. Our analysis revealed that 10% to 29% of papers screened failed to provide any scale information and that another third only provided incomplete scale information (Fig 1B). Scientists should consider the following points when displaying scale bars:

  • Every image type needs a scale bar: Authors usually add scale bars to microscope images but often leave them out in photos and clinical images, possibly because these depict familiar objects such a human or plant. Missing scale bars, however, adversely affect reproducibility. A size difference of 20% in between a published study and the reader’s lab animals, for example, could impact study results by leading to an important difference in phenotype. Providing scale bars allows scientists to detect such discrepancies and may affect their interpretation of published work. Scale bars may not be a standard feature of image acquisition and processing software for clinical images. Authors may need to contact device manufacturers to determine the image size and add height and width labels.

  • Scale bars and labels should be clearly visible: Short scale bars, thin scale bars, and scale bars in colors that are similar to the image color can easily be overlooked (Fig 4). In multicolor images, it can be difficult to find a color that makes the scale bar stand out. Authors can solve this problem by placing the scale bar outside the image or onto a box with a more suitable background color.

  • Annotate scale bar dimensions on the image: Stating the dimensions along with the scale bar allows readers to interpret the image more quickly. Despite this, dimensions were typically stated in the legend instead (Fig 1B), possibly a legacy of printing processes that discouraged text in images. Dimensions should be in high resolution and large enough to be legible. In our set, we came across small and/or low-resolution annotations that were illegible in electronic versions of the paper, even after zooming in. Scale bars that are visible on larger figures produced by authors may be difficult to read when the size of the figure is reduced to fit onto a journal page. Authors should carefully check page proofs to ensure that scale bars and dimensions are clearly visible.

Fig 4. Using scale bars to annotate image size.

Fig 4

Scale bars provide essential information about the size of objects, which orients readers and helps them to bridge the gap between the image and reality. Scales may be indicated by a known size indicator such as a human next to a tree, a coin next to a rock, or a tape measure next to a smaller structure. In microscope images, a bar of known length is included. Example images were created based on problems observed by reviewers. Poor scale bar examples (1 to 6), clear scale bar examples (7 to 12). Images 1, 4, 7: Microscope images of D. melanogaster nurse cell nuclei [18]; 2: Microscope image of Dictyostelium discoideum expressing Vps32-GFP (Vps32-green fluorescent protein shows broad signal in cells) and stained with dextran (spotted signal) after infection with conidia of Aspergillus fumigatus; 3, 5, 8, 10: Electron microscope image of mouse pancreatic beta-islet cells (Andreas Müller); 6, 11: Microscope image of Lendrum-stained human lung tissue (Haraszti, Public Health Image Library); 9: Photo of Arabidopsis thaliana; 12: Photograph of fossilized Orobates pabsti [19].

3. Use color wisely in images

Colors in images are used to display the natural appearance of an object or to visualize features with dyes and stains. In the scientific context, adapting colors is possible and may enhance readers’ understanding, while poor color schemes may distract or mislead. Images showing the natural appearance of a subject, specimen, or staining technique (e.g., images showing plant size and appearance, or histopathology images of fat tissue from mice on different diets) are generally presented in color (Fig 5). Images showing electron microscope images are captured in black and white (“grayscale”) by default and may be kept in grayscale to leverage the good contrast resulting from a full luminescence spectrum.

Fig 5. Image types and their accessibility in colorblind render and grayscale mode.

Fig 5

Shown are examples of the types of images that one might find in manuscripts in the biological or biomedical sciences: photograph, fluorescent microscope images with 1 to 3 color hues/LUT, electron microscope images. The relative visibility is assessed in a colorblind rendering for deuteranopia, and in grayscale. Grayscale images offer the most contrast (1-color microscope image) but cannot show several structures in parallel (multicolor images, color photographs). Color combinations that are not colorblind accessible were used in rows 3 and 4 to illustrate the importance of colorblind simulation tests. Scale bars are not included in this figure, as they could not be added in a nondistracting way that would not detract from the overall message of the figure. Images show: Row 1: Darth Vader being attacked, Row 2: D. melanogaster salivary glands [18], Row 3: D. melanogaster egg chambers [18], Row 4: D. melanogaster nurse cell nuclei [18], and Row 5: mouse pancreatic beta-islet cells. LUT, lookup table.

In some instances, scientists can choose whether to show grayscale or color images. Assigning colors may be optional, even though it is the default setting in imaging programs. When showing only one color channel, scientists may consider presenting this channel in grayscale to optimally display fine details. This may include variations in staining intensity or fine structures. When opting for color, authors should use grayscale visibility tests (Fig 6) to determine whether visibility is compromised. This can occur when dark colors, such as magenta, red, or blue, are shown on a black background.

Fig 6. Visibility of colors/hues differs and depends on the background color.

Fig 6

The best contrast is achieved with grayscale images or dark hues on a light background (first row). Dark color hues, such as red and blue, on a dark background (last row), are least visible. Visibility can be tested with mock grayscale. Images show actin filaments in Dictyostelium discoideum (LifeAct-GFP). All images have the same scale. GFP, green fluorescent protein.

4. Choose a colorblind accessible color palette

Fluorescent images with merged color channels visualize the colocalization of different markers. While many readers find these images to be visually appealing and informative, these images are often inaccessible to colorblind coauthors, reviewers, editors, and readers. Deuteranopia, the most common form of colorblindness, affects up to 8% of men and 0.5% of women of northern European ancestry [12]. A study of articles published in top peripheral vascular disease journals revealed that 85% of papers with color maps and 58% of papers with heat maps used color palettes that were not colorblind safe [20]. We show that approximately half of cell biology papers, and one-third of physiology papers and plant science papers, contained images that were inaccessible to readers with deuteranopia. Scientists should consider the following points to ensure that images are accessible to colorblind readers.

  • Select colorblind safe colors: Researchers should use colorblind safe color palettes for fluorescence and other images where color may be adjusted. Fig 7 illustrates how 4 different color combinations would look to viewers with different types of color blindness. Green and red are indistinguishable to readers with deuteranopia, whereas green and blue are indistinguishable to readers with tritanopia, a rare form of color blindness. Cyan and magenta are the best options, as these 2 colors look different to viewers with normal color vision, deuteranopia, or tritanopia. Green and magenta are also shown, as scientists often prefer to show colors close to the excitation value of the fluorescent dyes, which are often green and red.

  • Display separate channels in addition to the merged image: Selecting a colorblind safe color palette becomes increasingly difficult as more colors are added. When the image includes 3 or more colors, authors are encouraged to show separate images for each channel, followed by the merged image (Fig 8). Individual channels may be shown in grayscale to make it easier for readers to perceive variations in staining intensity.

  • Use simulation tools to confirm that essential features are visible to colorblind viewers: Free tools, such as Color Oracle (RRID:SCR_018400), quickly simulate different forms of color blindness by adjusting the colors on the computer screen to simulate what a colorblind person would see. Scientists using FIJI (RRID:SCR002285) can select the “Simulate colorblindness” option in the “Color” menu under “Images.”

Fig 7. Color combinations as seen with normal vision and 2 types of colorblindness.

Fig 7

The figure illustrates how 4 possible color combinations for multichannel microscope images would appear to someone with normal color vision, the most common form of colorblindness (deuteranopia), and a rare form of color blindness (tritanopia). Some combinations that are accessible to someone with deuteranopia are not accessible to readers with tritanopia, for example, green/blue combinations. Microscope images show Dictyostelium discoideum expressing Vps32-GFP (Vps32-green fluorescent protein shows broad signal in cells) and stained with dextran (spotted signal) after infection with conidia of Aspergillus fumigatus. All images have the same scale. GFP, green fluorescent protein.

Fig 8. Strategies for making 2- or 3-channel microscope images colorblind safe.

Fig 8

Images in the first row are not colorblind safe. Readers with the most common form of colorblindness would not be able to identify key features. Possible accessible solutions are shown: changing colors/LUTs to colorblind-friendly combinations, showing each channel in a separate image, showing colors in grayscale and inverting grayscale images to maximize contrast. Solutions 3 and 4 (show each channel in grayscale, or in inverted grayscale) are more informative than solutions 1 and 2. Regions of overlap are sometimes difficult to see in merged images without split channels. When splitting channels, scientists often use colors that have low contrast, as explained in Fig 6 (e.g., red or blue on black). Microscope images show D. melanogaster egg chambers (2 colors) and nurse cell nuclei (3 colors) [18]. All images of egg chambers and nurse cells respectively have the same scale. LUT, lookup table.

5. Design the figure

Figures often contain more than one panel. Careful planning is needed to convey a clear message, while ensuring that all panels fit together and follow a logical order. A planning table (Fig 9A) helps scientists to determine what information is needed to answer the research question. The table outlines the objectives, types of visualizations required, and experimental groups that should appear in each panel. A planning table template is available on OSF [14]. After completing the planning table, scientists should sketch out the position of panels and the position of images, graphs, and titles within each panel (Fig 9B). Audiences read a page either from top to bottom and/or from left to right. Selecting one reading direction and arranging panels in rows or columns helps with figure planning. Using enough white space to separate rows or columns will visually guide the reader through the figure. The authors can then assemble the figure based on the draft sketch.

Fig 9. Planning multipanel figures.

Fig 9

Planning tables and layout sketches are useful tools to efficiently design figures that address the research question. (A) Planning tables allow scientists to select and organize elements needed to answer the research question addressed by the figure. (B) Layout sketches allow scientists to design a logical layout for all panels listed in the planning table and ensure that there is adequate space for all images and graphs.

6. Annotate the figure

Annotations with text, symbols, or lines allow readers from many different backgrounds to rapidly see essential features, interpret images, and gain insight. Unfortunately, scientists often design figures for themselves, rather than their audience [7]. Examples of annotations are shown in Fig 10. Table 1 describes important factors to consider for each annotation type.

Fig 10. Using arrows, regions of interest, lines, and letter codes to annotate structures in images.

Fig 10

Text descriptions alone are often insufficient to clearly point to a structure or region in an image. Arrows and arrowheads, lines, letters, and dashed enclosures can help if overlaid on the respective part of the image. Microscope images show D. melanogaster egg chambers [18], with the different labeling techniques in use. The table provides an overview of their applicability and common pitfalls. All images have the same scale.

Table 1. Use annotations to make figures accessible to a broad audience.
Feature to be Explained Annotation
Size Scale bar with dimensions
Direction of movement Arrow with tail
Draw attention to:
    • Points of interest Symbol (arrowhead, star, etc.)
    • Regions of interest: black and white image Highlight in color if this does not obscure important features within the region OR
Outline with boxes or circles
    • Regions of interest: Color image Outline with boxes or circles
    • Layers Labeled brackets beside the image for layers that are visually identifiable across the entire image OR
A line on the image for wavy layers that may be difficult to identify
Define features within an image Labels

When adding annotations to an image, scientists should consider the following steps.

  • Choose the right amount of labeling. Fig 11 shows 3 levels of annotation. The barely annotated image (Fig 11A) is only accessible to scientists already familiar with the object and technique, whereas the heavily annotated version (Fig 11C) contains numerous annotations that obstruct the image and a legend that is time consuming to interpret. Fig 11B is more readable; annotations of a few key features are shown, and the explanations appear right below the image for easy interpretation. Explanations of labels are often placed in the figure legend. Alternating between examining the figure and legend is time consuming, especially when the legend and figure are on different pages. Fig 11D shows one option for situations where extensive annotations are required to explain a complex image. An annotated image is placed as a legend next to the original image. A semitransparent white layer mutes the image to allow annotations to stand out.

  • Use abbreviations cautiously: Abbreviations are commonly used for image and figure annotation to save space but inevitably require more effort from the reader. Abbreviations are often ambiguous, especially across fields. Authors should run a web search for the abbreviation [21]. If the intended meaning is not a top result, authors should refrain from using the abbreviation or clearly define the abbreviation on the figure itself, even if it is already defined elsewhere in the manuscript. Note that in Fig 11, abbreviations have been written out below the image to reduce the number of legend entries.

  • Explain colors and stains: Explanations of colors and stains were missing in around 20% of papers. Fig 12 illustrates several problematic practices observed in our dataset, as well as solutions for clearly explaining what each color represents. This figure uses fluorescence images as an example; however, we also observed many histology images in which authors did not mention which stain was used. Authors should describe how stains affect the tissue shown or use annotations to show staining patterns of specific structures. This allows readers who are unfamiliar with the stain to interpret the image.

  • Ensure that annotations are accessible to colorblind readers: Confirming that labels or annotations are visible to colorblind readers is important for both color and grayscale images (Fig 13). Up to one-third of papers in our dataset contained annotations or labels that would not have been visible to someone with deuteranopia. This occurred because the annotations blended in with the background (e.g., red arrows on green plants) or the authors use the same symbol in colors that are indistinguishable to someone with deuteranopia to mark different features. Fig 13 illustrates how to annotate a grayscale image so that it is accessible to color blind readers. Using text to describe colors is also problematic for colorblind readers. This problem can be alleviated by using colored symbols in the legend or by using distinctly shaped annotations such as open versus closed arrows, thin versus wide lines, or dashed versus solid lines. Color blindness simulators help in determining whether annotations are accessible to all readers.

Fig 11. Different levels of detail for image annotations.

Fig 11

Annotations help to orient the audience but may also obstruct parts of the image. Authors must find the right balance between too few and too many annotations. (1) Example with no annotations. Readers cannot determine what is shown. (2) Example with a few annotations to orient readers to key structures. (3) Example with many annotations, which obstruct parts of the image. The long legend below the figure is confusing. (4) Example shows a solution for situations where many annotations are needed to explain the image. An annotated version is placed next to an unannotated version of the image for comparison. The legend below the image helps readers to interpret the image, without having to refer to the figure legend. Note the different requirements for space. Electron microscope images show mouse pancreatic beta-islet cells.

Fig 12. Explain color in images.

Fig 12

Cells and their structures are almost all transparent. Every dye, stain, and fluorescent label therefore should be clearly explained to the audience. Labels should be colorblind safe. Large labels that stand out against the background are easy to read. Authors can make figures easier to interpret by placing the color label close to the structure; color labels should only be placed in the figure legend when this is not possible. Example images were created based on problems observed by reviewers. Microscope images show D. melanogaster egg chambers stained with the DNA dye DAPI (4′,6-diamidino-2-phenylindole) and probe for a specific mRNA species [18]. All images have the same scale.

Fig 13. Annotations should be colorblind safe.

Fig 13

(1) The annotations displayed in the first image are inaccessible to colorblind individuals, as shown with the visibility test below. This example was created based on problems observed by reviewers. (2, 3) Two colorblind safe alternative annotations, in color (2) and in grayscale (3). The bottom row shows a test rendering for deuteranopia colorblindness. Note that double-encoding of different hues and different shapes (e.g., different letters, arrow shapes, or dashed/nondashed lines) allows all audiences to interpret the annotations. Electron microscope images show mouse pancreatic beta-cell islet cells. All images have the same scale.

7. Prepare figure legends

Each figure and legend are meant to be self-explanatory and should allow readers to quickly assess a paper or understand complex studies that combine different methodologies or model systems. To date, there are no guidelines for figure legends for images, as the scope and length of legends varies across journals and disciplines. Some journals require legends to include details on object, size, methodology, or sample size, while other journals require a minimalist approach and mandate that information should not be repeated in subsequent figure legends.

Our data suggest that important information needed to interpret images was regularly missing from the figure or figure legend. This includes the species and tissue type, or object shown in the figure, clear explanations of all labels, annotations and colors, and markings or legend entries denoting insets. Presenting this information on the figure itself is more efficient for the reader; however, any details that are not marked in the figure should be explained in the legend.

While not reporting species and tissue information in every figure legend may be less of an issue for papers that examine a single species and tissue, this is a major problem when a study includes many species and tissues, which may be presented in different panels of the same figure. Additionally, the scientific community is increasingly developing automated data mining tools, such as the Source Data tool, to collect and synthesize information from figures and other parts of scientific papers. Unlike humans, these tools cannot piece together information scattered throughout the paper to determine what might be shown in a particular figure panel. Even for human readers, this process wastes time. Therefore, we recommend that authors present information in a clear and accessible manner, even if some information may be repeated for studies with simple designs.

Discussion

A flood of images is published every day in scientific journals and the number is continuously increasing. Of these, around 4% likely contain intentionally or accidentally duplicated images [3]. Our data show that, in addition, most papers show images that are not fully interpretable due to issues with scale markings, annotation, and/or color. This affects scientists’ ability to interpret, critique, and build upon the work of others. Images are also increasingly submitted to image archives to make image data widely accessible and permit future reanalyses. A substantial fraction of images that are neither human nor machine-readable lowers the potential impact of such archives. Based on our data examining common problems with published images, we provide a few simple recommendations, with examples illustrating good practices. We hope that these recommendations will help authors to make their published images legible and interpretable.

Limitations: While most results were consistent across the 3 subfields of biology, findings may not be generalizable to other fields. Our sample included the top 15 journals that publish original research for each field. Almost all journals were indexed in PubMed. Results may not be generalizable to journals that are unindexed, have low impact factors, or are not published in English. Data abstraction was performed manually due to the complexity of the assessments. Error rates were 5% for plant sciences, 4% for physiology, and 3% for cell biology. Our assessments focused on factors that affect readability of image-based figures in scientific publications. Future studies may include assessments of raw images and meta-data to examine factors that affect reproducibility, such as contrast settings, background filtering, and processing history.

Actions journals can take to make image-based figures more transparent and easier to interpret

The role of journals in improving the quality of reporting and accessibility of image-based figures should not be overlooked. There are several actions that journals might consider.

  • Screen manuscripts for figures that are not colorblind safe: Open source automated screening tools [22] may help journals to efficiently identify common color maps that are not colorblind safe.

  • Update journal policies: We encourage journal editors to update policies regarding colorblind accessibility, scale bars, and other factors outlined in this manuscript. Importantly, policy changes should be accompanied by clear plans for implementation and enforcement. Meta-research suggests that changing journal policy, without enforcement or implementation plans, has limited effects on author behavior. Amending journal policies to require authors to report research resource identifiers (RRIDs), for example, increases the number of papers reporting RRIDs by 1% [23]. In a study of life sciences articles published in Nature journals, the percentage of animal studies reporting the Landis 4 criteria (blinding, randomization, sample size calculation, exclusions) increased from 0% to 16.4% after new guidelines were released [24]. In contrast, a randomized controlled trial of animal studies submitted to PLOS ONE demonstrated that randomizing authors to complete the ARRIVE checklist during submission did not improve reporting [25]. Some improvements in reporting of confidence intervals, sample size justification, and inclusion and exclusion criteria were noted after Psychological Science introduced new policies [26], although this may have been partially due to widespread changes in the field. A joint editorial series published in the Journal of Physiology and British Journal of Pharmacology did not improve the quality of data presentation or statistical reporting [27].

  • Reevaluate limits on the number of figures: Limitations on the number of figures originally stemmed from printing costs calculations, which are becoming increasingly irrelevant as scientific publishing moves online. Unintended consequences of these policies include the advent of large, multipanel figures. These figures are often especially difficult to interpret because the legend appears on a different page, or the figure combines images addressing different research questions.

  • Reduce or eliminate page charges for color figures: As journals move online, policies designed to offset the increased cost of color printing are no longer needed. The added costs may incentivize authors to use grayscale in cases where color would be beneficial.

  • Encourage authors to explain labels or annotations in the figure, rather than in the legend: This is more efficient for readers.

  • Encourage authors to share image data in public repositories: Open data benefits authors and the scientific community [2830].

How can the scientific community improve image-based figures?

The role of scientists in the community is multifaceted. As authors, scientists should familiarize themselves with guidelines and recommendations, such as ours provided above. As reviewers, scientists should ask authors to improve erroneous or uninformative image-based figures. As instructors, scientists should ensure that bioimaging and image data handling is taught during undergraduate or graduate courses, and support existing initiatives such as NEUBIAS (Network of EUropean BioImage AnalystS) [31] that aim to increase training opportunities in bioimage analysis.

Scientists are also innovators. As such, they should support emerging image data archives, which may expand to automatically source images from published figures. Repositories for other types of data are already widespread; however, the idea of image repositories has only recently gained traction [32]. Existing image databases, which are mainly used for raw image data and meta-data, include the Allen Brain Atlas, the Image Data Resource [33], and the emerging BioImage Archives [32]. Springer Nature encourages authors to submit imaging data to the Image Data Resource [33]. While scientists have called for common quality standards for archived images and meta-data [32], such standards have not been defined, implemented, or taught. Examining standard practices for reporting images in scientific publications, as outlined here, is one strategy for establishing common quality standards.

In the future, it is possible that each image published electronically in a journal or submitted to an image data repository will follow good practice guidelines and will be accompanied by expanded “meta-data” or “alt-text/attribute” files. Alt-text is already published in html to provide context if an image cannot be accessed (e.g., by blind readers). Similarly, images in online articles and deposited in archives could contain essential information in a standardized format. The information could include the main objective of the figure, specimen information, ideally with RRID [34], specimen manipulation (dissection, staining, RRID for dyes and antibodies used), as well as the imaging method including essential items from meta-files of the microscope software, information about image processing and adjustments, information about scale, annotations, insets, and colors shown, and confirmation that the images are truly representative.

Conclusions

Our meta-research study of standard practices for presenting images in 3 fields highlights current shortcomings in publications. Pubmed indexes approximately 800,000 new papers per year, or 2,200 papers per day (https://www.nlm.nih.gov/bsd/index_stats_comp.html). Twenty-three percent [1], or approximately 500 papers per day, contain images. Our survey data suggest that most of these papers will have deficiencies in image presentation, which may affect legibility and interpretability. These observations lead to targeted recommendations for improving the quality of published images. Our recommendations are available as a slide set via the OSF and can be used in teaching best practice to avoid misleading or uninformative image-based figures. Our analysis underscores the need for standardized image publishing guidelines. Adherence to such guidelines will allow the scientific community to unlock the full potential of image collections in the life sciences for current and future generations of researchers.

Methods

Systematic review

We examined original research articles that were published in April of 2018 in the top 15 journals that publish original research for each of 3 different categories (physiology, plant science, cell biology). Journals for each category were ranked according to 2016 impact factors listed for the specified categories in Journal Citation Reports. Journals that only publish review articles or that did not publish an April issue were excluded. We followed all relevant aspects of the PRISMA guidelines [35]. Items that only apply to meta-analyses or are not relevant to literature surveys were not followed. Ethical approval was not required.

Search strategy

Articles were identified through a PubMed search, as all journals were PubMed indexed. Electronic search results were verified by comparison with the list of articles published in April issues on the journal website. The electronic search used the following terms:

Physiology: ("Journal of pineal research"[Journal] AND 3[Issue] AND 64[Volume]) OR ("Acta physiologica (Oxford, England)"[Journal] AND 222[Volume] AND 4[Issue]) OR ("The Journal of physiology"[Journal] AND 596[Volume] AND (7[Issue] OR 8[Issue])) OR (("American journal of physiology. Lung cellular and molecular physiology"[Journal] OR "American journal of physiology. Endocrinology and metabolism"[Journal] OR "American journal of physiology. Renal physiology"[Journal] OR "American journal of physiology. Cell physiology"[Journal] OR "American journal of physiology. Gastrointestinal and liver physiology"[Journal]) AND 314[Volume] AND 4[Issue]) OR (“American journal of physiology. Heart and circulatory physiology”[Journal] AND 314[Volume] AND 4[Issue]) OR ("The Journal of general physiology"[Journal] AND 150[Volume] AND 4[Issue]) OR ("Journal of cellular physiology"[Journal] AND 233[Volume] AND 4[Issue]) OR ("Journal of biological rhythms"[Journal] AND 33[Volume] AND 2[Issue]) OR ("Journal of applied physiology (Bethesda, Md.: 1985)"[Journal] AND 124[Volume] AND 4[Issue]) OR ("Frontiers in physiology"[Journal] AND ("2018/04/01"[Date—Publication]: "2018/04/30"[Date—Publication])) OR ("The international journal of behavioral nutrition and physical activity"[Journal] AND ("2018/04/01"[Date—Publication]: "2018/04/30"[Date—Publication])).

Plant science: ("Nature plants"[Journal] AND 4[Issue] AND 4[Volume]) OR ("Molecular plant"[Journal] AND 4[Issue] AND 11[Volume]) OR ("The Plant cell"[Journal] AND 4[Issue] AND 30[Volume]) OR ("Plant biotechnology journal"[Journal] AND 4[Issue] AND 16[Volume]) OR ("The New phytologist"[Journal] AND (1[Issue] OR 2[Issue]) AND 218[Volume]) OR ("Plant physiology"[Journal] AND 4[Issue] AND 176[Volume]) OR ("Plant, cell & environment"[Journal] AND 4[Issue] AND 41[Volume]) OR ("The Plant journal: for cell and molecular biology"[Journal] AND (1[Issue] OR 2[Issue]) AND 94[Volume]) OR ("Journal of experimental botany"[Journal] AND (8[Issue] OR 9[Issue] OR 10[Issue]) AND 69[Volume]) OR ("Plant & cell physiology"[Journal] AND 4[Issue] AND 59[Volume]) OR ("Molecular plant pathology"[Journal] AND 4[Issue] AND 19[Volume]) OR ("Environmental and experimental botany"[Journal] AND 148[Volume]) OR ("Molecular plant-microbe interactions: MPMI"[Journal] AND 4[Issue] AND 31[Volume]) OR (“Frontiers in plant science”[Journal] AND ("2018/04/01"[Date—Publication]: "2018/04/30"[Date—Publication])) OR (“The Journal of ecology” ("2018/04/01"[Date—Publication]: "2018/04/30"[Date—Publication])).

Cell biology: ("Cell"[Journal] AND (2[Issue] OR 3[Issue]) AND 173[Volume]) OR ("Nature medicine"[Journal] AND 24[Volume] AND 4[Issue]) OR ("Cancer cell"[Journal] AND 33[Volume] AND 4[Issue]) OR ("Cell stem cell"[Journal] AND 22[Volume] AND 4[Issue]) OR ("Nature cell biology"[Journal] AND 20[Volume] AND 4[Issue]) OR ("Cell metabolism"[Journal] AND 27[Volume] AND 4[Issue]) OR ("Science translational medicine"[Journal] AND 10[Volume] AND (435[Issue] OR 436[Issue] OR 437[Issue] OR 438[Issue])) OR ("Cell research"[Journal] AND 28[Volume] AND 4[Issue]) OR ("Molecular cell"[Journal] AND 70[Volume] AND (1[Issue] OR 2[Issue])) OR("Nature structural & molecular biology"[Journal] AND 25[Volume] AND 4[Issue]) OR ("The EMBO journal"[Journal] AND 37[Volume] AND (7[Issue] OR 8[Issue])) OR ("Genes & development"[Journal] AND 32[Volume] AND 7–8[Issue]) OR ("Developmental cell"[Journal] AND 45[Volume] AND (1[Issue] OR 2[Issue])) OR ("Current biology: CB"[Journal] AND 28[Volume] AND (7[Issue] OR 8[Issue])) OR ("Plant cell"[Journal] AND 30[Volume] AND 4[Issue]).

Screening

Screening for each article was performed by 2 independent reviewers (Physiology: TLW, SS, EMW, VI, KW, MO; Plant science: TLW, SJB; Cell biology: EW, SS) using Rayyan software (RRID:SCR_017584), and disagreements were resolved by consensus. A list of articles was uploaded into Rayyan. Reviewers independently examined each article and marked whether the article was included or excluded, along with the reason for exclusion. Both reviewers screened all articles published in each journal between April 1 and April 30, 2018, to identify full length, original research articles (S1S3 Tables, S1 Fig) published in the print issue of the journal. Articles for online journals that do not publish print issues were included if the publication date was between April 1 and April 30, 2018. Articles were excluded if they were not original research articles, or if an accepted version of the paper was posted as an “in press” or “early release” publication; however, the final version did not appear in the print version of the April issue. Articles were included if they contained at least one eligible image, such as a photograph, an image created using a microscope or electron microscope, or an image created using a clinical imaging technology such as ultrasound or MRI. Blot images were excluded, as many of the criteria in our abstraction protocol cannot easily be applied to blots. Computer generated images, graphs, and data figures were also excluded. Papers that did not contain any eligible images were excluded.

Abstraction

All abstractors completed a training set of 25 articles before abstracting data. Data abstraction for each article was performed by 2 independent reviewers (Physiology: AA, AV; Plant science: MO, TLA, SA, KW, MAG, IF; Cell biology: IF, AA, AV, KW, MAG). When disagreements could not be resolved by consensus between the 2 reviewers, ratings were assigned after a group review of the paper. Eligible manuscripts were reviewed in detail to evaluate the following questions according to a predefined protocol (available at: https://doi.org/10.17605/OSF.IO/B5296) [14]. Supplemental files were not examined, as supplemental images may not be held to the same peer review standards as those in the manuscript.

The following items were abstracted:

  1. Types of images included in the paper (photograph, microscope image, electron microscope image, image created using a clinical imaging technique such as ultrasound or MRI, other types of images)

  2. Did the paper contain appropriately labeled scale bars for all images?

  3. Were all insets clearly and accurately marked?

  4. Were all insets clearly explained in the legend?

  5. Is the species and tissue, object, or cell line name clearly specified in the figure or legend for all images in the paper?

  6. Are any annotations, arrows, or labels clearly explained for all images in the paper?

  7. Among images where authors can control the colors shown (e.g., fluorescence microscopy), are key features of the images visible to someone with the most common form of colorblindness (deuteranopia)?

  8. If the paper contains colored labels, are these labels visible to someone with the most common form of color blindness (deuteranopia)?

  9. Are colors in images explained either on the image or within the legend?

Questions 7 and 8 were assessed by using Color Oracle [36] (RRID:SCR_018400) to simulate the effects of deuteranopia.

Verification

Ten percent of articles in each field were randomly selected for verification abstraction, to ensure that abstractors in different fields were following similar procedures. Data were abstracted by a single abstractor (TLW). The question on species and tissue was excluded from verification abstraction for articles in cell biology and plant sciences, as the verification abstractor lacked the field-specific expertise needed to assess this question. Results from the verification abstractor were compared with consensus results from the 2 independent abstractors for each paper, and discrepancies were resolved through discussion. Error rates were calculated as the percentage of responses for which the abstractors’ response was incorrect. Error rates were 5% for plant sciences, 4% for physiology, and 3% for cell biology.

Data processing and creation of figures

Data are presented as n (%). Summary statistics were calculated using Python (RRID:SCR_008394, version 3.6.9, libraries NumPy 1.18.5 and Matplotlib 3.2.2). Charts were prepared with a Python-based Jupyter Notebook (Jupyter-client, RRID:SCR_018413 [37], Python version 3.6.9, RRID:SCR_008394, libraries NumPy 1.18.5 [38], and Matplotlib 3.2.2 [39]) and assembled into figures with vector graphic software. Example images were previously published or generously donated by the manuscript authors as indicated in the figure legends. Image acquisition was described in references (D. melanogaster images [18], mouse pancreatic beta islet cells: A. Müller personal communication, and Orobates pabsti [19]). Images were cropped, labeled, and color-adjusted with FIJI [15] (RRID:SCR_002285) and assembled with vector-graphic software. Colorblind and grayscale rendering of images was done using Color Oracle [36] (RRID:SCR_018400). All poor and clear images presented here are “mock examples” prepared based on practices observed during data abstraction.

Supporting information

S1 Fig. Flow chart of study screening and selection process.

This flow chart illustrates the number of included and excluded journals or articles, along with reasons for exclusion, at each stage of the study.

(JPG)

S1 Table. Number of articles examined by journal in physiology.

Values are n, or n (% of all articles). Screening was performed to exclude articles that were not full-length original research articles (e.g., reviews, editorials, perspectives, commentaries, letters to the editor, short communications, etc.), were not published in April 2018, or did not include eligible images. AJP, American Journal of Physiology.

(DOCX)

S2 Table. Number of articles examined by journal in plant science.

Values are n, or n (% of all articles). Screening was performed to exclude articles that were not full-length original research articles (e.g., reviews, editorials, perspectives, commentaries, letters to the editor, short communications, etc.), were not published in April 2018, or did not include eligible images. *This journal was also included on the cell biology list (Table S3). **No articles from the Journal of Ecology were screened as the journal did not publish an April 2018 issue.

(DOCX)

S3 Table. Number of articles examined by journal in cell biology.

Values are n, or n (% of all articles). Screening was performed to exclude articles that were not full-length original research articles (e.g., reviews, editorials, perspectives, commentaries, letters to the editor, short communications, etc.), were not published in April 2018, or did not include eligible images. *This journal was also included on the plant science list (Table S2).

(DOCX)

S4 Table. Scale information in papers.

Values are percent of papers.

(DOCX)

Acknowledgments

We thank the eLife Community Ambassadors program for facilitating this work, and Andreas Müller and John A. Nyakatura for generously sharing example images. Falk Hillmann and Thierry Soldati provided the amoeba strains used for imaging. Some of the early career researchers who participated in this research would like to thank their principal investigators and mentors for supporting their efforts to improve science.

Abbreviations

GFP

green fluorescent protein

LUT

lookup table

OSF

Open Science Framework

RRID

research resource identifier

Data Availability

The authors confirm that all data underlying the findings are fully available without restriction. The abstraction protocol, data, code and slides for teaching are available on an OSF repository (https://doi.org/10.17605/OSF.IO/B5296).

Funding Statement

TLW was funded by American Heart Association grant 16GRNT30950002 (https://www.heart.org/en/professional/institute/grants) and a Robert W. Fulk Career Development Award (Mayo Clinic Division of Nephrology & Hypertension; https://www.mayoclinic.org/departments-centers/nephrology-hypertension/sections/overview/ovc-20464571). LHH was supported by The Hormel Foundation and National Institutes of Health grant CA187035 (https://www.nih.gov). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Lee P, West JD, Viziometrics HB. Analyzing Visual Information in the Scientific Literature. IEEE Transactions on Big Data. 2018;4:117–29. [Google Scholar]
  • 2.Cromey DW. Digital images are data: and should be treated as such. Methods Mol Biol. 2013;931:1–27. 10.1007/978-1-62703-056-4_1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Bik EM, Casadevall A, Fang FC. The Prevalence of Inappropriate Image Duplication in Biomedical Research Publications. MBio. 2016;7. 10.1128/mBio.00809-16 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Laissue PP, Alghamdi RA, Tomancak P, Reynaud EG, Shroff H. Assessing phototoxicity in live fluorescence imaging. Nat Methods. 2017;14:657–61. 10.1038/nmeth.4344 [DOI] [PubMed] [Google Scholar]
  • 5.Marques G, Pengo T, Sanders MA. Imaging methods are vastly underreported in biomedical research. Elife. 2020;9. 10.7554/eLife.55133 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Pain E. How to (seriously) read a scientific paper. Science. 2016. [Google Scholar]
  • 7.Rolandi M, Cheng K. Perez-Kriz S. A brief guide to designing effective figures for the scientific paper. Adv Mater. 2011;23:4343–6. 10.1002/adma.201102518 [DOI] [PubMed] [Google Scholar]
  • 8.Canese K. PubMed® Display Enhanced with Images from the New NCBI Images Database. NLM Technical Bulletin. 2010;376:e14. [Google Scholar]
  • 9.Liechti R, George N, Gotz L, El-Gebali S, Chasapi A, Crespo I, et al. SourceData: a semantic platform for curating and searching figures. Nat Methods. 2017;14:1021–2. 10.1038/nmeth.4471 [DOI] [PubMed] [Google Scholar]
  • 10.Lindquist M. Neuroimaging results altered by varying analysis pipelines. Nature. 2020;582:36–7. 10.1038/d41586-020-01282-z [DOI] [PubMed] [Google Scholar]
  • 11.Botvinik-Nezer R, Holzmeister F, Camerer CF, Dreber A, Huber J, Johannesson M, et al. Variability in the analysis of a single neuroimaging dataset by many teams. Nature. 2020;582:84–8. 10.1038/s41586-020-2314-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.National Eye Institute. Facts about color blindness. 2015. https://nei.nih.gov/health/color_blindness/facts_about. [cited 2019. March 13]. [Google Scholar]
  • 13.Weissgerber TL. Training early career researchers to use meta-research to improve science: A participant guided, “learn by doing” approach. PLoS Biol. 2021;19:e3001073. 10.1371/journal.pbio.3001073 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Antonietti A, Jambor H, Alicea B, Audisio TL, Auer S, Bhardwaj V, et al. Meta-research: Creating clear and informative image-based figures for scientific publications. 2020. 10.17605/OSF.IO/B5296. March 5, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Schindelin J, Arganda-Carreras I, Frise E, Kaynig V, Longair M, Pietzsch T, et al. Fiji: an open-source platform for biological-image analysis. Nat Methods. 2012;9:676–82. 10.1038/nmeth.2019 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Rueden CT, Schindelin J, Hiner MC, DeZonia BE, Walter AE, Arena ET, et al. ImageJ2: ImageJ for the next generation of scientific image data. BMC Bioinformatics. 2017;18:529. 10.1186/s12859-017-1934-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Schmied C, Jambor HK. Effective image visualization for publications—a workflow using open access tools and concepts. F1000Research. 2020;9:1373. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Jambor H, Surendranath V, Kalinka AT, Mejstrik P, Saalfeld S, Tomancak P. Systematic imaging reveals features and changing localization of mRNAs in Drosophila development. Elife. 2015;4. 10.7554/eLife.05003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Nyakatura JA, Melo K, Horvat T, Karakasiliotis K, Allen VR, Andikfar A, et al. Reverse-engineering the locomotion of a stem amniote. Nature. 2019;565:351–5. 10.1038/s41586-018-0851-2 [DOI] [PubMed] [Google Scholar]
  • 20.Weissgerber TL, Winham SJ, Heinzen EP, Milin-Lazovic JS, Garcia-Valencia O, Bukumiric Z, et al. Reveal, Don’t Conceal: Transforming Data Visualization to Improve Transparency. Circulation. 2019;140:1506–18. 10.1161/CIRCULATIONAHA.118.037777 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Jambor H. Better figures for the life sciences. ecrLife. August 29, 2018. https://ecrlife420999811.wordpress.com/2018/08/29/better-figures-for-life-sciences/. [cited 2020 September 15]. [Google Scholar]
  • 22.Saladi S. JetFighter: Towards figure accuracy and accessibility. Elife. 2019. [Google Scholar]
  • 23.Bandrowski A, Brush M, Grethe JS, Haendel MA, Kennedy DN, Hill S, et al. Resource Identification Initiative Members. The Resource Identification Initiative: A cultural shift in publishing. F1000Res. 2015;4:134. 10.12688/f1000research.6555.2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.the NPQIP Collaborative Group. Did a change in Nature journals’ editorial policy for life sciences research improve reporting? BMJ Open Science. 2019;3:e000035. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Hair K, Macleod MR, Sena ES, Collaboration II. A randomised controlled trial of an Intervention to Improve Compliance with the ARRIVE guidelines (IICARus). Res Integr Peer Rev. 2019;4:12. 10.1186/s41073-019-0069-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Giofre D, Cumming G, Fresc L, Boedker I, Tressoldi P. The influence of journal submission guidelines on authors’ reporting of statistics and use of open research practices. PLoS ONE. 2017;12:e0175583. 10.1371/journal.pone.0175583 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Diong J, Butler AA, Gandevia SC, Heroux ME. Poor statistical reporting, inadequate data presentation and spin persist despite editorial advice. PLoS ONE. 2018;13:e0202121. 10.1371/journal.pone.0202121 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Piwowar HA, Vision TJ. Data reuse and the open data citation advantage. PeerJ. 2013;1:e175. 10.7717/peerj.175 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Markowetz F. Five selfish reasons to work reproducibly. Genome Biol. 2015;16:274. 10.1186/s13059-015-0850-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Colavizza G, Hrynaszkiewicz I, Staden I, Whitaker K, McGillivray B. The citation advantage of linking publications to research data. arXiv. 2020. 10.1371/journal.pone.0230416 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Cimini BA, Norrelykke SF, Louveaux M, Sladoje N, Paul-Gilloteaux P, Colombelli J, et al. The NEUBIAS Gateway: a hub for bioimage analysis methods and materials. F1000Res. 2020;9:613. 10.12688/f1000research.24759.1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Ellenberg J, Swedlow JR, Barlow M, Cook CE, Sarkans U, Patwardhan A, et al. A call for public archives for biological image data. Nat Methods. 2018;15:849–54. 10.1038/s41592-018-0195-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Williams E, Moore J, Li SW, Rustici G, Tarkowska A, Chessel A, et al. The Image Data Resource: A Bioimage Data Integration and Publication Platform. Nat Methods. 2017;14:775–81. 10.1038/nmeth.4326 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Bandrowski AE, Martone MERRID. A Simple Step toward Improving Reproducibility through Rigor and Transparency of Experimental Methods. Neuron. 2016;90:434–6. 10.1016/j.neuron.2016.04.030 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Moher D, Liberati A, Tetzlaff J, Altman DG. Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement. J Clin Epidemiol. 2009;62:1006–12. 10.1016/j.jclinepi.2009.06.005 [DOI] [PubMed] [Google Scholar]
  • 36.Jenny B and Kelso NV. Color Oracle. 2018. https://colororacle.org. [cited 2020 March 3]. [Google Scholar]
  • 37.Kluyver T, Ragan-Kelley B, Pérez F and Granger B. Jupyter Notebooks—a publishing format for reproducible computational workflows. In: Scmidt F. L. a. B., ed. Positioning and Power in Academic Publishing: Players, Agents and Agendas Netherlands: IOS Press; 2016. [Google Scholar]
  • 38.Harris CR, Millman KJ, van der Walt SJ, Gommers R, Virtanen P, Cornapeau D, et al. Array programming with NumPy. Nature. 2020;585:357–62. 10.1038/s41586-020-2649-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Hunter JD. Matplotlib: A 2D graphics environment. Comput Sci Eng. 2007;9:90–5. [Google Scholar]

Decision Letter 0

Roland G Roberts

28 Oct 2020

Dear Dr Weissgerber,

Thank you for submitting your manuscript entitled "Creating Clear and Informative Image-based Figures for Scientific Publications" for consideration as a Meta-Research Article by PLOS Biology.

Your manuscript has now been evaluated by the PLOS Biology editorial staff as well as by an academic editor with relevant expertise and I am writing to let you know that we would like to send your submission out for external peer review.

However, before we can send your manuscript to reviewers, we need you to complete your submission by providing the metadata that is required for full assessment. To this end, please login to Editorial Manager where you will find the paper in the 'Submissions Needing Revisions' folder on your homepage. Please click 'Revise Submission' from the Action Links and complete all additional questions in the submission questionnaire.

Please re-submit your manuscript within two working days, i.e. by Oct 30 2020 11:59PM.

Login to Editorial Manager here: https://www.editorialmanager.com/pbiology

Once your full submission is complete, your paper will undergo a series of checks in preparation for peer review, after which it will be sent out for review.

Given the disruptions resulting from the ongoing COVID-19 pandemic, please expect some delays in the editorial process. We apologise in advance for any inconvenience caused and will do our best to minimize impact as far as possible.

Feel free to email us at plosbiology@plos.org if you have any queries relating to your submission.

Kind regards,

Roland G Roberts, PhD,

Senior Editor

PLOS Biology

Decision Letter 1

Roland G Roberts

9 Dec 2020

Dear Dr Weissgerber,

Thank you very much for submitting your manuscript "Creating Clear and Informative Image-based Figures for Scientific Publications" for consideration as a Meta-Research Article at PLOS Biology. Your manuscript has been evaluated by the PLOS Biology editors, an Academic Editor with relevant expertise, and by five independent reviewers. I must apologise for the excessive number of reviewers; we usually aim for three or four, but an administrative oversight led to us recruiting an extra one. I hope that you nevertheless find all the comments useful.

You'll see that the reviewers are broadly positive about your study, but each raises a number of concerns and makes suggestions for improvement. In light of the reviews (below), we are pleased to offer you the opportunity to address the from the reviewers in a revised version that we anticipate should not take you very long. We will then assess your revised manuscript and your response to the reviewers' comments and we may consult the reviewers again.

We expect to receive your revised manuscript within 1 month.

Please email us (plosbiology@plos.org) if you have any questions or concerns, or would like to request an extension. At this stage, your manuscript remains formally under active consideration at our journal; please notify us by email if you do not intend to submit a revision so that we may end consideration of the manuscript at PLOS Biology.

**IMPORTANT - SUBMITTING YOUR REVISION**

Your revisions should address the specific points made by each reviewer. Please submit the following files along with your revised manuscript:

1. A 'Response to Reviewers' file - this should detail your responses to the editorial requests, present a point-by-point response to all of the reviewers' comments, and indicate the changes made to the manuscript.

*NOTE: In your point by point response to the reviewers, please provide the full context of each review. Do not selectively quote paragraphs or sentences to reply to. The entire set of reviewer comments should be present in full and each specific point should be responded to individually.

You should also cite any additional relevant literature that has been published since the original submission and mention any additional citations in your response.

2. In addition to a clean copy of the manuscript, please also upload a 'track-changes' version of your manuscript that specifies the edits made. This should be uploaded as a "Related" file type.

*Resubmission Checklist*

When you are ready to resubmit your revised manuscript, please refer to this resubmission checklist: https://plos.io/Biology_Checklist

To submit a revised version of your manuscript, please go to https://www.editorialmanager.com/pbiology/ and log in as an Author. Click the link labelled 'Submissions Needing Revision' where you will find your submission record.

Please make sure to read the following important policies and guidelines while preparing your revision:

*Published Peer Review*

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. Please see here for more details:

https://blogs.plos.org/plos/2019/05/plos-journals-now-open-for-published-peer-review/

*PLOS Data Policy*

Please note that as a condition of publication PLOS' data policy (http://journals.plos.org/plosbiology/s/data-availability) requires that you make available all data used to draw the conclusions arrived at in your manuscript. If you have not already done so, you must include any data used in your manuscript either in appropriate repositories, within the body of the manuscript, or as supporting information (N.B. this includes any numerical values that were used to generate graphs, histograms etc.). For an example see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5

*Blot and Gel Data Policy*

We require the original, uncropped and minimally adjusted images supporting all blot and gel results reported in an article's figures or Supporting Information files. We will require these files before a manuscript can be accepted so please prepare them now, if you have not already uploaded them. Please carefully read our guidelines for how to prepare and upload this data: https://journals.plos.org/plosbiology/s/figures#loc-blot-and-gel-reporting-requirements

*Protocols deposition*

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosbiology/s/submission-guidelines#loc-materials-and-methods

Thank you again for your submission to our journal. We hope that our editorial process has been constructive thus far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Roli Roberts

Roland G Roberts, PhD,

Senior Editor,

rroberts@plos.org,

PLOS Biology

*****************************************************

REVIEWERS' COMMENTS:

Reviewer #1:

[identifies herself as Elisabeth Bik]

In this paper, the authors screened hundreds of papers from three different scientific fields (physiology, cell biology, and plant sciences) and selected 580 papers that included photographic images. They analyzed the papers containing photographic images for the presence of scale bars, inset annotation, clear labeling, colorblindness-friendly color scheme, adequate description of the specimen etc. The majority of the papers failed one of these criteria. Examples of good and bad image labeling are given throughout the manuscript.

The paper is a welcome addition to the field of meta-science (science about science papers, and provides clear guidelines about what constitutes good labeling and color-use of photographic images in biomedical papers. The search strategy is clearly described and reproducible, and the paper was easy to read and understand. Also, kudos to the authors for including an image featuring Darth Vader.

I have some minor comments.

General comments:

It would be nice if the Abstract should include the total number of papers (580) screened for this study - that number is somewhat hard to find. It is included in Figure S1 (flow chart) and the discussion but it would be good to include it in the abstract and the first paragraph of the Results (see below).

The term "Microphotograph" might benefit from a definition. It appears the authors mean a photo taken from a specimen under a microscope (e.g. of cells or tissues), but I am not sure. Is a "Photograph" then defined as a photo of something visible to the eye such as a plant or a petridish? One could call all the image types mentioned in Figure 1A "photographs", so maybe consider using the term "macrophotograph" for a photo that is not a microphotograph.

Are the examples shown in Figure 4-6 from the papers that were screened for this paper? Or were they taken from public sources (as indicated for some photos) and then manipulated digitally to either remove or add a scale bar (see fig 4)? It would be nice to clearly define that in the Methods (or maybe I missed that).

Specific comments

Page 1, Affiliations of the authors: Typo: "Uterecht"

Introduction. At the end of the Introduction, and the end of "Using a science of science approach...." on Page 4, there are several references to specific figures. I would personally not expect these in the Results, but rather in the Introduction, so maybe consider removing part of that last paragraph of "Using a science...." to the beginning of the Results?

Results. Page 4. It would be more clear to start the Results section by mentioning how many papers (580) were screened.

Results. Page 4. "More than half of the papers in the sample contained images (plant science: 68%, cell biology: 72%, physiology: 55%)." - These numbers do not seem to match the data provided in Supplemental Tables 1-3. Maybe I am misunderstanding something, but Supplemental Tables 1-3 mention 39.9, 51.2, and 38.9% of papers, which are much lower numbers.

Physiology: 431 screened - 172 included (39.9%)

Plant science: 502 screened - 257 included (51.2%)

Cell Biology: 409 screened - 159 included (38.9%)

On page 6, "Approximately half of the papers (47-58%) also failed or partially failed to adequately explain insets. " appears to refer to Figure 1C, right panel, but the figure number/panel is not mentioned. Maybe add that?

Page 11, under 3 "Use Color wisely in images", "Images showing ielectron micrographs" should perhaps read "Images showing electron microphotographs"

Page 13, Maybe write "Deuteranopia, the most common form of colorblindness..." to remind the reader of what the term means (used a lot in the following paragraph)

Discussion. Page 22: "intentionally or accidentally manipulated images" - should be "intentionally or accidentally duplicated images"

Page 22: What is meant by "Error rates" here? The numbers listed here do not appear to match anything else in the paper. Maybe a reference or reminder needs to be included here.

Discussion. Page 22: "Actions journals can take to make image-based figures more transparent and easier to interpret". An important item not listed here, but that I personally think is very important, is to add particular requirements about e.g. the use of colorblind-safe colors and inclusion of scale bars in the journal's guidelines for figure preparation/guidelines for authors. Many of these requirements could be listed to the guidelines that many journals already have online. It is much easier to have these requirements up front instead of trying to fix them during the manuscript reviewing stage.

Page 23. "of which 500 are estimated to contain images" - do the authors mean photographic images? What is this number based on?

Figure 1B and Figure 1C layout could be more similar to each other

Figure 1C - right hand panel not described in Results, and not clear how it differs from what is shown in the left panel

In Figure 4, Square = 1cm, should this be 1cm2?

Figure 4 refers to 1-3 and 4-6 but there are no numbers in the figure itself.

Figure 4 typo: "Micropcope"

Figure 12: In top right, I did not think the color annotation was that clear ; I liked the solution used in the top left, although that is not color blind safe - could something similar be used in the top right? The line to the mRNA appears to land in an area that has both colors, which was not very clear. Maybe moving it a bit to the left so that it would land in a clear green area would help.

Methods. Page 25, under "Screening" what is meant by "using Rayyan software"? I was not familiar with that tool.

Supplemental materials. The Plant Cell articles were included twice in Tables S2 and S3, which was potentially confusing, since now the totals of Tables S1-S3 cannot be summed. I would recommend leaving them out of the Cell Biology table (S3), with a little note under the table, so that there are no duplicate values across the tables.

Table S1-S3: maybe include percentages in the top row, e.g., n=409 n=159 (38.9%)

Page 29, under Table S2, should be "This journal was also included on the cell biology list (Table S3)." instead of "(Table S2)".

Reviewer #2:

In general, I find this paper to be excellent and to be potentially a very valuable resource to the community. I appreciate the large amount of work their initial quantitative findings must have required, and the thoroughness of the recommendations they have put together.

My largest critique (the only one I feel would be NECESSARY to address before publication is that in general), the authors prescribe certain things readers should do when authoring their own papers, but are inconsistent in whether or not they tell readers how to do that (or point them to an educational resource). This is not universal- they do, for example, point the reader to resources for simulating colorblindness in the text around Figures 7 and 8, but not how to do the inversions or greyscale testing in Figure 6, how to generate labels ala Figures 10 and 11, etc. Obviously it would be outside the scope of this paper to teach readers to do every task in every POSSIBLE software it could be done in, but the authors could select one or two commonly used tools (such as FIJI, Photoshop/Illustrator, etc, though for maximum utility my vote would be for something free to use) and provide guidance in those. This could be done along the way, and/or as part of a section at the beginning describing what are some commonly used tools for figure creation (and pointing to resources for each to learn to do common tasks). In that vein it would also be nice for the authors to more fully credit the tools that were used to make their own figures (they describe which python libraries are used in the creation of their bar graphs, but don't cite the relevant publications for those libraries or for the jupyter project itself (which according to the OSF project is how those figures were created), nor do they describe which software tool(s) they used to create the rest of the figures (They mention the QuickFigures tool at one point, though it's not clear that is what's used in this work or not).

An additional few smaller critiques-

1) The degree to which the authors obey their own rules for best practices vary; many of the images in the paper lack scale bars, for example, or have illegible bars (figure 6). I understand in most cases that is not the point being illustrated in that particular figure, and would not see it as a blocker for publication, but it would be nice to see them used more consistently, especially in the "good" images.

2) The text in the table in Figure 10 is VERY small, it might be better to move it below rather than beside the figure so it can more easily be enlarged. The text in other figures (such as 9 and 11 is also borderline tiny)

3) I personally find the broken-up-bar-graph in figure 1B a bit hard to read, especially as the bars for "Some scale bar dimensions" and "All/some magnification in legend" are overlapping; breaking it into multiple bar plots ala 1A lacks the "nice" effect of seeing how things add to 100 but might be more clear.

Reviewer #3:

The manuscript starts with quantification of image usage in publications and is followed by quantification of correct/incorrect image reporting (usage of scale information, insets etc.). The analysis of the published papers served the authors to discover problems and to come-up with suggestions that are presented in the following - core part of the manuscript. Here the authors give clear suggestions to relevant steps of image representation and figure preparation. Each step is visualized comparing wrong and right/improved approaches, such that the readers can compare the differences immediately by themselves. The manuscript ends with a final discussion that includes action points suggested to journal and the scientific community. The manuscript is very clearly written and gives the reader clear recommendations on how to improve image display.

Novelty and significance

While the single steps addressed (scale bar, color scheme, annotations) are not novel, the way of presenting it with the comparison in figures and the focus on the "colorblind safe" images is. The discussion in context of modern publishing (online) and the connection to online image repositories is timely.

The manuscript gives the reader a very clear "workflow" of what to do in different cases (e.g. 2 color image vs. 3 color image, or EM image vs. color photo) in order to avoid pitfalls. With this I expect it to be of great use, especially (but not only) for early career scientists.

Points of criticism:

I would have wished for a discussion around the flexibility of the rules and a potential of "miscounting" in the quantification of fig 1. E.g. also in this manuscript the scale bar is missing in most figures and would have been counted accordingly as "Partial scale information" in figure 1. (The reason why the scale bar is missing is written in the text of the manuscript.)

Also, I would have wished for a discussion whether/whether not it is important to include details in the figure legend, especially about tissue specification. Under section 7 (prepare figure legends) it is written that some journals require details, while others not - which clearly shows different opinions about this topic. Figure 2B "Are species/tissue/object clearly described in the legend?" shows to me rather different opinions on this topic rather than clear errors in image representation.

Minor comments:

- Fig 1: Include to the supplementary examples of images classified as e.g. "insets inaccurately marked, some marked " etc. if this is possible following copyright of already published figures.

- Fig 3A, subcellular scale image is saturated

- Fig 3B. Solution (cell image): inset marking is not fully transparent

- Fig 4: Ruler as scale bar - Square: 1cm; square not visible in this magnification

- Fig. 5: "Darth Vader being attached" - kids playing Star Wars?

- Section 5. Design the figure: "either from top to bottom and/or from right to left" should presumably read as "left to right"

- Fig 6 scale bar not visible in the print as it is for now

- Fig 8 Split the color channel: blue described as "least visible" in Fig. 6, but used anyway

- Same in Fig. 12 (red), described as "least visible" in Fig. 6, but used anyway

Reviewer #4:

[identifies herself as Perrine Paul-Gilloteaux]

This paper proposes a systematic review of figures in literature in biology-related fields, following some of the PRISMA guidelines, to assess the quality of these published figures. The criteria assessed are the accessibility of figures for color-blindness scientist, the presence of some minimal information as defined by the authors in the legend, the clarity of annotations or insets as assessed by the authors, the presence and clarity of the scale bar. The minimal information (in addition to the scale bar) that should be reported in the legend, as defined by the authors, are defined as the species (or cell lines) observed and the explanation of colors shown. Statistics on the binary fulfilment of these criteria are reported on the selected sample of publications.

The main message reported is that a majority of figures manually inspected by the authors did not fulfil all these criteria.

In addition the authors provides some examples of DO and DON'T for these points and provide guidelines to design good quality figures, according to these criteria.

While the study is certainly a considerable amount of work, and may point out that editors and reviewers did not do their job (PLOS Biology was not assessed) (reporting scale bar is at least known and required to be present and all figures by editors), I am questioning the choice of the criteria assessed. In particular, authors stated that these criteria serve the reproducibility, I do not understand how badly presented insets may reduce the reproducibility, as stated by the authors. It may unserve the readability, or send a bad message of the rigour of the study, but even this would need to be supported as statements, since in the study the figures which were not filling these criteria did not need them to be understood by the reader. More important guidelines, such as the one asked by journal publishing guidelines (contrast settings, background filtering, process history) would be more important as they can lead to wrong and false messages. The choice of these particular criteria should have been defended against some data or example about how they prevent reproducibility.

Then, showing with the permission of editors/authors, some example of badly assessed figures would have been useful: in particular I am doubtful about the unvisible annotation due to the blending with background color and how it can escape, the example shown of DON'T would serve better the message if taken from real published papers. Real example from real papers of figures assessed as not filling some of the criteria would serve better the message of the paper. Or even more ambitious, adding some reporting on the subjective loss of information and understanding in these papers by the authors of this meta analysis?

For example, even if it is indeed not deserving the main message of the paper, scale bar is not reported in most of the figures of this paper itself (it would have been expected at least for the example of different scale of images Figure 3 ) and in the same time species is reported for all figures when it brings no element to the main message, which is not biologically-related.

Also in the reporting of the method, I could not get how was defined the error rate mentioned: discrepancy in the binary answer of reviewers on each criteria? Are the scripts to compute the statistics provided? I could not find it on the link provided by the authors.

In addition, one of the main conclusion is also that these recommendations could help in designing the minimal information required when depositing data, but actually the repositories mentioned (IDR, Cell Atlas) store the raw data, not the figures, so the criteria and factors assessed are not applicable. Could the authors comment on this point or clarify this?

In conclusion, while the topic is timely relevant in the time of the reproducibility crisis, the authors are sending some messages that should be in the hand of the editors while editing the final proof of papers, in particular with the limited amount and impact of the criteria assessed. The two parts of the paper: constatation of the state of figure published in April 2018 against the criteria defined by the authors, followed by related guidelines and recommendations, are coherent together but the angle taken is too narrow:, in particular when stating as a main mission the reproducibility of papers. It may be of relevance for teaching courses but I am not sure about its categorization as a research paper as it is. The meta analysis could be of further interest if the support of the message was stronger by proving how this failure in criteria deserves reproducibility and interpretation of the data, as I am not convinced the ones chosen are the more important.

Reviewer #5:

[identifies himself as Simon F. Nørrelykke]

* Summary of the research and my overall impression

** 1. summarise what the ms claims to report

This manuscript details the results from a group of researchers across the globe who got together to document the state of image-based figures in scientific publications. The results obtained show that there is ample room for improvement and the authors proceed by giving figure-creation recommendations that, if followed by authors and journals, should greatly increase the quality of published figures.

Fraudulent image manipulation and how to acquire images is not the focus of this manuscript. Microscopy images, both transmitted, fluorescent, and electron, as well and photographs, are the focus; medical images (MRI, ultrasound, etc) were allowed but rare in the three fields studied.

All papers published during April 2018 in 15 journals (45 journals in total) in the three fields of plant science, cell biology, and physiology were manually examined and scored along several dimensions according to a shared protocol, available online and discussed in the manuscript.

580 papers were examined by "eLife Community Ambassadors from around the world" working together.

Only 2--16% of these papers met all the criteria set for good practices.

Detailed recommendations are given for the preparation of figures with microscopy images. These include discussions of scale bars, insets, colors/colorblindness, label, annotations, legends etc.

Though figures are ideally be designed to reach a wide audience, incl. scientists in other fields, they are typically only interpretable by a very narrow one, if at all.

The advise given on selecting the relevant magnification, how and where to include scale bars, and usage of color, should all be common sense, but apparently is not (behold the results of the investigation reported in this manuscript.) They are thus valuable, even if not novel or thought-provoking, and should be mandatory reading for every student preparing their first manuscript - and perhaps for a majority of PIs, reviewers, and editors alike.

** 2. give overview of the strengths and weaknesses of the ms

- Well written manuscript that reads well (except, perhaps, for the results section)

- The results section is very dry. Six paragraphs lists a large number of percentages. This is data but almost not information. An actuarian may disagree. Figures contain slightly more data and in a more digestible format (graphical).

- Data-acquisition: The number of journals assessed and the approach taken (two reviewers per paper and a clear protocol) is scientific and convincing

- The recommendations are clear and well illustrated

- Though most/all of the points are not new to anyone used to working with images (colorblindness, contrast, scale bars etc), it is useful to see them all collected and commented on in one place - also, every number of years it is useful to remind the community that these things are still (or increasingly? we don't know) an issue.

- Being literal about PLOS criteria:

+ Originality :: this is, as far as I know, the first papers reporting solidly on image-based figure quality

+ Importance to researchers in its field :: Important enough that it should be mandatory reading for any figure-creating scientist

+ Interest to scientists outside the field :: The findings and recommendations cover three fields and easily generalise to other fields

+ Rigorous methodology and substantial evidence for its conclusions :: Yes! Details given elsewhere in report.

** 3. recommended course of action

Publish after revision.

Highlight with editorial mention and Twitter activity.

This paper may do more for science than many a pure research manuscript.

* Specific areas that could be improved

** Major issues

- Major, somewhat, because pointing to conceptual issues

+ p. 6 "We evaluated only images in which the authors could have adjusted the image colors (i.e. fluorescence microscopy)"

+ Unless I misunderstand, it is perfectly possible to adjust the colors in any image, so this limitation to fluorescent microscopy images seems to not be justified by the argument given.

+ Example: In an RGB image, e.g. a photo of a flower, the user can set a different color for each of the three channels. This is easily done in, e.g. Imagej/Fiji using the channel tool

* https://imagej.net/docs/guide/146-28.html#toc-Subsection-28.5

* https://imagej.net/docs/guide/146-28.html#sub:Channels...[Z]

+ Fix: redo research or reformulate sentence to simply state which images you comment on.

+ Or, did you perhaps mean "e.g." and not "i.e."?

- Major, but fixable, because pointing to conceptual issues

+ p. 12: "Digital microscope setups capture each channel's intensities in greyscale values."

+ Nope: Some do, some don't.

+ Fluorescent microscopes equipped with filter cubes and very light sensitive CCDs (CMOSs) tend to, as do confocals

+ Slides scanners (also microscopes) are usually equipped with RGB cameras.

+ Suggested fix: delete sentence after understanding why it is wrong

- Suggestion for how to lead by example and in the interest of reproducibility

+ Share the data in an interoperable manner (FAIR principles)

+ Share the Python notebooks used for statistical analysis

+ Share the scripts used to create figures (unless assembled by hand)

+ Do this in GitHub, Zenodo, or the journal website

** Minor issues

- p3: EMBO's Source Data tool (RRID:SCR_015018)

+ Is this supposed to be a link or reference?

- p6: "Color Oracle (https://colororacle.org/, RRID:SCR_018400)."

+ What is RRID? Not explained until p. 23.

- p. 5, Figure 1

+ Please give n in subpanel B, similar to A and C, or Fig 2 A, B, C.

+ Or state that numbers are the same as in A

- p. 11, Figure 4

+ This figure would be more powerful if the problems were 1-1 mirrored by solutions

+ Only two of the five problem images are solved

+ The ruler shown in the bottom right corner is too small to illustrate the point otherwise made: Zooming in, in the pdf, does not give clearly resolved 1cm squares, perhaps due to jpg effect.

+ Alternatively, rename from "problem" and "solution" to something not evoking expectations of solutions to the problems, e.g. by removing those two words.

- p. 12, Figure 5, top row

+ This is a very unlikely example of a scientific image

+ Resist temptation of including photos of family members ;-)

+ If you cannot find a natural, scientific, example, perhaps this is not an actual problem?

- p. 12, Figure 5, third and fourth row

+ Recommendations: the splitting should be in addition to, not instead of, adjusting for colorblindness in a merged image

+ Yes, you refer to Fig 8, but here is a natural place to mention it

- p. 13, Figure 6

+ This figure ought to be redundant, to the extent that the reader knows that higher contrast has higher contrast

+ If, however, the authors saw many examples of dark colors on dark background during their scans of papers, this could still seem a justified figure

- p. 14

+ "Free tools, such as Color Oracle (RRID:SCR_018400)"

+ Also available, for images, in the very popular open source software Fiji under "Image > Color > Simulate Color Blindness"

- p. 15, Figure 8

+ You show possible solutions but do not say what you recommend.

+ Please, do that and argue for the choice!

- p. 16

+ "QuickFigures (RRID:SCR019082)"

+ Does this software support reproducibility (creates scripts that can generate entire figure)?

+ Please comment in manuscript

- p. 17, Figure 10

+ Text in right half of figure is too small to comfortably read

- p. 21 Figure 13

+ Add title to third column

- p. 23

+ "increase training opportunities in bioimaging"

+ Should, likely, read "increase training opportunities in bioimage analysis"

- p. 35, Figure S1

+ Please create higher quality figure that better supports zooming in

- Suggestion

+ Cite first author's recent paper in F1000R-NEUBIAS on same topic

Decision Letter 2

Roland G Roberts

26 Feb 2021

Dear Tracey,

I've obtained advice from two of the previous reviewers, and on behalf of my colleagues and the Academic Editor, Jason Swedlow, I'm pleased to say that we can in principle offer to publish your Meta-Research Article "Creating Clear and Informative Image-based Figures for Scientific Publications" in PLOS Biology, provided you address any remaining formatting and reporting issues. These will be detailed in an email that will follow this letter and that you will usually receive within 2-3 business days, during which time no action is required from you. Please note that we will not be able to formally accept your manuscript and schedule it for publication until you have made the required changes.

Please take a minute to log into Editorial Manager at http://www.editorialmanager.com/pbiology/, click the "Update My Information" link at the top of the page, and update your user information to ensure an efficient production process.

PRESS: We frequently collaborate with press offices. If your institution or institutions have a press office, please notify them about your upcoming paper at this point, to enable them to help maximise its impact. If the press office is planning to promote your findings, we would be grateful if they could coordinate with biologypress@plos.org. If you have not yet opted out of the early version process, we ask that you notify us immediately of any press plans so that we may do so on your behalf.

We also ask that you take this opportunity to read our Embargo Policy regarding the discussion, promotion and media coverage of work that is yet to be published by PLOS. As your manuscript is not yet published, it is bound by the conditions of our Embargo Policy. Please be aware that this policy is in place both to ensure that any press coverage of your article is fully substantiated and to provide a direct link between such coverage and the published work. For full details of our Embargo Policy, please visit http://www.plos.org/about/media-inquiries/embargo-policy/.

Thank you again for supporting Open Access publishing. We look forward to publishing your paper in PLOS Biology. 

Best wishes,

Roli

Roland G Roberts, PhD 

Senior Editor 

PLOS Biology

_______________

REVIEWERS' COMMENTS:

Reviewer #1:

[identifies herself as Elisabeth M Bik]

I thank the authors for addressing all of the comments raised by the reviewers. I look forward to see this paper published.

Reviewer #2:

[identifies herself as Beth Cimini]

The authors have satisfied my concerns and I can happily recommend this work for publication.

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Fig. Flow chart of study screening and selection process.

    This flow chart illustrates the number of included and excluded journals or articles, along with reasons for exclusion, at each stage of the study.

    (JPG)

    S1 Table. Number of articles examined by journal in physiology.

    Values are n, or n (% of all articles). Screening was performed to exclude articles that were not full-length original research articles (e.g., reviews, editorials, perspectives, commentaries, letters to the editor, short communications, etc.), were not published in April 2018, or did not include eligible images. AJP, American Journal of Physiology.

    (DOCX)

    S2 Table. Number of articles examined by journal in plant science.

    Values are n, or n (% of all articles). Screening was performed to exclude articles that were not full-length original research articles (e.g., reviews, editorials, perspectives, commentaries, letters to the editor, short communications, etc.), were not published in April 2018, or did not include eligible images. *This journal was also included on the cell biology list (Table S3). **No articles from the Journal of Ecology were screened as the journal did not publish an April 2018 issue.

    (DOCX)

    S3 Table. Number of articles examined by journal in cell biology.

    Values are n, or n (% of all articles). Screening was performed to exclude articles that were not full-length original research articles (e.g., reviews, editorials, perspectives, commentaries, letters to the editor, short communications, etc.), were not published in April 2018, or did not include eligible images. *This journal was also included on the plant science list (Table S2).

    (DOCX)

    S4 Table. Scale information in papers.

    Values are percent of papers.

    (DOCX)

    Attachment

    Submitted filename: Response_to_reviewers_R1_20200126.docx

    Data Availability Statement

    The authors confirm that all data underlying the findings are fully available without restriction. The abstraction protocol, data, code and slides for teaching are available on an OSF repository (https://doi.org/10.17605/OSF.IO/B5296).


    Articles from PLoS Biology are provided here courtesy of PLOS

    RESOURCES