Abstract
Effective visual communication is a core competency for pharmacometricians, statisticians, and, more generally, any quantitative scientist. It is essential in every step of a quantitative workflow, from scoping to execution and communicating results and conclusions. With this competency, we can better understand data and influence decisions toward appropriate actions. Without it, we can fool ourselves and others and pave the way to wrong conclusions and actions. The goal of this tutorial is to convey this competency. We posit three laws of effective visual communication for the quantitative scientist: have a clear purpose, show the data clearly, and make the message obvious. A concise “Cheat Sheet,” available on https://graphicsprinciples.github.io, distills more granular recommendations for everyday practical use. Finally, these laws and recommendations are illustrated in four case studies.
The goal of quantitative science is to facilitate informed decisions and actions through a data‐driven understanding of complex scientific questions. It is the role of any quantitative scientist (pharmacometrician, statistician, econometrician, etc.) to support this goal through (i) appropriate quantitative methods (experimental design, statistical models, etc.) and (ii) effective communication of results. There should be no conflict between technical and communicative skills—on the contrary, both of these aspects work in concert; either one without the other is not sufficient. Often, however, scientists focus on the former and neglect the latter, and sophisticated investigations remain without impact.1 The goal of this tutorial is to help close this gap.
Scientific influence relies on effective communication,2 and visual communication is one of the most effective channels of communication. Effective visual communication means using the visual channel to deliver the right information or messages clearly and concisely. Quoting Chambers et al.,3 “there is no single statistical tool that is as powerful as a well‐chosen graph.” Indeed, effective visual communication is a core competency for the quantitative scientist: he or she must not only “get the question right” (understand contextual subject matter) and “get the methods right” (technical expertise) but also “get the message right.”
Visualization and the use of graphics can help at every stage of a quantitative workflow, from the very first data explorations to the final communication of conclusions and recommendations. For example, we often switch across different modes of working, from learning to confirming4 or from the “subjective” to the “objective” (see also Gelman and Hennig5). Effective visualization can help in all of these modes; see, for example, Gabry et al.6 for a demonstration in a Bayesian workflow. In drug development, visual communication is required at all stages, from designing, analyzing, and reporting clinical trials to communicating results and supporting subsequent decision making. The role of the quantitative scientist in this process is to ensure that relevant information (concepts, assumptions, patterns, trends, signals, and conclusions) is clearly described and easy to interpret. For this, we must understand the laws and principles of effective visual communication, such as the grammar of a (visual) language (see also Wilkinson7).
If we get this right, we will be more successful in the scoping (e.g., visually clarifying the research question), execution (e.g., finding patterns in data), and communication (e.g., displaying results and conclusions) of our work. We can find important information in complex data and help people understand it. We can positively influence decisions and actions. We can create trust, partnership, and engagement in cross‐functional teams and audiences. We can increase our personal effectiveness. However, if we fail, we can fool ourselves and others. We can fail to see patterns in inappropriate displays. We can confuse teams, detract from the message, and pave the way to wrong conclusions and decisions.8
Much work has been done on this theme. Tukey,9, 10 Tufte,11 and Cleveland12 laid the foundations for good quantitative graphics. Amit et al.,13 Bradstreet et al.,14 Cabanski et al.,8 Donahue,15 Doumont,16, 17 Duke et al.,18 Few,19 Gordon and Finch,1 Krause and O'Connell,20 Matange and Heath,21 Nolan and Perrett,22 Nussbaumer Knaflic,23 Robbins,24 Vandemeulebroecke et al.,25 Wainer,26 Wong,27 and Wong28 have all fostered an intelligent and impactful use of visual communication and graphics. Collaborative initiatives such as CTSpedia29 have emerged. Importantly, the theme extends beyond technical “tips and tricks” for good graphics. More fundamentally, it includes the focus on the right purpose, scientific question, situation, and audience. With this in mind, it is the goal of this tutorial to distill and convey the main principles of effective visual communication for the quantitative scientist in simple, useful, and actionable terms.
The remainder of this tutorial is structured as follows. In the next section, we posit three laws of visual communication for the quantitative scientist. Second, we provide more granular recommendations for good visual display, conveniently compiled in a single‐page reference sheet. Next, four complete use cases are given to illustrate the application of these laws and recommendations in practice. Finally, we close with a discussion. Although most of the content is inspired by our work in pharmaceutical development, the same principles apply in any quantitative science.
THE THREE LAWS OF VISUAL COMMUNICATION FOR THE QUANTITATIVE SCIENTIST
In quantitative sciences, effective visual communication follows three laws:
Have a clear purpose
Show the data clearly
Make the message obvious
These three laws correspond to the three main ingredients of any quantitative work: purpose, data, and message. Getting these right leads to success; failing in any of them leads to overall failure. In the next subsections, we discuss each of these laws in turn.
Law 1: Have a clear purpose
Why? 30, 31 What is the purpose of this display or that communication? Doumont17 states this as the “zeroth law” of professional communication, “a principle so obvious that it had long been overlooked.” Be clear and explicit about what you want to achieve. Is it to explore some data, to convey an inferential analysis, to deliver a message, convince an audience, or support a decision? It may be a mixture of these—for example, even seemingly simple exploratory plots should serve some (perhaps implicit) decision (e.g., on how to explore further). Every graph, and, more generally, every communication, must be tailored to its specific purpose.
It helps to carve out the scientific question you are trying to address, ideally in discussion with partners, and to write it down explicitly. Try not to look at any data before formulating your question. This is the concept of “question‐based visualizations”:25 let the scientific question determine what data to display and how. (For example, combine data from different domains if it helps address the question effectively. Do not only produce standard outputs by data domain—a display should be determined by the question it addresses, not by the way the data are organized.) As Diggle32 put it, we “analyze problems, not data.” This does not mean that the question could not be refined after seeing the data. We may well iterate over the problem space and the solution space, as long as we do it consciously. Senn33 illustrates many examples of wrongly framed research questions. A common one is to focus on the wrong comparison, such as comparing a posttreatment value with the corresponding baseline value instead of with the value under a control treatment. Most quantitative graphs display comparisons,34 and it always helps to ask, “compared with what?”11 If the comparison is not clear to the author, it will also not be clear to the reader.
Part of this first law (and of the third; see below) is also to be clear about your audience. Then, to adapt to your audience. Do not assume it will adapt to you. You cannot control your audience, but you can control what information and messages you deliver to it, and how. Is your audience just you (trying to see patterns in data), you in a few years (trying to remember what you did), quantitative experts such as your peers (interested in your methods), subject matter experts (eager for your main message), decision makers (headlines only), or a mixture of these? Your visual communication will need to be different accordingly. Your communication (plot, presentation, and report) is for the audience, not for you.
Clarity on the purpose and the scientific question of interest will help choose appropriate quantitative methods to address them. This, plus clarity on your audience, will help define the key messages and how to deliver them. (On the aspect of delivery, see also Law 3 below.)
Of note, this first law is so important that it may occasionally defy other good principles. If your primary goal is to catch attention, then you may choose an iconic graphical representation that does this well, even if it violates some of the recommendations given further below.35 However, you should never distort the data.
Law 2: Show the data clearly
This is Tufte's11 maxim: “Above all else, show the data.” Show them accurately and clearly. This law has several faces:
Simplify! “Simplify to clarify.”36 It is the prime task of quantitative scientists to make the complex simple: reveal structure in data through models, make inference through analyses, and distill and convey conclusions through (visual) communication. Choose the simplest appropriate graph type; prefer familiar designs over fancy ones (see also the Cleveland‐McGill effectiveness ranking in Law 3 below). Avoid fake dimensions and pie charts. Make your plot “as simple as it can be, but not simpler” (attributed to Albert Einstein; also “Occam's razor” or the law of parsimony). “Understand, edit, and simplify the information and design with your readers in mind.”28 Do not be confused: it is hard to make things simple. This is an iterative process: “edit and revise,”11 and repeat.
Maximize the data‐to‐ink ratio (also “data density index”11) within reason. Maximize the signal over the noise by removing the noise: remove anything that distracts from the purpose of the graph. Nothing is neutral: the choice of symbols or colors, background, fonts, line style and annotations. These elements are noise if they do not serve a clear purpose. Choose them wisely and parsimoniously; make the data stand out. Do not trust defaults in graphical software packages. Often, intelligent use of white space can structure a display better than a lot of ink. (The same holds for tables: these are often most effectively structured by reserving black lines for the horizontal direction and using simple alignment in the vertical direction.) Never clutter your graph with “chart junk.”37
Display the relevant data directly. In a quantitative workflow, this often means to look at the raw data and not just rely on summary statistics. Cabanski et al.8 illustrate this with nine data sets that show completely different patterns despite identical marginal means, SDs, and correlation coefficients (see also Anscombe38 and Matejka and Fitzmaurice39). Ask yourself what is the best way of summarizing the relevant features of the data; it may not be the mean +/‐ SE. When fitting a (statistical, compartmental, mechanistic, etc.) model to the data to draw inference or make predictions, model‐derived quantities may be the “relevant data” to display. In this case, a plot of the raw data may be misleading if it does not account for important covariates. In simulations (e.g., for study planning), this second law may be applied to simulated data and/or derived quantities as appropriate. In a final communication, display concisely what best supports your message (see also Law 3 below).
Figure 1 illustrates some aspects of this second law. Wainer26 has turned this law around, noting that “methods for displaying data badly have been developed for many years, and a wide variety of interesting and inventive schemes have emerged.” He provides 12 highly amusing rules for “how to display data badly,” with striking examples. He then concludes more seriously: “The rules for good display are quite simple. Examine the data carefully enough to know what they have to say, and then let them say it with a minimum of adornment. Do this while following reasonable regularity practices in the depiction of scale, and label clearly and fully.”
Figure 1.

Law 2: show the data clearly. The pie and donut charts in panels (a) and (b) make it difficult to see the order of magnitude of some of the segments. The eye needs to compare areas, bent lengths (of the contour), or angles, – graphical attributes that are not easily decoded. The donut chart even omits the angles. The mosaic plot in panel (c) only relies on areas; again, it is hard to tell the order of magnitude. It is better to use lengths with a common baseline or positions on a common scale, such as in a bar chart or dot plot (see Cleveland‐McGill effectiveness ranking in Law 3). The bar chart in panel (d), however, introduces a fake dimension, which is unnecessary and makes it hard to decode the numerical values from the height of the bars. Panels (e) and (f) are appropriately simple and show the data clearly. They also order the data by magnitude to aid comparison even further. The dot plot in panel (f) uses minimal amount of ink and draws the eye to the position of the dots; it is the most effective way of displaying these data.
Law 3: Make the message obvious
If the second law focused on the data (with the aim to minimize noise), then the third is all about the message (and amplifying the signal). This assumes that you do have a message to tell and that this message is clear at least to yourself. If there is any doubt on this, return to the first law.
The third law mandates to make your message as obvious as possible. Quoting Krzywinski and Cairo,40 “inviting readers to draw their own conclusions is risky.” Do not only make your message easy to get. Make it impossible to miss. This extends beyond graphical elements and involves all aspects of communication.
Clarity on your audience, mentioned already in the first law, is also a prerequisite for the third. It is needed for carving out the message to tell (Law 1) and for adapting its way of delivery (Law 3).
Specific examples for the third law include the following:
Choose wisely how to encode the data you display. Color and area are good for drawing attention, but a viewer can decode positions on a common scale much more easily and accurately. Consider the effectiveness ranking of graphical attributes for encoding numerical values, as proposed by Cleveland 12, 41 and Cleveland and McGill42, 43, 44 (see also Cairo,45 Munzner,46 and Heer and Bostock47). See Figure 4 for a representation of this ranking.
- Exploit preattentive processing as much as possible.19, 48 Some graphical features “jump to the eye,” whereas others require careful inspection. Consider this in your choice of how to encode the data (Cleveland‐McGill effectiveness ranking; see above) and in your choice of symbols, colors, line types, etc. See Figure 2 for an illustration.
Figure 2.
The power of preattentive processing. In the top row, one data point is marked by a different symbol (left) or color (right). Discerning the different symbol requires attention, whereas the different color “jumps out” preattentively. In the bottom row, left panel, we are drawn to a comparison of lengths. In the right panel, we introduce color to draw the eye to the bottom two bars first. Although length is already a strong preattentive attribute, color is even stronger. We can use this to guide the viewers’ attention through a plot and let them follow the story we want to convey. Avoid mental arithmetic. If differences or ratios are the main interest, show them directly. If both raw values and differences are of interest, consider showing both. Use identical axis ranges across multiple plots or panels where appropriate.
- Exploit the principles of visual grouping.49 Graphical entities are most effectively grouped by enclosure, connection, proximity, and similarity (in this order). That is, similar objects are perceived as belonging together, as are objects close to each other, connected by lines, or enclosed in a common subspace. See Figure 3 for an illustration: these mechanisms can provide contextual information to a plot in a simple yet powerful way.
Figure 3.
The principles of visual grouping: enclosure, connection, proximity, and similarity. The first panel shows two groups of points, identified by similarity (of plot symbols). Here, proximity is a consequence of the data display and cannot be chosen deliberately. The second panel introduces color to further enhance the grouping by similarity (using effective redundancy and preattentive processing). The third panel groups the points even more effectively by connecting lines (in addition to the plot symbols and colors). Special attention is drawn to a pair of points by an enclosing ellipse. Minimize the viewer's eye movement. Place elements that are to be compared close to each other. Prefer direct labeling over a legend. See also Figure 5.
Draw the reader's attention to the main points. Use appropriate graphical features (e.g., bold or colored highlighting, reference lines, circling, etc.). Follow up with explicit labeling (e.g., “treatment A outperforms treatment B by X%”).
Add meaningful information to a graph to tell the whole story (e.g., include reference lines, benchmark effects, inferences, etc.).
Use effective redundancy. Convey the same message through multiple channels, to amplify it and give the audience a second chance to get it. Use words and pictures in unison.16 For example, in addition to showing the data, consider annotating the “good” or “bad” axis direction, and state what is seen in plain words. Do not confuse redundancy (pointlessly cluttering the graph) with effective redundancy (conveying a message through multiple complementary channels).
Let every plot stand on its own. Use informative labels and captions, and explain abbreviations. Do not require the reader to search through text in order to understand a figure.
Always add a title to your plot. Consider phrasing it as a conclusion, not a description (e.g., “plasma concentration depends on body weight” rather than “plot of plasma concentration vs. body weight”). Think of your audience when making this choice.
THE GRAPHICS PRINCIPLES CHEAT SHEET
The three laws outlined above provide overarching principled advice and should serve as a guiding star toward effective visual communication for the quantitative scientist. To further ease their implementation in practice, it helps to distill even more detailed recommendations and to illustrate them concretely.
To this end, we have introduced the Graphics Principles Cheat Sheet.25, 50 This single‐page reference sheet is an integral part of this tutorial. It was carefully designed as a concise and accessible resource for everyday practical use. Yet, it draws from a wide range of sources.7, 9, 10, 11, 12, 15, 16, 18, 19, 20, 24, 26, 27, 28, 41, 42, 43, 44, 45, 46, 47 We hope that it proves useful for putting the three laws into practice.
We highlight major parts of the Cheat Sheet in Figures 4, 5, 6. The full version is available on https://graphicsprinciples.github.io, along with corresponding programming code in R.51 Figure 4 exemplifies the conscious choice of the right base graph, an essential step toward any good data display. The same figure also illustrates the Cleveland‐McGill effectiveness ranking of graphical attributes (see Law 3 in the previous section). According to this ranking, numbers are most effectively encoded by position or length, less effectively by angle or area, and least effectively by color hue or volume. Figure 5 provides recommendations for facilitating comparisons and effective use of color. Finally, Figure 6 shows some advice for displaying data more clearly and enhancing legibility and clarity of the narrative.
Figure 4.

Selecting the right base graph; effectiveness ranking. A conscious choice of the most appropriate graph type is an essential step toward any good data display. Typical examples are illustrated here. The Cleveland‐McGill effectiveness ranking of graphical attributes12, 41, 42, 43, 44 posits that numbers are most effectively encoded by position or length, less effectively by angle or area, and least effectively by color hue or volume. Reproduced from Margolskee et al.50 with permission.
Figure 5.

Facilitating comparisons; color for emphasis or distinction. Most quantitative graphs display comparisons.34 Comparisons can be facilitated by the effective use of proximity, by making visual inspections easy, and by reducing mental arithmetic. Color is a powerful stimulus. It is effective for drawing attention and organizing a narrative, but it should be used with caution and restraint. Reproduced from Margolskee et al.50 with permission.
Figure 6.

Implementation considerations; legibility and clarity. Various tips and tricks for displaying data more clearly. Effective graphs stand on their own; they include all necessary elements for the intended narrative. Their implementation, graphical design, and typography support legibility. Reproduced from Margolskee et al.50 with permission. AUC, area under the curve; CL, clearance.
CASE STUDIES
We now apply the laws and recommendations from the previous two sections in four short case studies to illustrate their use in practice. The case studies are inspired by common examples from a pharmacometrician's (or biostatistician's) work. The same principles apply in any quantitative science. Although the case studies resemble realistic scenarios, the data used in generating the graphs are simulated and do not represent any particular drug or trial. Details on how these case studies were generated (with programming code in R) can be found on https://graphicsprinciples.github.io.
Case study 1: Exploratory exposure–response analysis
This case study illustrates the importance of understanding the scientific context when exploring data graphically. An exploratory data analysis is more than just “plotting data”; it can lead to a deeper understanding and inform next steps.6, 52 However, like an analysis that is poorly thought through, a poorly implemented graph can also deceive.
Consider an inhaled drug intended to improve lung function, with the target site of action in the lungs. The drug is also absorbed systemically from the lungs. Suppose that the team wants to fine‐tune the choice of a recommended dose. A typical starting point for this question is often a plot of the response variable of interest against a summary measure of plasma concentration (e.g., the area under the concentration time curve (AUC)). Figure 7 a shows such a plot generated using the default settings of the R package ggplot2.53
Figure 7.

Exploratory exposure–response analysis for an inhaled drug that acts locally in the lung. (a) A scatterplot of response vs. exposure is improved by (b) applying good graphical principles, and fundamentally changed by (c) revisiting the question of interest. AUC0−24, 0–24‐hour area under the concentration‐time curve; FEV1, forced expiratory volume in 1 second
In terms of good graphical principles, this plot leaves much to be desired. Several improvements are warranted, including proper axis scaling, gridlines, annotation, font size, etc. One particularly egregious issue is the lack of care in selecting axis labels, leaving programming labels for the plotted variables (presumably only then to make the effort of explaining them in a caption). An improved version is shown in Figure 7 b, addressing many of these deficits. With an added locally estimated scatterplot smoother,54 we see a positive nonlinear trend, suggesting a shallow sigmoidal exposure–response relationship.
It is tempting, especially when presented with a suboptimal graph, to immediately set about fixing the various graphical imperfections and produce a more appropriate and visually appealing version of the same graph. This is an example of selective attention,55 focusing on the detail but overlooking the higher purpose of the task (i.e., the “why”). Instead, let us now take a step back and revisit this example in the context of the first law of visual communication: have a clear purpose.
Why are we conducting an exposure–response analysis? Recall that the scientific interest is to fine‐tune the dose and that the drug is inhaled and acting locally in the lungs. The implicit assumption of an exposure–response analysis is one of causality. Here, however, plasma concentration is unlikely to be on the causal path from dose to response. What would be a better way to address the scientific question of interest?
Consider Figure 7 c, where instead of estimating an overall trend, we now look at the trends within dose. Clearly, any apparent trends within dose do not follow a consistent pattern across doses. The only reason why exposure and response seemed associated in the previous two plots is that they share a common cause, namely, dose. In other words, dose is a confounder in those plots, and, in this case study, systemic concentration is not a better predictor of response than dose. We should build a dose–response model, rather than an exposure–response model, and choose a recommended dose based on this (and any information on safety and tolerability).
Case study 2: Pharmacokinetic exposure by ethnicity
This case study is inspired by a publication56 comparing the pharmacokinetic (PK) exposure of a drug in Caucasian vs. Japanese subjects. The same single oral dose was administered to all subjects. The original graph displays mean ± SDs of the plasma concentration over time, grouped by ethnicity (Caucasian or Japanese), as in Figure 8 a.
Figure 8.

Pharmacokinetic exposure by ethnicity. (a)A graph that looks fair at first sight reveals important information after (b) two simple changes, namely, scaling the y‐axis differently and plotting confidence intervals instead of SDs. Key messages are made more obvious by directly plotting (c) the pharmacokinetic parameters of interest, or (d) even only their ratios. AUClast, area under the concentration‐time curve from time of administration up to the time of the last quantifiable concentration; CI, confidence interval; Cmax, peak plasma concentration; Ctrough, trough plasma concentration.
This time, let us start with Law 1. What is the purpose of this graph? For drugs that are mainly developed in a Caucasian population, Japanese drug regulation requires sponsors to investigate whether the PKs are similar or different between Caucasian and Japanese populations. The purpose of this graph is to help address this question. Looking at it, we may be tempted to say that the PKs are reasonably similar. But are they really? If they are not, then in what way?
This leads us to Law 2: show the data clearly. The graphical attributes in Figure 8 a seem to be wisely chosen: the symbols and labels are clear, the gridlines are supportive and stay in the background, and there is no unnecessary adornment. However, at least two things obscure the answer to our question of interest. First, the concentrations are plotted on a linear scale, which makes it difficult to distinguish them at the lower end of the profile. Concentrations should be plotted on a logarithmic scale because they are log‐normally distributed. In fact, for concentrations in particular (but not generally for any log‐normally distributed data), we should produce both types of display: one on a log‐linear scale (to assess the elimination phase) and one on a linear scale (to see the peak more clearly). Second, it is hard to determine whether any differences are significant when SDs are plotted instead of SEs or confidence intervals. SDs show the variation in the data; they do not shrink when more data are collected. SEs show the variation in the means. Confidence intervals may be the best choice as they directly show the uncertainty about the means.
These issues have been fixed in Figure 8 b (we only show the log‐linear version). To reduce cluttering, the ticks at the end of the whiskers have also been omitted (nondata ink). If the graph displayed more than two profiles, we might consider replacing the whiskers by (shaded) confidence bands or separating the graph out in panels or “small multiples” (see bottom of the backside of the Cheat Sheet on https://graphicsprinciples.github.io). From Figure 8 b, it seems that the higher concentrations are not meaningfully different, but the elimination phase does differ between the two ethnicities. This could also translate to different average exposures. That is, based on two simple changes in the plot, we now see answers emerging for our initial question about PK differences. We see them emerging with respect to three key PK characteristics: peak, elimination/trough, and overall exposure. Although (depending on the drug) similarity in the peak may be reassuring from a safety point of view, a lower overall exposure in Japanese subjects could be a concern for efficacy.
Moving on to Law 3, let us now make the message obvious. We could choose a completely different graph type to strengthen the message. Figure 8 c shows the three (noncompartmentally derived) quantities peak plasma concentration (Cmax), trough plasma concentration (Ctrough), and AUC from time of administration up to the time of the last quantifiable concentration (AUClast) with 95% confidence intervals by ethnicity. Clearly, Ctrough and AUClast are different between the two ethnicities, but Cmax is not. We could derive these quantities from a compartmental model fit and produce the same plot, or we could go one step further and show directly their geometric mean ratio, Japanese vs. white subjects, as in Figure 8 d. This last plot answers the initial question most succinctly, and its graphical appearance has also been further simplified (no frame, minimal gridlines, and mildly highlighted line of equality), to not distract from the message. In practice, Figures 8 b,c and/or 8 d together may be most informative, covering the time course as well as differences in key parameters.
Case study 3: Improving a “waterfall plot”
This case study illustrates the importance of aligning a graph with the scientific question it should address, the option of filtering signals through a model, and finally the display of a scientific answer in a condensed messaging graph.
Consider a small early development trial, randomized and placebo controlled (2:1 randomization), with a continuous primary end point measured at baseline and longitudinally over a period of 4 weeks. Lower outcome values are better, and there are no dropouts and no missing data. Suppose that the team is interested in the effect of the drug at the last relevant measurement timepoint, as is often the case. A common approach in early development trials is to simply plot the observed change scores in a so‐called “waterfall plot,” such as Figure 9 a.
Figure 9.

Improving a “waterfall plot.” A waterfall plot focusing on (a) the last observed timepoint, common in early development, is improved by (b) a side‐by‐side boxplot, which shows the treatment comparison more clearly. (c) Showing the totality of the data during the first 4 weeks facilitates an even richer understanding, including a suggested linear trend of the treatment effect over time. This may justify the fit of (d) a linear model and ultimately (optionally) (e) a condensed messaging graph on the treatment effect based on this fit.
To probe Law 1, what is the question addressed by this plot? It asks about the treatment effect after 4 weeks of treatment. Is this the right question? Let us assume for a moment that it is. Then a waterfall plot is not ideal for addressing it.57 Small treatment effects are difficult to discern, especially with an unbalanced randomization ratio. The audience must observe the distribution of color across the entire plot just to determine which treatment is more beneficial; this can become even more difficult with a larger sample size or more than two treatment groups. In Figure 9 a, one might see a treatment benefit, but how large is it and how certain of it are we? The popularity of waterfall plots is a mystery.
If we insist on week 4 as the only timepoint of interest, we could present overlaid density plots or side‐by‐side boxplots for a better appreciation of the difference in distribution between the two treatment arms. Figure 9 b shows an example with the raw data points included, which is a much better alternative to Figure 9 a. The side‐by‐side placement facilitates the treatment comparison, and the plot is simple and familiar and uses minimal ink for what it shows. Graphical attributes (colors, font size, etc.) are easily readable.
However, with such rich longitudinal data, it may be more informative to ask the question about the treatment effect during—not just after—the first 4 weeks of treatment. This is especially relevant in the early, more exploratory development phase (and it would be even more relevant if there were dropouts). As a rule, the recommended first step is to visualize the totality of the data. Figure 9 c does this and includes means by treatment and timepoint. We see large interindividual variability and overlap between the treatment groups. We also start to get an appreciation for the time course of a mean effect. We see linear trajectories of the means over time, with the active arm seeming to improve and the placebo arm staying fairly constant. We cannot exclude that the apparent gap might continue to increase beyond 4 weeks of treatment. This plot, while doing little more than displaying the raw data, is already worth sharing with the project team. It facilitates a much richer understanding of the data than the previous two plots. It shows the data clearly (Law 2).
Depending on the goal of the analysis, we could stop here. However, if we want to quantify the treatment difference while adjusting for important covariates, we should proceed with a statistical model. Based on Figure 9 c, a linear model seems appropriate. We fit a linear model with treatment, patient‐specific intercept, and slope, and we now also adjust for the baseline value of the primary end point and for any other important covariates. We can then visualize the data filtered through this model, omitting the raw data but displaying longitudinal point estimates of the means and some form of uncertainty intervals for both treatment groups (Figure 9 d). This gets closer to the nature of a messaging graph, focusing directly on the results of our model. Optionally—and depending on the audience!—we could even go one final step further and display the treatment difference directly, as in Figure 9 e. This plot addresses the question about the treatment effect over time without requiring any mental arithmetic. We can read off approximate estimates for the treatment effect, and the level of confidence is easily appreciable from the added credible band (which does include zero!). Appropriate and parsimonious annotations make the message even more obvious (Law 3), also through “effectively redundant” information (stating what can be seen).
It is worth emphasizing that this last plot should not be the only one generated, and probably not the only one shown either. Strongly reduced messaging graphs require a robust understanding of the underlying data, which can only be built through a workflow, such as the one described above. Further, depending on the situation and the audience, they might be challenged as loaded or unscientific. (For example, the apparently perfect linear trend in Figure 9 e appears “unrealistic.”) It is therefore important to ensure and emphasize that this last plot derives from a model that (as every model) is intended to separate the signal from the noise and that the choice of this model is justified by a thorough inspection of the data.
Case study 4: Post hoc subgroup analysis
Post hoc exploratory subgroup analyses are common, especially after borderline or failed clinical trials.58 Often, the objective of such analyses is to understand why the study failed or to identify a subgroup of patients who did show sufficient response to the treatment.59, 60 In this fourth case study, we illustrate the challenge of navigating this type of analysis. The objective is to present to decision makers a recommendation whether to proceed with further investigation of a genetic marker that may be predictive of response to treatment.
Figure 10 a displays the desired effect, percentage body weight loss from baseline, for an active treatment and a placebo control arm. The primary end point is 12 weeks after randomization. Although the active treatment reduced body weight, the average extent of the effect was not considered clinically meaningful. However, the team found a subgroup of patients identified by a biologically plausible genetic marker who seemed to benefit more strongly from the compound, as shown in Figure 10 b.
Figure 10.

Post hoc subgroup analysis. A project team saw (a) an overall insufficient treatment effect regarding weight loss but (b) a seemingly stronger benefit in a subgroup of patients identified by a biologically plausible genetic marker. (c) Getting the comparison right, however, it becomes clear that the marker, although prognostic for weight loss in general, is not predictive for a treatment effect vs. placebo. (d) This is made more obvious by displaying the treatment difference directly. (e) Finally, we can craft a messaging slide suited for conveying this message to a group of executives within seconds. CI, confidence interval.
At this point, we need to intervene. The first plot did not show the data clearly (Law 2), using barely distinguishable plot symbols and huge ticks at the end of the error bars, too many ticks on the y‐axis, and—the worst offense—visit numbers on the x‐axis (equally spaced!). The second plot, in addition, is fundamentally flawed because it displays the wrong comparison (Law 1, wrong scientific question). The treatment effect in this trial is weight loss under active treatment compared with weight loss under placebo, and this subgroup analysis removes the placebo arm completely.
Let us put the placebo group back in and at the same time improve the various graphical shortcomings of the plot (using appropriately spaced times on the x‐axis, distinguishable plot symbols, etc.). We obtain Figure 10 c, which displays the treatment‐by‐subgroup interaction over time (and clarifies that the error bars are 95% confidence intervals). Interestingly, we see some weight loss also in the biomarker‐positive placebo group and weight gain in the biomarker‐negative placebo group. The treatment effect (active drug vs. placebo) is similar within each biomarker group. In other words, the biomarker seems prognostic for weight loss (at least under the conditions of this clinical trial, which may include dietary or exercise advice) but not predictive for a treatment effect (see also Lipkovich et al.60). Figure 10 d displays the treatment effect directly, to make it obvious that it is the same irrespective of the genetic marker (Law 3).
We have carved out the main point, but we can still improve the analysis, e.g., by performing an analysis of covariance on the absolute scale (with baseline body weight as covariate), instead of displaying the percent change. To complete our objective, we should present these results and our recommendation about the genetic marker to decision makers. We need to show the results clearly and make the message obvious. In this context, the message is that the genetic marker is not predictive for a treatment effect; whether it is prognostic for weight loss is less important. Therefore, we finally craft a messaging slide that is suited for conveying this message in a meeting to a group of executives within seconds.
Figure 10 e shows this final slide, based on the analysis of covariance. Guided by Laws 2 and 3, we arrived at this slide by considering the following principles and recommendations.
Facilitate comparisons. We zoom in on the comparison of interest by appropriately restricting the y‐axis. A reference line at zero indicates the position of no treatment effect, and a labeled arrow shows the direction of treatment benefit.
Reduce cognitive load. We use direct labeling instead of a legend to identify the two subgroups. We select clearly distinguishable colors that also carry through to the subgroup labels.
Adapt to your audience. We avoid writing the y‐axis label vertically, by placing it above the plot and using it as its title. This is to prevent the audience from straining to read information during the meeting. We spell out that this is a treatment comparison (a difference) in terms of weight loss. We also reduce the size of the symbols representing the point estimates, to avoid the audience focusing too much on these and too little on the uncertainty intervals.
Remove nondata ink. We reduce the number of ticks and points displayed on the x‐axis and y‐axis. They do not convey useful information for a short presentation, as we do not expect the audience to extract exact values. The connecting lines are pushed to the background by introducing a degree of transparency.
Effective redundancy. We add a caption that spells out what is seen at week 12 (the primary end point), and we highlight the corresponding treatment effect estimates by an enclosing circle (a powerful grouping technique to draw attention). Finally, we place the main message directly as the title of the slide (setting the most important word, “not,” in bold), and as subtitle we explicitly put our recommendation: “Genetic marker does not warrant further investigation.”
DISCUSSION
Effective visual communication seems obvious, but creating it requires skill. The purpose of this tutorial was to convey some of this skill set. We have proposed three laws of visual communication for the quantitative scientist:
Have a clear purpose
Show the data clearly
Make the message obvious
The first law is about doing the right thing; the two others are about doing things right. We have distilled a range of more detailed recommendations on the Graphics Principles Cheat Sheet,50 a concise reference sheet available on https://graphicsprinciples.github.io. Finally, we have discussed four use cases to illustrate the application of these laws and recommendations in practice.
No system is perfect. First, there is simply no tension‐free one‐size‐fits‐all approach to visual communication: the given laws and recommendations may occasionally cause conflict. Should we show raw data or summaries? More or less detail? Confidence intervals or not? Individual treatment effects or only their difference? These and other choices depend on the purpose and situation of the communication, the audience, and the author. Importantly, the format of a display or presentation may impose practical constraints. (Are colors available? How much time will I have? Etc.) Good judgement is always due, and sometimes a conscious compromise
Second, the distinction between the three given laws may be challenged as occasionally blurred:
We have already mentioned the need to adapt to the audience, which is part of Law 1 (have a clear purpose—with respect to a target audience) as well as Law 3 (make the message obvious—to a specific audience). If your audience does not understand your message, do not repeat it in the same way (and never louder!). Adapt and find other ways to convey it (effective redundancy).
Another such general principle is “do no harm.” In medical ethics and bioethics (“primum non nocere”), it is thought to derive from the oath of Hippocrates. However, it also applies to graphics and visual communication.61 For example, do not get carried away by technology (templates, frames, animation, etc.) that clouds the data (Law 2) or draws attention away from the message (Law 3; see also Baillie's and Vandemeulebroecke's comment62 in the discussion of Bowman63). The removal of “chart junk” (in Law 2) is another example.
Bonate31 mandates a principle that may belong to Law 1 and Law 3: “don't be boring.” If you engage your audience through clarity, visual cues, and perhaps humor, it will be easier for them to pay attention and understand your message. Your visual communication will achieve its purpose. However, if you bore your audience with unnecessary detail, an odd visual pace, by clouding your main point or by not even having one,then it will just turn away.
The Cleveland‐McGill effectiveness ranking, introduced in Law 3, serves Law 2 just as much.
Third, we may have missed some special considerations and applications. For example, we did not say much about an appropriate choice of color, considering human color perception, psychological connotations, color blindness, etc. For further information, we refer to Wong,64 Wilke,65 and resources such as Munsell66 and ColorBrewer.67 We did not discuss any particular software or tool, because the skill set we hope to convey does not depend on technology. In addition, we did not cover interactive graphics, a wide and increasingly popular field. We want to stress the importance of mastering the basic principles, which are at the core of this tutorial. They also form the basis for more advanced applications.
Despite these limitations, we believe that it helps to think of the three proposed laws as major maxims for visual communication for the quantitative scientist. Keeping the Graphics Principles Cheat Sheet “at your fingertips” may provide additional practical support. Finally, our advice is to practice. To quote Tufte11: “Graphical competence demands three quite different skills: the substantive, statistical, and artistic.” These skills cannot be learned by reading an article. Adopt visualization in every part of your workflow; make it a habit. Think graphically. Use pencil and paper before coding in software.68, 69 Calculate and communicate.68 Test and repeat.
Funding
No funding was received for this work.
Conflict of Interest
The authors declared no competing interests for this work.
Supporting information
Graphics Principles Cheat Sheet.
Acknowledgments
We would like to thank David Carr, Julie Jones, Jürgen Löffler, Oliver Sander, Ivan‐Toma Vranesic, and Andrew Wright. We also thank the editor for inviting us to submit this tutorial, and two anonymous reviewers for helpful comments.
[Correction added on September 6, 2019, after first online publication: The word “white” was changed to “Caucasian” throughout the paper.]
References
- 1. Gordon, I. & Finch, S. Statistician heal thyself: have we lost the plot? J. Comput. Graph. Stat. 24, 1210–1229 (2015). [Google Scholar]
- 2. Gibson, E. Leadership in statistics: increasing our value and visibility. Am. Stat. 73, 109–116 (2019). [Google Scholar]
- 3. Chambers, J.M. , Cleveland, W.S. , Kleiner, B. & Tukey, P.A. Graphical Methods for Data Analysis (Wadsworth, Pacific Grove, CA, 1983). [Google Scholar]
- 4. Sheiner, L.B. Learning versus confirming in clinical drug development. Clin. Pharmacol. Ther. 61, 275–291 (1997). [DOI] [PubMed] [Google Scholar]
- 5. Gelman, A. & Hennig, C. Beyond subjective and objective in statistics (with discussion). J. R. Statist. Soc. A 180, 967–1033 (2017). [Google Scholar]
- 6. Gabry, J. , Simpson, D.P. , Vehtari, A. , Betancourt, M. & Gelman, A. Visualization in Bayesian workflow (with discussion). J. R. Statist. Soc. A 182, 389–402 (2019). [Google Scholar]
- 7. Wilkinson, L. The Grammar of Graphics (Springer, Berlin, Germany, 2005). [Google Scholar]
- 8. Cabanski, C. , Gilbert, H. & Mosesova, S. Can graphics tell lies? A tutorial on how to visualize your data. Clin. Transl. Sci. 11, 371–377 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9. Tukey, J. Some graphic and semigraphic displays In Statistical Papers in Honor of George W. Snedecor. (eds. Bancroft T.), 293‐316 (Iowa State University Press, Iowa, 1972). [Google Scholar]
- 10. Tukey, J. Exploratory Data Analysis. (Addison‐Wesley, Reading, MA, 1977). [Google Scholar]
- 11. Tufte, E.R. The Visual Display of Quantitative Information (Graphics Press, Cheshire, CT, 1983(2nd ed. 2001)). [Google Scholar]
- 12. Cleveland, W.S. The Elements of Graphing Data. (Chapman and Hall, New York, NY, 1985). [Google Scholar]
- 13. Amit, O. , Heiberger, R.M. & Lane, P.W. Graphical approaches to the analysis of safety data from clinical trials. Pharm. Stat. 7, 20–35 (2008). [DOI] [PubMed] [Google Scholar]
- 14. Bradstreet, T.E. , Nessly, M.L. & Short, T.H. A course in constructing effective displays of data for pharmaceutical research personnel. Pharm. Stat. 12, 174–184 (2013). [DOI] [PubMed] [Google Scholar]
- 15. Donahue, R.M.J. Fundamental statistical concepts in presenting data: principles for constructing better graphics. Version 2.11. <http://biostat.mc.vanderbilt.edu/wiki/Main/RafeDonahue>. (2011). Accessed March 11, 2019.
- 16. Doumont, J.‐L. Trees, Maps, and Theorems: Effective Communication for Rational Minds. (Principiae, Brussels, Belgium, 2009). [Google Scholar]
- 17. Doumont, J.‐L. The three laws of professional communication. IEEE Trans. Prof. Commun. 45, 291–296 (2002). [Google Scholar]
- 18. Duke, S. , Bancken, F. , Crowe, B. , Soukup, M. , Botsis, T. & Forshee, R. Seeing is believing: good graphic design principles for medical research. Stat. Med. 34, 3040–3059 (2015). [DOI] [PubMed] [Google Scholar]
- 19. Few, S. Show Me The Numbers—Designing Tables and Graphs to Enlighten, 2nd edn (Analytics Press, Burlingame, CA, 2012). [Google Scholar]
- 20. Krause, A. & O'Connell, M. , eds.) A Picture is Worth a Thousand Tables: Graphics in Life Sciences. (Springer, Heidelberg, Germany, 2012). [Google Scholar]
- 21. Matange, S. & Heath, D. Statistical Graphics Procedures by Example. (SAS Institute Inc., Cary, NC, 2011). [Google Scholar]
- 22. Nolan, D. & Perrett, J. Teaching and learning data visualization: ideas and assignments. Am. Stat. 70, 260–269 (2016). [Google Scholar]
- 23. Nussbaumer Knaflic, C. Storytelling with Data. (Wiley, Hoboken, NJ, 2015). [Google Scholar]
- 24. Robbins, N.B. Creating More Effective Graphs. (Chart House, Wayne, NJ, 2013). [Google Scholar]
- 25. Vandemeulebroecke, M. et al How can we make better graphs? An initiative to increase the graphical expertise and productivity of quantitative scientists. Pharm. Stat. 18, 106–114 (2019). [DOI] [PubMed] [Google Scholar]
- 26. Wainer, H. How to display data badly. Am. Stat. 38, 137–147 (1984). [Google Scholar]
- 27. Wong, B. Points of view. Series of columns in Nat. Methods 2010–2015.
- 28. Wong, D.M. The Wall Street Journal Guide to Information Graphics: The Dos and Don'ts of Presenting Data, Facts, and Figures. (WW Norton, New York, NY, 2013). [Google Scholar]
- 29. CTSpedia . < https://www.ctspedia.org>. Accessed March 8, 2019.
- 30. Sinek, S. Start With Why: How Great Leaders Inspire Everyone to Take Action. (Portfolio, New York, 2009). [Google Scholar]
- 31. Bonate, P.L. Be a model communicator and sell your models to anyone(Author, ISBN 978‐0692323816, 2014).
- 32. Diggle, P.J. Analyse problems, not data: one world, one health. Spatial Statistics 28, 4–7 (2018). [Google Scholar]
- 33. Senn, S. Statistical Issues in Drug Development, 2nd edn (John Wiley & Sons, New York, 2008). [Google Scholar]
- 34. Gelman, A. , Pasarica, C. & Dodhia, R. Let's practice what we preach. Am. Stat. 56, 121–130 (2002). [Google Scholar]
- 35. Gelman, A. & Unwin, A. Infovis and statistical graphics: different goals, different looks. J. Comput. Graph. Stat. 22, 2–28 (2011). [Google Scholar]
- 36. Wong, B. Points of view: simplify to clarify. Nat. Methods 8, 611 (2011). [DOI] [PubMed] [Google Scholar]
- 37. Tufte, E.R. Visual Explanations: Images and Quantities, Evidence and Narrative. (Graphics Press, Cheshire, CT, 1997). [Google Scholar]
- 38. Anscombe, F.J. Graphs in statistical analysis. Am. Stat. 27, 17–21 (1973). [Google Scholar]
- 39. Matejka, J. & Fitzmaurice, G. Same stats, different graphs: generating datasets with varied appearance and identical statistics through simulated annealing. In: Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems (CHI 2017) 1290‐1294 (ACM, New York, NY, 2017). [Google Scholar]
- 40. Krzywinski, M. & Cairo, A. Reply to: “Against storytelling of scientific results”. Nat. Methods 10, 1046 (2013). [DOI] [PubMed] [Google Scholar]
- 41. Cleveland, W.S. Graphs in scientific publications. Am. Stat. 38, 261–269 (1984). [Google Scholar]
- 42. Cleveland, W.S. & McGill, R. Graphical perception: theory, experimentation and application to the development of graphical methods. J. Am. Stat. Assoc. 79, 531–554 (1984). [Google Scholar]
- 43. Cleveland, W.S. & McGill, R. Graphical perception: the visual decoding of quantitative information on graphical displays. J. R. Statist. Soc. A 150, 192–229 (1987). [Google Scholar]
- 44. Cleveland, W.S. & McGill, R. Graphical perception and graphical methods for analyzing scientific data. Science 229, 828–833 (1985). [DOI] [PubMed] [Google Scholar]
- 45. Cairo, A. The Truthful Art: Data, Charts, and Maps for Communication. (New Riders Publishing, Thousand Oaks, CA, 2016). [Google Scholar]
- 46. Munzner, T. Visualization Analysis and Design. (CRC Press, Visualization Series, New York, NY, 2014). [Google Scholar]
- 47. Heer, J. & Bostock, M. Crowdsourcing graphical perception: using mechanical Turk to assess visualization design. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI 2010) 203–212 (ACM, New York, NY, 2010). [Google Scholar]
- 48. Ware, C. Information Visualization: Perception for Design, 3rd edn (Morgan Kaufmann Publishers Inc., San Francisco, CA, 2012). [Google Scholar]
- 49. Wong, B. Points of view: Gestalt principles (Part 1). Nat. Methods 7, 863 (2010). [DOI] [PubMed] [Google Scholar]
- 50. Margolskee, A. , Baillie, M. , Magnusson, B. , Jones, J. & Vandemeulebroecke, M. Graphics Principles Cheat Sheet. (Biostatistical Sciences and Pharmacometrics, Novartis Institutes for Biomedical Research, Cambridge, 2018). <https://graphicsprinciples.github.io/> [Google Scholar]
- 51. Development Core Team, R. R: a language and environment for statistical computing. (R Foundation for Statistical Computing, Vienna, Austria, 2008) <http://www.R-project.org>. [Google Scholar]
- 52. Gelman, A. Exploratory data analysis for complex models. J. Comput. Graph. Stat. 13, 755–779 (2004). [Google Scholar]
- 53. Wickham, H. Ggplot2: Elegant Graphics for Data Analysis. (Springer, New York, NY, 2009). [Google Scholar]
- 54. Cleveland, W.S. Robust locally weighted regression and smoothing Scatterplots. J. Am. Stat. Assoc. 74, 829–836 (1979). [Google Scholar]
- 55. Chabris, C.F. & Simons, D. The Invisible Gorilla: And Other Ways Our Intuitions Deceive Us. (Crown, New York, NY, 2010). [Google Scholar]
- 56. Sunkara, G. , Yeh, C. , Ligueros‐Saylan, M. , Kawashita, H. , Koseki, N. & Fukui, Y. Assessment of ethnic differences in the pharmacokinetics and pharmacodynamics of valsartan. J. Bioequiv. Availab. 2, 120–124 (2010). [Google Scholar]
- 57. Mercier, F. , Consalvo, N. , Frey, N. , Phipps, A. & Ribba, B. From waterfall plots to spaghetti plots in early oncology clinical development. Pharm. Stat. (2019) [e‐pub ahead of print]. 10.1002/pst.1944 [DOI] [PubMed] [Google Scholar]
- 58. Wang, R. , Lagakos, S.W. , Ware, J.H. , Hunter, D.J. & Drazen, J.M. Statistics in medicine—reporting of subgroup analyses in clinical trials. N. Engl. J. Med. 357, 2189–2194 (2007). [DOI] [PubMed] [Google Scholar]
- 59. Ruberg, S.J. & Shen, L. Personalized medicine: four perspectives of tailored medicine. Stat. Biopharm. Res. 7, 214–229 (2015). [Google Scholar]
- 60. Lipkovich, I. , Dmitrienko, A. & D'Agostino, R.B. Sr . Tutorial in biostatistics: data‐driven subgroup identification and analysis in clinical trials. Stat. Med. 36, 136–196 (2017). [DOI] [PubMed] [Google Scholar]
- 61. Tufte, E. Envisioning Information. (Graphics Press, Cheshire, CT, 1990). [Google Scholar]
- 62. Baillie, M. & Vandemeulebroecke, M. Discussion on the meeting on ‘Data visualization’. J. R. Statist. Soc. A 182, 419–441 (2019). [Google Scholar]
- 63. Bowman, A.W. Graphics for uncertainty. J. R. Statist. Soc. A 182, 403–418 (2019). [Google Scholar]
- 64. Wong, B. Points of view: color coding. Nat. Methods 7, 573 (2010). [DOI] [PubMed] [Google Scholar]
- 65. Wilke, C.O. Fundamentals of Data Visualization: A Primer on Making Informative and Compelling Figures. (O'Reilly, Sebastopol, CA, 2019). [Google Scholar]
- 66. Munsell, A.H. A Color Notation, 2nd edn (Geo. H. Ellis Co., Boston, MA, 1907). [Google Scholar]
- 67. Brewer, C.A. , Hatchard, G.W. & Harrower, M.A. ColorBrewer in print: a catalog of color schemes for maps, cartography and geographic information. Science 30, 5–32 (2003). [Google Scholar]
- 68. Grant, R. Calculate and communicate. Significance 15, 42–44 (2018). [Google Scholar]
- 69. Wong, B. & Schmidt Kjærgaard, R.S. Points of view: pencil and paper. Nat. Methods 9, 1037 (2012). [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Graphics Principles Cheat Sheet.
