Table 5.
Representative qualitative comments regarding interpretation and clarity of the graph formats
| Positive comments | Negative comments | Other insights | |
|---|---|---|---|
| Comments on line-graph type | |||
| “More” line type | As plain graphs these are quite clear [C01]a | They are somewhat confusing…whether it’s physical or fatigue is in one graph lower and in one graph higher…requires very close attention to detail [C04] | Regarding if higher is better or worse—need consistency [R]b |
| It’s confusing whether lines going up or down should have negative or positive connotations because it’s mixed on the four graphs [C07] | They would be better if have the same direction (for example up = better) [R] | ||
| “Better” line type | Reviewing the graph, I understand the scale now and it was fairly simple to figure out [C09] | This one is more confusing in that severe fatigue is at the bottom as opposed to the top…my inclination would be that as fatigue worsens it would go up [C09] | It’s a bit unusual to have “reverse” scaling on pain and fatigue (i.e., lines going up means less pain or doing better) [R] |
| The two lower graphs were harder to digest…I’d expect it was plotting pain level with higher on the y-axis reporting MORE pain [R] | |||
| “Normed” line type | The graphs are quite clear and descriptive [C06] | The contrast between treatments is clear, but the magnitude of the effect is absent [R] | |
| No anchors on the Y axis makes it hard to tell if differences are meaningful [R] | |||
| Comments on line-graph variations | |||
| Plain graphs | The graphs themselves are quite clear [C01] | There is no confidence interval, there is no asterisk, so my only reference point was this little thing that says P = 0.01…it’s hard to know at what time points that was determined at [C02] | I like that the treatment is linked to the specific lines (versus a traditional legend) [R] |
| They’re a little bit easier to read [C07] | [The other formats] offer more statistical information that is helpful to the clinician [C05] | What does the p value represent? You have repeated measures, is it the difference at the final time point? [R] | |
| The graphs are clear and if these were the only graphs provided to me I wouldn’t know what I was missing [C05] | |||
| Indication of clinical significance | I believe the asterisk format is the easiest in showing patient results without the confidence intervals [C06] | The legend says that the difference is determined to be clinically important, although I don’t exactly understand what that means [C02] | Presumably, the method of determining a “clinically important difference” would be explained in the article text [R] |
| The asterisks were actually helpful at the different months… because then I know if the differences were significant or not at that point in time [C07] | Using asterisk to denote clinical importance was confusing since it is often used in other studies to reflect statistical significance [R] | How am I supposed to determine clinically significant differences based on these graphs … What measure was used? Is there an established MCID? [R] | |
| The “clinically important difference” was indicated by an asterisk. That was helpful [R] | They’ve eliminated the confidence intervals so I don’t really have a sense of how statistically different everything is [C04] | The graph does not indicate how the lines were estimated/calculated and as a result no information how the p-values were calculated [R] | |
| Indication of confidence limits | The asterisks and the demarcation of the confidence interval at every time point is very helpful [C05] | I think the confidence intervals plus the asterisks are redundant to see which time points are clinically significant...they were a little harder to read [C07] | The * and p value meanings differ, and the curves therefore highlight the ambiguity concerning statistical significance and clinically important differences [R] |
| It gave both standard errors to give you an idea of the disbursement of the data as well as the asterisk to demonstrate statistical significancec in the difference between them [C04] | There are a lot of lines and they’re overlapping so it’s a little bit less easy to interpret, so I think I like the simplicity of the (clinical significance) [C08] | May be difficult for some clinicians to interpret 95% confidence limits as currently explained [R] | |
| What’s also quite interesting is that there is a significant clinical change between these two treatments, however, the confidence intervals themselves overlap [C01] | While the error bars are a little bit helpful for understanding the spread of the data better, visually it gets a bit cumbersome [C09] | ||
| Comments on formats showing proportions changed | |||
| Pie charts | I think it’s just easier for my brain to see and compare the two charts…the bar graph takes me a little bit longer to compare the treatments [C08] | It’s not a format that I’m used to seeing to have the data presented and so it did catch me off guard initially [C09] | At 9 months Treatment Y had 50% improvement, Treatment X had 40% improvement, but the P value is 0.10 demonstrating non-significance. I put down treatments are about the same, but I’m confused as to what the p value truly means [C03] |
| A pie chart is always easier on the eye [R] | Bar graph is easier to describe patient results compared to the pie graph [C06] | What I find ambiguous is that…because we have three variables for each comparison, we don’t know if we’re comparing improved versus improved or the ratio of improved to same to worsened [C01] | |
| The pie chart gives me percentages of patients stating improvement, worsening or about the same and I think that’s important information in discussions with the patient about treatment decision making [C05] | Please read any graphics design book that you can grab hold of …Never use pie charts! [R] | For the first question on physical function, the answer depended on whether an alpha of 0.05 or 0.10 was being considered. Since none was stated, I assumed the classical alpha = 0.05 [R] | |
| Bar charts | (Bar charts)...can show each category, improved, about the same or worsened, head to head against the two treatments…for pie graphs you have to bounce back and forth to see the direct comparisons [C01] | The bar graph takes me a little bit longer to compare the two (treatments) [C08] | We’re given a test of significance but we don’t know what is being compared in that test [C01] |
| Beauty is in the eye of the beholder I guess, but the bar graph…when you’re talking with patients I think clearly shows what you’re trying to describe [C06] | These bar charts are very confusing…there is no confidence interval at all…I don’t have a score to determine if patients improved how much did they improve; if patients worsened how much did they worsen [C03] | I think what’s confusing to me is that the p value is in the bottom right of each of the graphs, but it’s just unclear what was being compared [C02] | |
| I find this graph to be much easier to read than the pie charts [R] | I find these bar charts to be difficult to interpret. They take more time and likely are going to be more prone to error in interpretation [C05] | Technically the overall p-value that probably derives from a X2 type of test cannot be used to justify statements about individual comparisons as the two statements were phrased…instead the comparative residuals of the expected comparison could be used to highlight where differences occur [R] | |
Clinician comments are taken from the in-person interviews. The number in parentheses denotes the specific clinician providing the comment
Representative researchers’ comments are taken from the online open-ended text boxes
This respondent may have misinterpreted the meaning of the asterisks (clinical not statistical significance)