Abstract
A human decision-maker benefits the most from an AI assistant that corrects for their biases. For problems such as generating interpretation of a radiology report given findings, a system predicting only highly likely outcomes may be less useful, where such outcomes are already obvious to the user. To alleviate biases in human decision-making, it is worth considering a broad differential diagnosis, going beyond the most likely options. We introduce a new task, “less likely brainstorming,” that asks a model to generate outputs that humans think are relevant but less likely to happen. We explore the task in two settings: a brain MRI interpretation generation setting and an everyday commonsense reasoning setting. We found that a baseline approach of training with less likely hypotheses as targets generates outputs that humans evaluate as either likely or irrelevant nearly half of the time; standard MLE training is not effective. To tackle this problem, we propose a controlled text generation method that uses a novel contrastive learning strategy to encourage models to differentiate between generating likely and less likely outputs according to humans. We compare our method with several state-of-the-art controlled text generation models via automatic and human evaluations and show that our models’ capability of generating less likely outputs is improved.1
1. Introduction
Cognitive errors occur when an abnormality is identified, but its importance is incorrectly understood, resulting in an incorrect final diagnosis (Onder et al., 2021; Bruno et al., 2015). For example, radiologists may look for confirmatory evidence to support a diagnostic hypothesis and ignore or discount evidence that refutes the hypothesis (confirmation bias; Busby et al. (2018); Onder et al. (2021)). One way to reduce the likelihood of such cognitive errors is to provide cognitive “help” by having a devil’s advocate (Seah et al., 2021; Waite et al., 2017). For this purpose, we propose a new text generation task called “less likely brainstorming” to produce less likely but relevant consultations to bring fresh eyes to examine a case—a powerful way to correct diagnostic errors.
Here, we consider less likely hypotheses in two scenarios. First, they can be hypotheses that humans think are likely but not among the most likely to happen. These hypotheses are critical to providing second opinion of a prior clinical study but are often difficult to generate by traditional decoding techniques. Second, they can be hypotheses that are indeed impossible according to humans, but are close to being true if certain counterfactual assumptions about the input hold. These hypotheses are also helpful as they are often ignored by clinicians. There is a tendency for clinicians to look for a confirmatory diagnostic hypothesis but ignore a refutable one. Note that a less likely hypothesis reflects the likelihood of a potential diagnosis from the human perspective, not from the probability of model output.
We propose Brainstorm, a novel contrastive learning strategy to generate “less likely” hypotheses. We treat this problem as a text generation task as text generation models are the most flexible for providing predictions and explanations for complex tasks; they can generalize to new examples and produce complex, structured diagnoses in many formats. Generation of the “less likely hypotheses” is conditioned on an indicator variable set to trigger the model to prefer outputs are less likely according to humans. For this purpose, we propose two additional loss objectives to effectively learn the relationship between input context, the indicator, and outputs. Without our training strategy, using naive controlled generation training, we find that conditioning on the indicator often leads to generating “highly likely” or irrelevant outputs.
We explore this task in two settings: everyday commonsense reasoning and brain magnetic resonance imaging (MRI) interpretation generation (more details in Section 5). In the everyday commonsense reasoning setting, we adapt Art (Bhagavatula et al., 2020) and E-CARE (Du et al., 2022), which both contain “less plausible” or “implausible” hypotheses that fit our definition of less likely. An illustrative example asking for less likely hypotheses can be found in Figure 1. We show that our approach can generate more “less likely” hypotheses than baselines, including models directly fine-tuned on this set, past controllable generation approaches (Lu et al., 2022), or models with alternate decoding (Li et al., 2022; Liu et al., 2021). In the brain MRI interpretation setting, we experiment with predicting diagnoses from brain MRI reports (see Figure 1). Assessment by a neurologist reveals that our model successfully shifts the distribution of generated diagnoses further toward the tail while still generating relevant diagnoses.
Figure 1:
Examples from MRIInterpret and E-CARE datasets. The task is to generate interpretations or hypotheses that humans would consider to be “less likely” to happen but still relevant to the context. “+” and “~” represent likely and less likely outputs, respectively.
2. Related Work
Uncertainty in Radiology Interpretation
Uncertainty plays a significant role in the process of clinical decision making (Croskerry, 2013). When facing uncertainty, physicians may resort to various erroneous strategies, such as denying the presence of uncertainty resulting in various interpretation biases. These biases could lead to unexpected consequences (Kim and Lee, 2018; Eddy, 1984), including missed diagnoses, misdiagnoses, unnecessary diagnostic examinations and even life-threatening situations (Farnan et al., 2008). Recent work (Seah et al., 2021; Waite et al., 2017) have provided deep-learning based methods and suggestions in reducing errors from interpretation bias on medical imaging. To the best of our knowledge, we are the first to explore reducing bias from interpreting radiology reports via our less likely text generation framework.
Controllable text generation and decoding methods
Controllable text generation is the task of generating text that adheres certain attributes, such as language detoxification (Zhang and Song, 2022; Liu et al., 2021; Dathathri et al., 2020), formality modification (Mireshghallah et al., 2022; Yang and Klein, 2021) and open-ended story generation (Mori et al., 2022; Lin and Riedl, 2021; Fan et al., 2018). The task of controllable text generation encompasses both training-time and decoding-time methods. Training-time approaches include CTRL (Keskar et al., 2019), which learns to utilize control codes to govern attributes in order to generate the desired text, and Quark (Lu et al., 2022), which leverages a strong attribute classifier as a reward function to unlearn unwanted attributes. These methods typically rely on training data that contains both the desired and undesired attributes to be effective in the supervised setting. Our method falls into this category.
On the other hand, decoding-time methods utilize off-the-shelf pre-trained LMs (PLMs) and aim to re-rank the probability of generated text based on specific constraints. PPLM (Dathathri et al., 2020) and FUDGE (Yang and Klein, 2021) are typical methods in this category that train an attribute classifier to guide PLMs to generating desired text. DExperts (Liu et al., 2021) and Contrastive Decoding (Li et al., 2022) are more recent methods that re-weight generation probabilities by contrasting the output distributions between different LMs. We select those two as strong baselines for comparison against our proposed model.
Contrastive Learning in NLP
Contrastive learning (CL) has been applied to a wide range of representation learning tasks in NLP, such as learning task-agnostic sentence representation (Gao et al., 2021) and improving natural language understanding (Jaiswal et al., 2021; Qu et al., 2021). It has recently been applied to text generation tasks as well (An et al., 2022; Cao and Wang, 2021; Lee et al., 2021) where additional hard positive or negative examples are created through techniques such as back-translation or perturbation.
3. Problem Setting
The problem we tackle in this work can be viewed as a controllable text generation task. Let be a premise or a brain MRI report findings, we want a model to generate a likely/less likely hypothesis or interpretation given an indicator by drawing from the distribution . The indicator can take two values: + to indicate generating likely outputs and ~ to generate less likely outputs.
For example, given a premise x = “Tom goes to the gym every day.” in Figure 1 from the E-CARE dataset (more details in Section 5), we want a model to generate a hypothesis that is less likely to happen after , such as “He gets a promotion from his manager who saw him in the gym.”. Although this hypothesis fits into the same scenario as the premise as it directly connects to the premise involving Tom’s daily gym attendance, it is less likely to happen since the causal relationship between going to the gym and receiving a promotion is not common. The understanding of what is “less likely” can be based on the concept of bounded rationality (Simon, 1955), where likely hypotheses are those that are likely given known premises, but less likely hypotheses may stem from additional unknown premises.
It is important to note that when we refer to an output as “less likely/likely”, we mean that it is less likely/likely based on human understanding of . All models we experiment with in this work generate outputs that have high probability according to the model, regardless of whether they are likely or less likely to happen according to humans.
4. Methodology
In this section, we present our method as well as baseline models we compare against. Requirements for these models can be found in Table 1. We use BART (Lewis et al., 2020) as the backbone LM for all experimental settings.
Table 1:
Requirements for various methods. pair means a method requires both for . Quark can take any type of data as inputs but requires a trained classifier. We use Brainstorm′ as an alternative of Brainstorm if and are not both available for . DExperts and CD require that both and could be available for (which is not the case for MRIInterpret, Section 7).
Methods | Data | Need Clf. | ||
---|---|---|---|---|
+ | ~ | pair | ||
Training-time Method | ||||
Mle-LL | ✓ | |||
Mle | ✓ | |||
Quark | ✓ | ✓ | ✓ | ✓ |
Brainstorm | ✓ | |||
Brainstorm′ | ✓ | ✓ | ||
Decoding-time Method | ||||
DExperts | ✓ | |||
CD | ✓ |
4.1. Brainstorm
Our encoder-decoder system takes the concatenation of a pair as input and returns one or multiple generated output sequences . At decoding time , our model iteratively decodes the next token conditioned on the left-hand context, i.e., :
(1) |
where is the next token distribution given the context. The task inputs are described in Section 5.
Besides the standard maximum likelihood training with human reference, we incorporate two additional loss objectives to guide models to associate the context, indicators, and target sequences. The training approach is illustrated in Figure 2.
Figure 2:
An overview of Brainstorm using an example from E-CARE, which consists of three objectives. is the encoder representation of the input conditioned on an indicator and are the decoder representations of positive, hard negative, and other negative target sequences within the same batch, respectively. The objective is highlighted in red where it requires both likely and less likely data.
Margin Loss
First, given the indicator , we want the model to assign a higher estimated probability to human reference than its opposite indicator . Therefore, we apply a margin-based loss:
(2) |
where is the margin value. This loss objective tells models that if the indicator is modified, then the target sequence should have lower probability. Margin loss does not require both likely and less likely outputs and .
Similarity Loss
We propose two versions of a contrastive similarity loss based on the availability of examples that can be used in CL. When both positive and negative examples are available in the same batch, we define the similarity loss as
(3) |
Here, , and represent the hidden representations of input , human reference , and an output in the same batch. encourages the model to maximize the agreement between and its corresponding output . This loss objective encourages a model to learn the relation between certain indicators and the target sequence by contrasting the target sequence with all negative outputs in the batch.
This objective term resembles that in CoNT (An et al., 2022) which takes self-generated outputs as negative samples; here, we conditioned the input on special indicators. Note that at the training time, the indicator could be either + or . When the indicator , the hard negative is the human reference of , and vice versa. We set the weight of the term in Equation (3) associated with the hard negative to 10 throughout the experiment to increase its importance relative to in-batch negatives.
When positive and negative examples are not available at the same time (denoted by a lack of a “pair” check in Table 1), we propose an alternative similarity loss objective that minimizes the similarity of encoder representation and , without comparing to outputs in the batch:
(4) |
We use cosine similarity for both versions.
Final Loss
The overall training objective of Brainstorm is the combination of the standard maximum likelihood estimation (MLE) , margin loss, and similarity loss:
(5) |
where and are hyperparameters. Brainstorm′ replaces by .
4.2. Baselines
4.2.1. Training-Time Baselines
Mle and Mle-LL
Mle is trained on all data. It is a conditional model that learns to generate both and depending on . Mle-LL learns to generate less likely outputs by only training on . Both models are trained with standard MLE.
Quark
(Lu et al., 2022) is a state-of-the-art controllable text generation method that outperforms methods such as unlikelihood training (Welleck et al., 2020). Quark trains an LM to generate text with fewer undesirable properties by maximizing rewards assigned by a reward function. In this study, we use the DeBERTa model (He et al., 2020) as the reward function to help generate more (more details in Section 6).
4.2.2. Decoding-Time Baselines
Modified DExperts
DExperts (Liu et al., 2021) combines a base LM along with two language models called “expert” and “antiexpert” that model text with desired and undesired properties, respectively. The next token distribution is determined by where is the logits for the next token and is the truncated logits from under any truncation sampling methods such as top- sampling. For simplicity, we omit the preceding context in the notation. The hyperparameter controls how far the final token distribution deviates from model M.
In our setting, we modify this definition to be
(6) |
Here, is from the model that learns to generate by only training on pairs. is from the model that learns to generate both and conditioned on the indicator. Unlike Mle, this model does not condition on indicators to generate hypotheses. Instead, it leverages text with both desired (generating ) and undesired properties (generating ). It is shown to effectively maintain the fluency of the generated text (Liu et al., 2021). is from a base LM that generates only. It can be Mle -LL or Brainstorm.
Modified Contrastive Decoding
Contrastive Decoding (CD) combines a larger and a smaller “amateur” model ( and searches for text under a constrained search space (Li et al., 2022). The resulting outputs are intended to amplify the strengths of and remove undesired properties that appear in . A scaling factor controls the penalties of the amateur model in CD.
In our setting, two models have the same size. learns to generate can be Mle-LL or Brainstorm. Intuitively, the ability to generate is preserved, while the tendency to generate is factored out.
Hyperparameters
We experiment with a wide range of values for in DExperts and in CD and show how the fraction changes across these values in Figure 3. We keep the recommended value for the remaining hyperparameters. Unless specified otherwise, we generate outputs using diverse beam search (Vijayakumar et al., 2016).
Figure 3:
Fraction-perplexity trade-off of decoding-time methods CD and DExperts on ART test set and original E-CARE validation set (our test set). We show the trade-off across various values for in CD> and in DExperts. Both CD and DExperts can improve the fraction of less likely hypotheses, but at a very high cost to perplexity.
5. Experimental Settings
We investigate our methods in both brain MRI settings and everyday commonsense reasoning settings (Table 5).
5.1. Everyday Commonsense Reasoning
Two datasets from the commonsense reasoning domain were adapted. See examples in Figure 4 from Appendix.
Art
(Abductive Reasoning in narrative Text; Bhagavatula et al. (2020)) is a large-scale benchmark dataset that tests models’ language-based abductive reasoning skills over narrative contexts. Each instance in the dataset consists of two observations and happened before , as well as a likely and a less likely hypothesis event (happening in between and ) collected from crowd workers. Each “likely” hypothesis is causally related to two observations and each “less likely”; hypothesis is created by editing each “likely” hypothesis. The original task is to generate a likely hypothesis given the observation pair .
E-CARE
(Explainable CAusal REasoning; Du et al. (2022)) tests models’ causal reasoning skills. Each instance in the dataset consists of a premise, a “likely” and a “less likely” hypothesis, and a conceptual explanation of the causality. The likely hypothesis can form a valid causal fact with the premise. Two tasks are introduced: (1) causal reasoning: choosing the “likely” hypothesis given a premise and (2) explanation generation: generating an explanation for the causal fact.
Adapted Setting
In our adapted setting, we want a model to generate given either an observation pair (Art) or a premise (E-CARE) . Formally, let be a binary evaluator that classifies an output into either or based on . We want a model that generates , where .
Evaluation
For Art, we use the default training, validation and test sets to evaluate our models. For E-CARE, we randomly construct training and validation sets from the original training set and use the default validation set as the test set since the original test set is not available. All hyperparameters are determined on the validation set.
For each instance in the test set, we ask a model to generate , then measure the fraction of less likely hypotheses according to an evaluator .
To reduce ambiguity and encourage more consistent human evaluations, we formally define all relevancy categories from rounds of pilot studies. More detailed definitions and annotation instructions can be found in Appendix B and C. We measure both the (1) relevancy and (2) fluency of generated hypothesis in human evaluation.
5.2. MRIInterpret
We present a new dataset MRIInterpret based on the findings and impression sections of a set of de-identified radiology reports we collected from brain MRIs. Each instance consists of a findings , an indicator , and a likely/less likely interpretation of the findings depending on .
Dataset Construction
We first find phrases such as “likely represents”, “consistent with”, and “may be unrelated to” that represent uncertainty from each sentence of reports. We view these phrases as indicators of the presence of interpretations; denote them by or . A likely or less likely indicator (Appendix F) suggests a likely or less likely interpretation of a finding. For each likely indicator , we treat the sub-sentence preceding concatenated with prior 6 sentences as the findings , and the completion of the sentence following as the likely interpretation of the findings . We include prior sentences to provide more context for reaching interpretations. For less likely indicators , we treat the sub-sentence either following or preceding as the less likely interpretation of the findings depending on how is stated. An example can be found in Figure 4.
Indicator Unification
We have collected a variety of indicators and decided to unify them into a minimum set for both likely and less likely indicators. More details of indicator unification can be found in Appendix F.
Evaluation
To ensure the human evaluation for MRIInterpret to be as reliable as possible, we carefully curate a thorough annotation instruction guideline with precise definitions for all relevancy labels in Section 7 and Appendix E.
6. Evaluation on Commonsense Reasoning
6.1. Automatic Evaluation
Our first evaluation relies on automatically assessing whether system outputs are likely or less likely according to humans. We fine-tune DeBERTa models (He et al., 2020) for our automatic evaluation on two everyday commonsense datasets. They take the pair of as input and predict whether is a likely or less likely hypothesis. In our settings, the fine-tuned DeBERTa model achieves accuracy on the test set of Art and achieves on the original validation set of E-CARE.
Table 2 compares a number of methods on our commonsense reasoning datasets. We answer several questions based on these results. We perform a paired bootstrap test for each result by comparing to Mle-LL. We highlight results that are better at 0.05 level of significance.
Table 2:
Performance of generating less likely hypothesis on Art test set and E-CARE validation set. For DExperts and CD, we list the fractions where models reach minimum PPL. The ablation study of our proposed method is shown at the bottom.
ART | E-CARE | |||
---|---|---|---|---|
Model | Frac (↑) | PPL (↓) | Frac (↑) | PPL (↓) |
Mle | 54.1 | 42.6 | 54.5 | 80.4 |
Mle-LL | 56.6 | 42.5 | 52.6 | 84.8 |
+ CD | 59.9 | 49.8 | 63.4 | 107.3 |
+ DExperts | 56.2 | 51.7 | 57.2 | 108.3 |
Brainstorm | 79.4 | 40.7 | 58.1 | 69.2 |
+ CD | 79.7 | 50.2 | 67.2 | 88.1 |
+ DExperts | 79.0 | 51.5 | 58.1 | 89.3 |
Quark | 85.9 | 27.5 | 68.2 | 80.8 |
Brainstorm | ||||
69.3 | 44.9 | 54.6 | 73.2 | |
58.2 | 52.6 | 53.2 | 83.7 | |
Brainstorm′ | 58.3 | 52.0 | 55.1 | 71.2 |
Can we just train on ?
Interestingly, the baseline model Mle-LL that only trained on pairs generates “likely” hypotheses approximately half of the time. This is possibly an effect of the pre-training regimen; furthermore, generating likely hypotheses may be easier and past work has shown that seq2seq models can amplify behaviors like copying that are easy to learn (Goyal et al., 2022).
Are the proposed two loss objectives effective?
We see that compared to Mle-LL, our proposed Brainstorm method achieves substantially higher fractions of less likely hypotheses with no cost to quality in terms of perplexity. At the bottom of Table 2, we show that ablating either of the proposed loss objectives worsens performance (and note that ablating both yields Mle). Brainstorm′ is not as effective since it does not compare with outputs in the batch, but we can see its merits in MRIInterpret (Section 7).
Can decoding-time methods alleviate the problem of generating likely outputs?
We explore whether DExperts and CD can further raise the fraction of less likely generations when combined with either Mle-LL or Brainstorm. These methods have hyperparameters that trade off how much of the “undesired” behavior each can remove from the system. We compute several fraction-perplexity trade-off curves in Figure 3. Notably, although the fraction of less likely outputs can improve, both of these methods significantly increase the perplexity of generations, which corresponds with notably worse fluency of the text. Although these points apparently have high less likely fractions, we caution that the distribution of the text may deviate from the text that DeBERTa was fine-tuned on, meaning that our classifiers may not work well in these ranges. The green lines reflect thresholds where we observe serious degradation in output quality starting to occur. Below this perplexity threshold, the automatic evaluation suggests that both methods demonstrate some capability in alleviating the models’ tendency in generating “likely” hypotheses without too great a cost to perplexity. Note that DExperts is more effective than CD in Art and vice versa in E-CARE.
Table 2 reports the settings where models achieve the minimum perplexities; at these points, perplexity is substantially increased but the fraction of less likely hypotheses is not substantially changed for the majority of results.
Can Quark yield improvement?
In Table 2, the automatic evaluation results show that Quark exceeds Brainstorm by generating 6% more “less likely” hypothesis in ART and 10% more in E-CARE. It also has lower perplexity in Art. To further compare the two models, we conducted a human evaluation on the outputs from two models, and the result shows that Quark generates lower-quality “less likely” hypotheses (Section 6.2).
6.2. Human Evaluation
To further validate the results, we conduct a finer-grained human evaluation on a sample of 100 examples from the test sets of both datasets along two axes – relevancy and fluency. We refined our relevancy evaluation by dividing the “relevancy” category into four subcategories, resulting in a total of five categories for evaluation.: (1) Likely; (2) Less likely; (3) Contradictory - the output is impossible if we assume the input is true; (4) Repetition - the output is describing the same meaning as the input; and (5) Irrelevant - the output has little connection with input. More thorough category definitions with examples, annotation instruction and quality checks for AMT annotators can be found in Appendix C. We compare the performance of three models: Mle-LL, Brainstorm, and Quark (Table 3). As Quark demonstrates better performance in automatic evaluation, we include its generated text in our human evaluation.
Table 3:
Human evaluations on Art and E-CARE. We see that our method is able to produce more “less likely” (L-Likely) outputs on both datasets. We calculated the mean of the ratings from multiple annotators for each sample.
Model | Art | E-CARE | ||||||||
---|---|---|---|---|---|---|---|---|---|---|
Likely (↓) | L-Likely (↑) | Contra. (?) | Rep. (↓) | Irrel. (↓) | Likely (↓) | L-Likely (↑) | Contra. (?) | Rep. (↓) | Irrel. (↓) | |
Mle-LL | 42.3 | 15.2 | 22.7 | 9.5 | 10.3 | 35.4 | 15.6 | 5.7 | 18.6 | 24.7 |
Quark | 14.7 | 20.8 | 51.0 | 4.3 | 9.2 | 35.2 | 15.1 | 5.7 | 3.3 | 40.7 |
Brainstorm | 20.9 | 20.2 | 41.3 | 4.8 | 12.8 | 37.1 | 20.1 | 4.7 | 12.7 | 25.4 |
Our results show a high level of agreement between the automatic evaluation (Table 2) and human evaluation (Table 3) regarding the fraction of “likely” hypotheses on both datasets. On Art, Quark and Brainstorm decrease the fraction of “likely” hypotheses by and , respectively, compared to Mle-LL. However, on E-CARE, the human evaluation indicates that all three models generate an equivalent number of “likely” hypotheses. By further breaking down the “relevancy” category used in the automatic evaluation, we then have a clearer understanding of the distribution of categories among the models’ outputs.
Low-Quality Hypotheses
It is not desirable for models to generate outputs that are repetitions of the input (Repetition) or have little connection to the input (Irrelevant). On the Art dataset, all models generate a small proportion of irrelevant outputs, with Quark and Brainstorm reducing the fraction of “Repetition” hypotheses by half, compared to Mle-LL. However, we get more low-quality outputs on E-CARE. While Brainstorm is able to reduce the fraction of Repetition hypotheses by a large margin, it is not as effective as Quark. One possible reason for this is that Quark is trained to generate outputs that the DeBERTa classifier (the reward model) predicts as less likely; Repetition cases are rarely classified as less likely due to their similarity with the input, but Irrelevant outputs are more likely to be classified this way.
Less Likely versus Contradictory
While less likely hypotheses are desirable, contradictory hypotheses are less so. A typical way of generating a contradictory hypothesis is by simply adding negation: Lisa went laptop shopping yesterday Lisa didn’t go laptop shopping yesterday. However, such examples have little value as the negation brings no new information to the input and is not a useful counterfactual for a user to see.
We evaluate the models’ outputs on the Art dataset, where a significant number of contradictory hypotheses are generated, and find that 43 out of 100 hypotheses generated by Quark include the words “didn’t” or “not,” while only 10 hypotheses generated by Brainstorm and Mle-LL did so. We posit that this is likely due to the DeBERTa classifier assigning high rewards for hypotheses that include negation words, and Quark effectively learning this shortcut.
7. Human Evaluation on MRIInterpret
To evaluate the models’ performance on the radiological interpretation generation setting, we select 30 findings from our validation set that ask for less likely interpretation. For each finding, we select the human reference and generate the top 5 less likely interpretations from 2 baselines (Mle-LL and Mle) and Brainstorm′, resulting in interpretations. We randomized the order of these interpretations before evaluation.
Due to the structure of the indicators in this dataset, methods that require examples to have both and for the same data (see “pair” in Table 1) are not able to be used. Since Quark relies on a trained classifier, we choose not to use Quark as well. A trained classifier on MRIInterpret is not reliable since the training set only consists of naturally occurring data, which is highly imbalanced (see Table 5 in Appendix). This leads the classifier to perform poorly on the “less likely” class, which is the minority class but is also the class of greatest interest in this study. We find that augmenting the training data with counterfactual cases is not easy. For example, “the lack of evidence of restricted diffusion makes it less likely to be” is a naturally occurring prompt from a less likely example, and attempting to change it to a sentence such as “the lack of evidence of restricted diffusion could represent” yields a statement that turns out to be out of distribution from the training data and models do not behave reliably in these counterfactual cases.
For each generated interpretation, we evaluate its (1) relevancy to the findings and (2) whether it contains any hallucinations about findings (Appendix E.2). For relevancy, we asked a neurologist to classify each interpretation into: (1) Relevant and likely; (2) Relevant and less likely; and (3) Irrelevant. Further, for those classified as “Relevant and less likely”, we further evaluate how well the interpretation fits into the context of the findings by grading them on three levels: high, medium and low, ranging from high matches that represent the most obvious less likely interpretations to low matches that represent relevant but exceedingly rare diagnosis. We provide detailed definitions for these categories and include comprehensive annotation guidelines in Appendix E to facilitate consistency in future studies.
Results are shown in Table 4. Most human references (which the neurologist was blinded to) are annotated as either a high or medium match under the relevant but less likely category, suggesting the reliability of the neurologist’s annotation. We find that training on all data (Mle) instead of exclusively on less likely data (Mle-LL) would effectively help generate more relevant but less likely interpretations and reduce the amount of irrelevant ones. One possible reason is that MRIInterpret is a highly imbalanced dataset (Table 5).
Table 4:
Human Evaluation on MRIInterpret. Results are shown as percentages. We evaluated less likely interpretations generated from each model and 30 less likely interpretations from human reference. Results show that our proposed model successfully shifts the distribution of generated interpretations further toward the tail of the “relevant but less likely” category but still generates relevant diagnoses.
Model | Likely | Less likely | Irrel. | ||
---|---|---|---|---|---|
High | Med. | Low | |||
Mle-LL | 6.7 | 40.7 | 21.2 | 14.7 | 16.7 |
Mle | 7.3 | 50.0 | 22.1 | 13.3 | 7.3 |
Brainstorm′ | 6.7 | 42.0 | 32.6 | 8.7 | 10.0 |
Reference | 3.3 | 76.7 | 13.4 | 3.3 | 3.3 |
By comparing the outcomes between human reference and Brainstorm, we find that Brainstorm tends to shift the distribution of generated interpretations towards generating lower matched interpretations, which effectively extends the beam of potential diagnoses that meet the criteria of “relevant but less likely” based on refuting findings. Anecdotally, interpretations in this medium category reflect the sort of alternative hypotheses and “outside-the-box” suggestions that represent the original goal of our approach.
8. Conclusion
In this work, we propose a new text generation task “less likely brainstorming” for reducing cognitive errors in interpreting findings of MRI reports. We found that simply training on less likely data does not help with generating less likely interpretations and hence propose a novel CL method to tackle the problem. In two settings, we show that our proposed training technique can effectively generate more “less likely” hypotheses, producing interpretations that radiologists may not think of, outperforming past training- and decode-time modifications to generation models.
Limitations
Our brain MRI interpretations were evaluated by a single neurologist. Such annotations require deep expertise and are not easily carried out with high quality by trainees, which limited the amount of data we were able to collect. To ensure that the annotation would be as reliable as possible, we carefully thought of the dimensions in evaluating the generated interpretations and proposed a thorough annotation instruction guideline. We believe that future work can conduct more extensive studies using our annotation guidelines as a starting point. Further, the radiology reports we experiment with are from a single academic medical center, which makes the generalizability unclear. Future work is needed to evaluate the performance of our models on data from different medical centers. Finally, future work is needed to evaluate relevant and likely outputs from MRI interpretations to address different forms of interpretation bias and to expand the beam of potential likely diagnoses based on the findings.
Beyond the brain MRI interpretation experiments, our generation experiments are limited to a set of pre-trained models optimized for carrying out generation tasks in English. It is possible that multilingual models generating in languages other than English will show different properties. We are limited by the availability of resources for automatic evaluation in these settings, but a more extensive multilingual evaluation with human users could be conducted in the future.
Ethical Risks
We are proposing better ways for incorporating systems into the radiological diagnostic process. This is aimed at helping improve human decision-making and mitigating the limitations of traditional fully-automatic approaches. However, we believe that it is imperative to rigorously test and evaluate these methods before they can be put into practical clinical settings. We are not claiming that these methods are ready for real-world adoption at this stage.
Acknowledgments
We would like to thank Darcey Riley and TAUR lab at UT for discussion about DExperts and for providing feedback on this work. We acknowledge the funding support from National Science Foundation AI Center Institute for Foundations of Machine Learning (IFML) at University of Texas at Austin (NSF 2019844), as well as NSF CAREER Award IIS-2145280 and IIS-2145640, National Library of Medicine under Award No. 4R00LM013001, and a gift from Salesforce, Inc.
A. Dataset statistics
Dataset statistics can be found in Table 5.
B. Definition of Relevancy Categories on Everyday Commonsense
To encourage more consistent human evaluations, we formally define all relevancy categories as the following. These definitions are refined from rounds of pilot studies to reduce ambiguity for human annotations. Example outputs and explanations for each relevancy category can be found in the annotation interface (Figure 5 and 7).
B.1. E-CARE
Relevant
A hypothesis is relevant if it fits with the same scenario as the premise. It should not introduce new people, places, or things that are not at least plausibly in the same source scenario.
Likely
For the hypothesis to be likely, it must also be causally related to the premise – either the premise causes the hypothesis or the hypothesis causes the premise (you will see both versions of the task below). There should not be clearly more likely hypotheses than it.
Relevant and Less likely
The hypothesis is still the same scenario as the premise (relevant). However, it is less likely to be causally related to the premise. There could be other hypotheses that are superior to the given hypothesis.
Irrelevant
The generated hypothesis does not describe the same scenario as the premise or is not causally related to the premise.
Contradictory
The hypothesis contradicts the premise – it says something that is impossible if we assume the premise to be true (e.g., the premise states that something happened and the hypothesis states that that thing did not happen).
Repetition
The hypothesis is very similar to the premise – it either contains a text span that is a repetition of the premise, or it is expressing nearly the same meaning as the premise.
B.2. Art
Relevant
A hypothesis is relevant if it fits with the same scenario as the observation pair. It should not introduce new people, places, or things that are not at least plausibly in the same source scenario.
Likely
For the hypothesis to be likely, it must also be strongly related to and in a causal fashion – to the extent possible, the first observation should cause the hypothesis and the hypothesis causes the second observation . There should not be clearly more likely hypotheses than it.
Relevant and Less likely
The hypothesis is still the same scenario as the observation pair (relevant). However, it is less likely to be causally related to the observation pair – maybe it could happen following , but not necessarily. There could be other hypotheses that are superior to the given hypothesis.
Irrelevant
The hypothesis does not describe the same scenario as the observation pair: it either involves different people, places, or things, or the events it describes have very little connection to and .
Contradictory
The hypothesis contradicts either observation or observation – it says something that is impossible if we assume and to be true (e.g., states that something happened and the hypothesis states that that thing did not happen).
Repetition
The hypothesis is very similar to either or – it either contains a text span that is a repetition of or , or it is expressing nearly the same meaning as or .
C. Annotation on Everyday Commonsense
The human evaluation by crowdworkers has been judged to be IRB exempt. We hired crowd annotators from US through Amazon Mechanical Turk. These annotators have lifetime approval rates over 99% and more than 1000 approved HITs. We first conducted a quality check on Art and E-CARE. For each dataset, we randomly selected 100 examples from the test set and each example is evaluated by 7 annotators, resulting in annotations for each dataset. We finally selected 7 qualified crowdworkers from each of the datasets. The procedure of filtering out non-qualified workers is shown below. For qualified crowdworkers, we randomly select another 100 examples from each dataset and conduct a final annotation round, resulting in annotations in total. We set maximum time on completing each HIT to 1 hour and each HIT takes approximately 1.5 minutes. We paid annotators , which
Table 5:
A summary of dataset statistics. All datasets are in English. For Art and E-CARE, we show the stats of our adapted versions. Since E-CARE has a hidden test set, we randomly split the original training set into a training and a validation set, and we use the original validation set as our test set. Note that each example in E-CARE asks for either the cause or the effect of the premise.
Dataset | Train | Val | Test | |||||
---|---|---|---|---|---|---|---|---|
Likely | Less Likely | Less Likely | Less Likely | |||||
MRIInterpret | 10097 | 1005 | 121 | — | ||||
Art | 50509 | 50509 | 1781 | 3562 | ||||
E-CARE | cause | effect | cause | effect | cause | effect | cause | effect |
6855 | 6580 | 6855 | 6580 | 762 | 731 | 1088 | 1044 |
is equivalent to and is higher than the minimum USA wage.
Category definitions and annotation instructions with examples are shown in Figure 5, 6, 7 and 8.
Selecting Qualified Workers
After we collected all annotations from the pilot study. We filter out workers by following these steps:
We first filter out workers that annotated less than 4 HITs. With limited amount of annotated HITs, it is hard to evaluate the consistency of their annotations.
For any HIT, if two output sequences are exactly the same but the annotator assigned them different categories, then we remove the worker. For example, in E-CARE, if the premise is “Tom goes to the gym every day.”. and we have the hypotheses “He gets a promotion from his manager who saw him in the gym.” that appears twice, then if one hypothesis is classified as “Relevant and Likely” and another one is classified as “Relevant but Less Likely”, we will filter out this annotator.
We use the “Repetition” category to further filter out annotators. We believe “Repetition” is the least subjective category in our annotation instruction, and using this category to filter out annotations would lead to minimum bias we can project to the selected annotators. This consists of two steps: (1) A model many generate an output that is exactly the input. For example, a model takes as input “Tom goes to the gym every day.” and generate “Tom goes to the gym every day.” as well. This happens sometimes across all models. For those cases, we will filter out annotators that assigned categories other than “Repetition”; (2) Besides the exact match, there are cases where a model’s output is a paraphrase of the input. For these, to minimize our bias, we choose to use models’ outputs that only differs from the input by at most two words to filter out annotators. For example, in Art, if one observation is “Lisa went laptop shopping yesterday”, and the model’s output is “She went laptop shopping yesterday”, then we filter out annotators that do not assign “Repetition” to it.
After we collected all the annotations from qualified workers, we use the above steps to further filter out works that do not meet our standard. Finally, we got valid annotations by three annotators from each datasets. We use Fleiss kappa to calculate the agreement between annotators. The annotators achieve moderate agreement on Art and fair agreement on E-CARE for relevancy evaluation. This is within our expectation since evaluating whether a hypothesis is likely or less likely is subjective.
D. Fluency Evaluation on Everyday Commonsense Reasoning
Fluency evaluation can be found in Table 6. Most of generations from models are fluent and grammatically correct.
E. Annotation on Brain MRI Interpretation
The use of the brain MRI data is covered by an IRB. A neurologist reviewed each finding sample and evaluated the interpretation on multiple metrics.
E.1. Relevancy
The overall objective of the interpretation generation was to produce less likely diagnoses, or interpretations, based on the absence of specific findings. The findings followed a common pattern of “Absence of [finding x] makes it unlikely to
Table 6:
Human evaluation of fluency on everyday commonsense reasoning datasets. Annotators reached substantial agreement on both datasets.
Model | Art | E-CARE | ||
---|---|---|---|---|
Gram. Correct Fluent | Contain Flu. Errors | Gram. Correct Fluent | Contain Flu. Errors | |
Mle-LL | 93.9 | 6.1 | 99.0 | 1.0 |
Quark | 94.6 | 5.4 | 98.0 | 2.0 |
Brainstorm | 93.5 | 6.6 | 95.9 | 4.1 |
Figure 4:
Examples from MRIInterpret, Art and E-CARE. The example shown in the table for E-CARE asks for a likely/less likely effect of the premise. “+”/”~” indicates whether humans would consider the output to be likely/less likely according to the context under the Examples column. We explain why humans would consider these outputs as likely/less likely in the Explanation column (this is not in the training data).
Table 7:
Human evaluation on hallucinations. The result shows the percentage of hallucinations found in 150 generated interpretations from each model.
Model | Hallucination (%) |
---|---|
Mle-LL | 23.3 |
Mle | 30.0 |
Brainstorm | 33.3 |
Reference | 6.6 |
be [interpretation y].” The finding of interest was modified to be standardized across all findings if it used varying terminologies in a similar pattern (see Appendix F for more details). Because the interpretations are oriented in this negated valence, the objective of the output is to produce “relevant but unlikely” interpretations. The annotator rated the interpretation on 3 metrics: (1) relevant and likely, (2) relevant but less likely, and (3) irrelevant.
Relevant and Likely
Output was judged as “relevant and likely” if the interpretation erroneously suggested a diagnosis that would be likely, not unlikely, despite the absence of [finding x]. For instance, “Absence of restricted diffusion within the previously described fluid collections along the right convexity makes it unlikely to be”. An interpretation of “the presence of a small subdural hematoma” is actually a likely diagnosis given the lack of restricted diffusion in the fluid collection since subdural hematomas do not normally demonstrate restricted diffusion.
Relevant but Less Likely
Output was judged as “relevant but less likely” if the interpretation correctly provides a less likely diagnosis due to the absence of [finding x]. For example, “absence of restricted diffusion makes it unlikely to be”. An interpretation of “acute ischemia” is unlikely since diffusion restriction is often associated with acute ischemia.
If the interpretation was judged as “relevant but unlikely”, the degree to which the interpretation fits with the findings was graded on three levels: (1) high, (2) medium, and (3) low.
Less likely interpretations were high matches if they were within the top 5 diagnoses to fit the statement. These were the most obvious interpretations.
Less likely interpretations were medium matches if they were further down the bar of potential interpretations. They still were relevant to the findings and made sense as being less likely given the absence of the finding of interest, but are less obvious and fall outside of the top 5 diagnoses.
Less likely interpretations were low matches if the interpretation was relevant to the findings, but was an exceedingly rare diagnosis to make it of low value to mention as an interpretation.
Irrelevant
Output was judged as “irrelevant” if it was not related to the finding of interest or the structure that the finding of interest is referring to.
E.2. Presence of Hallucination
Lastly, no matter the rating of relevance, presence or absence of hallucination was noted. It was possible to have a relevant but unlikely interpretation with high degree of fit with the finding, but a hallucination that does not appear in the original findings was added. We therefore evaluate whether each interpretation contains hallucinations.
The results are shown in Table 7. The models listed contain a large proportion of hallucinated content especially for Mle and Brainstorm. We examined what these hallucinations look like. We found that in the most cases, models hallucinate about the findings (generating some findings that do not actually written in the report) and concatenate those hallucinated findings after their interpretations. For examples, a generated interpretation would be “an acute infarction although this is limited by the presence of contrast enhancement”, “intracranial abscess although this is limited by the presence of significant soft tissue swelling”, or “blood products in the ventricular system as seen on prior CT.”
However, unlike other text generation tasks such as text summarization where hallucinations are hard to identify, hallucinations in MRIInterpret follow a pattern of interpretation followed by the non-existent findings. Although future work could work on how to directly generate interpretations without hallucination, a rule-based heuristics can remove the majority of hallucinations in the current version of our system.
Table 8:
Examples of generating less likely hypotheses across different methods on Art.
O1: Riley went to the store with her mother. | |
O2: Riley wore her cowboy boots to school the next day. | |
Mle-LL | Riley’s mother bought her cowboy boots. |
+ CD (τCD = 0.5) | Riley had bought cowboy shoes that she had not worn before. |
+ CD (τCD = 1.0) | Her mother bought a new cowboy shirt for Riley. |
+ CD (τCD = 1.5) | Riiley got her new cowboy boots torn. |
+ DExperts (α = 0.1) | Riley’s mother bought her cowboy boots. |
+ DExperts (α = 0.5) | Riley’s mother bought her cowboy boots. |
+ DExperts (α = 1.0) | Riley didn’t have cowboy boots. |
Brainstorm | Riley bought a pair of cowboy boots for her mother. |
+ CD (τCD = 0.5) | Rice bought cowboy hats with her money and jeans. |
+ CD (τCD = 1.0) | Rice’s mother bought her a pair of cowboy boots to wear to school next to her cowboy outfit. |
+ CD (τCD = 1.5) | Her mom bought her a new pair of cowboy boots. |
+ DExperts (α = 0.1) | Riley’s mother bought her a pair of cowboy boots. Riley hated them. |
+ DExperts (α = 0.5) | Riley bought a pair of cowboy pants. |
+ DExperts (α = 1.0) | Riley bought a pair of cowboy pants. |
Quark | Riley’s mother bought her cowboy boots at the store and Riley hated them. |
Table 9:
Examples of generating less likely hypotheses across different methods on E-CARE.
Premise: The whole kitchen was burnt down more quickly. | |
Ask for: less likely cause | |
Mle-LL | The whole kitchen was burnt down more quickly. |
+ CD (τCD = 0.5) | Tom poured water in the pan of acid. |
+ CD (τCD = 1.0) | There was a big thunderstorm yesterday. |
+ CD (τCD = 1.5) | There is a big storm outside. |
+ DExperts (α = 0.1) | The whole kitchen was burnt down more quickly. |
+ DExperts (α = 0.5) | The fire in the kitchen spread outwards. |
+ DExperts (α = 1.0) | There are more and more fires in this place. |
Brainstorm | Tom put a lot of fuel on the fire. |
+ CD (τCD = 0.5) | Tom poured a bucket of water to a sink which has a high temperature. |
+ CD (τCD = 1.0) | There was an accident at night. |
+ CD (τCD = 1.5) | Tom poured gasoline to the stove. |
+ DExperts (α = 0.1) | There is a fire in the kitchen. |
+ DExperts (α = 0.5) | The whole kitchen was filled with smoke. |
+ DExperts (α = 1.0) | Tom’s kitchen is leaking water. |
Quark | The fire in the kitchen was very hot. |
F. Indicator Unification for MRIInterpret
We narrowed down the indicators to a smaller set to ensure that our model sees sufficient data for each indicator during training. The indicator mappings are shown in Figure 9 and 10. We also include the way we flip these indicators for the margin loss objective.
G. Example of generated outputs
We show examples of generated outputs for both everyday commonsense reasoning datasets in Table 8 and 9.
H. Implementation Details
H.1. Significance Test
We perform a paired bootstrap test for each result by comparing to Mle-LL. We highlight results that are better at 0.05 level of significance.
H.2. Computing Infrastructure
We use BART from HuggingFace Transformers (Wolf et al., 2020), which is implemented in the PyTorch framework.
H.3. Training Details
We fine-tune BART-Large (400M parameters) with 1 NVIDIA RTX A6000 GPU on all experiments and it converges in 2 epochs. We use AdamW as our optimizer with adam epsilon set to 1e-8. Learning rate is set to 5e-5 with linear schedule warmup. There is no warm-up step.
H.3.1. Everyday Commomsense Reasoning
We initialize the model from facebook/bartx–large. The batch size is set to 64 if only using MLE objective and 42 otherwise. We set maximum input length to 100 and maximum output length to 64. Most text should fit into these lengths. The average training time for each model is around 0.8 GPU hours if only using MLE objective and 1.5 GPU hours otherwise.
H.3.2. MRIInterpret
We initialize the model from GanjinZero/biobart–large (Yuan et al., 2022). The batch size is set to 32. We set maximum input length to 256 and maximum output length to 60. Most text should fit into these lengths. The average training time for each model is around 0.8 GPU hours if only using MLE objective and 1.2 GPU hours otherwise.
H.4. Hyperparameter Setups
Brainstorm
For the margin loss (Equation (2)), we chose within in the range of and and set it to 0.005 in the log space as it works well throughout our experiments. and are set to 1.0 and 10.0, respectively, as they achieve the best result on the validation set.
Quark
We follows the default parameter setups in the original work with 6000 training steps for both commonsense reasoning datasets.
Decoding
We use diverse beam search for all experiments with diversity penalty set to 1.0. We set in CD from to , and in DExperts from to 1. We keep the recommended values for the remaining hyperparameters.
Figure 5:
Annotation Interface (I) for Art.
Figure 6:
Annotation Interface (II) for Art.
Figure 7:
Annotation Interface (I) for E-CARE.
Figure 8:
Annotation Interface (II) for E-CARE.
Figure 9:
Unifying “likely” indicators in MRIInterpret.
Figure 10:
Unifying “less likely” indicators in MRIInterpret and how we map flipped indicators.
Footnotes
Code is available at https://github.com/Liyan06/Brainstorm.
References
- An Chenxin, Feng Jiangtao, Lv Kai, Kong Lingpeng, Qiu Xipeng, and Huang Xuanjing. 2022. Cont: Contrastive neural text generation. abs/2205.14690.
- Bhagavatula Chandra, Bras Le Ronan, Malaviya Chaitanya, Sakaguchi Keisuke, Holtzman Ari, Rashkin Hannah, Downey Doug, Yih Wen tau, and Choi Yejin. 2020. Abductive commonsense reasoning. In International Conference on Learning Representations. [Google Scholar]
- Bruno Michael A., Walker Eric A., and Abujudeh Hani H.. 2015. Understanding and confronting our mistakes: The epidemiology of error in radiology and strategies for error reduction. RadioGraphics, 35(6):1668–1676. [DOI] [PubMed] [Google Scholar]
- Busby Lindsay P., Courtier Jesse L., and Glastonbury Christine M.. 2018. Bias in radiology: The how and why of misses and misinterpretations. RadioGraphics, 38(1):236–247. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cao Shuyang and Wang Lu. 2021. CLIFF: Contrastive learning for improving faithfulness and factuality in abstractive summarization. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 6633–6649, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics. [Google Scholar]
- Croskerry Pat. 2013. From mindless to mindful practice — cognitive bias and clinical decision making. New England Journal of Medicine, 368(26):2445–2448. [DOI] [PubMed] [Google Scholar]
- Dathathri Sumanth, Madotto Andrea, Lan Janice, Hung Jane, Frank Eric, Molino Piero, Yosinski Jason, and Liu Rosanne. 2020. Plug and play language models: A simple approach to controlled text generation. In International Conference on Learning Representations. [Google Scholar]
- Du Li, Ding Xiao, Xiong Kai, Liu Ting, and Qin Bing. 2022. e-CARE: a new dataset for exploring explainable causal reasoning. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 432–446, Dublin, Ireland. Association for Computational Linguistics. [Google Scholar]
- Eddy David M.. 1984. Variations in physician practice: The role of uncertainty. Health Affairs, 3(2):74–89. [DOI] [PubMed] [Google Scholar]
- Fan Angela, Lewis Mike, and Dauphin Yann. 2018. Hierarchical neural story generation. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 889–898, Melbourne, Australia. Association for Computational Linguistics. [Google Scholar]
- Farnan JM, Johnson JK, Meltzer DO, Humphrey HJ, and Arora VM. 2008. Resident uncertainty in clinical decision making and impact on patient care: a qualitative study. Quality and Safety in Health Care, 17(2):122–126. [DOI] [PubMed] [Google Scholar]
- Gao Tianyu, Yao Xingcheng, and Chen Danqi. 2021. SimCSE: Simple contrastive learning of sentence embeddings. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 6894–6910, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics. [Google Scholar]
- Goyal Tanya, Xu Jiacheng, Li Junyi Jessy, and Durrett Greg. 2022. Training dynamics for text summarization models. In Findings of the Association for Computational Linguistics: ACL 2022, pages 2061–2073, Dublin, Ireland. Association for Computational Linguistics. [Google Scholar]
- He Pengcheng, Liu Xiaodong, Gao Jianfeng, and Chen Weizhu. 2020. Deberta: Decoding-enhanced bert with disentangled attention. ArXiv, abs/2006.03654. [Google Scholar]
- Jaiswal Ajay, Tang Liyan, Ghosh Meheli, Rousseau Justin F., Peng Yifan, and Ding Ying. 2021. Radbert-cl: Factually-aware contrastive learning for radiology report classification. In Proceedings of Machine Learning for Health, volume 158 of Proceedings of Machine Learning Research, pages 196–208. PMLR. [PMC free article] [PubMed] [Google Scholar]
- Keskar Nitish Shirish, McCann Bryan, Varshney Lav R., Xiong Caiming, and Socher Richard. 2019. Ctrl: A conditional transformer language model for controllable generation. ArXiv, abs/1909.05858. [Google Scholar]
- Kim Kangmoon and Lee Young-Mee. 2018. Understanding uncertainty in medicine: concepts and implications in medical education. Korean Journal of Medical Education, 30(3):181–188. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lee Seanie, Lee Dong Bok, and Hwang Sung Ju. 2021. Contrastive learning with adversarial perturbations for conditional text generation. In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3–7, 2021. OpenReview.net. [Google Scholar]
- Lewis Mike, Liu Yinhan, Goyal Naman, Ghazvininejad Marjan, Mohamed Abdelrahman, Levy Omer, Stoyanov Veselin, and Zettlemoyer Luke. 2020. BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 7871–7880, Online. Association for Computational Linguistics. [Google Scholar]
- Li Xiang Lisa, Holtzman Ari, Fried Daniel, Liang Percy, Eisner Jason, Hashimoto Tatsunori, Zettlemoyer Luke, and Lewis Mike. 2022. Contrastive decoding: Open-ended text generation as optimization.
- Lin Zhiyu and Riedl Mark. 2021. Plug-and-blend: A framework for controllable story generation with blended control codes. In Proceedings of the Third Workshop on Narrative Understanding, pages 62–71, Virtual. Association for Computational Linguistics. [Google Scholar]
- Liu Alisa, Sap Maarten, Lu Ximing, Swayamdipta Swabha, Bhagavatula Chandra, Smith Noah A., and Choi Yejin. 2021. DExperts: Decoding-time controlled text generation with experts and anti-experts. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 6691–6706, Online. Association for Computational Linguistics. [Google Scholar]
- Lu Ximing, Welleck Sean, Jiang Liwei, Hessel Jack, Qin Lianhui, West Peter, Ammanabrolu Prithviraj, and Choi Yejin. 2022. Quark: Controllable text generation with reinforced unlearning. ArXiv, abs/2205.13636. [Google Scholar]
- Mireshghallah Fatemehsadat, Goyal Kartik, and Berg-Kirkpatrick Taylor. 2022. Mix and match: Learning-free controllable text generationusing energy language models. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 401–415, Dublin, Ireland. Association for Computational Linguistics. [Google Scholar]
- Mori Yusuke, Yamane Hiroaki, Shimizu Ryohei, and Harada Tatsuya. 2022. Plug-and-play controller for story completion: A pilot study toward emotion-aware story writing assistance. In Proceedings of the First Workshop on Intelligent and Interactive Writing Assistants (In2Writing 2022), pages 46–57, Dublin, Ireland. Association for Computational Linguistics. [Google Scholar]
- Onder Omer, Yarasir Yasin, Azizova Aynur, Durhan Gamze, Onur Mehmet Ruhi, and Ariyurek Orhan Macit. 2021. Errors, discrepancies and underlying bias in radiology with case examples: a pictorial review. Insights into Imaging, 12(1). [DOI] [PMC free article] [PubMed] [Google Scholar]
- Qu Yanru, Shen Dinghan, Shen Yelong, Sajeev Sandra, Chen Weizhu, and Han Jiawei. 2021. Coda: Contrast-enhanced and diversity-promoting data augmentation for natural language understanding. In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3–7, 2021. OpenReview.net. [Google Scholar]
- Seah Jarrel C Y, Tang Cyril H M, Buchlak Quinlan D, Holt Xavier G, Wardman Jeffrey B, Aimoldin Anuar, Esmaili Nazanin, Ahmad Hassan, Pham Hung, Lambert John F, Hachey Ben, Hogg Stephen J F, Johnston Benjamin P, Bennett Christine, Oakden-Rayner Luke, Brotchie Peter, and Jones Catherine M. 2021. Effect of a comprehensive deep-learning model on the accuracy of chest -ray interpretation by radiologists: a retrospective, multi-reader multicase study. The Lancet Digital Health, 3(8):e496–e506. [DOI] [PubMed] [Google Scholar]
- Simon Herbert A.. 1955. A behavioral model of rational choice. The Quarterly Journal of Economics, 69(1):99. [Google Scholar]
- Vijayakumar Ashwin K, Cogswell Michael, Selvaraju Ram-prasath R, Sun Qing, Lee Stefan, Crandall David, and Batra Dhruv. 2016. Diverse beam search: Decoding diverse solutions from neural sequence models. arXiv preprint arXiv:1610.02424.
- Waite Stephen, Scott Jinel, Gale Brian, Fuchs Travis, Kolla Srinivas, and Reede Deborah. 2017. Interpretive error in radiology. American Journal of Roentgenology, 208(4):739–749. [DOI] [PubMed] [Google Scholar]
- Welleck Sean, Kulikov Ilia, Roller Stephen, Dinan Emily, Cho Kyunghyun, and Weston Jason. 2020. Neural text generation with unlikelihood training. In International Conference on Learning Representations. [Google Scholar]
- Wolf Thomas, Debut Lysandre, Sanh Victor, Chaumond Julien, Delangue Clement, Moi Anthony, Cistac Pierric, Rault Tim, Louf Remi, Funtowicz Morgan, Davison Joe, Shleifer Sam, von Platen Patrick, Ma Clara, Jernite Yacine, Plu Julien, Xu Canwen, Le Scao Teven, Gugger Sylvain, Drame Mariama, Lhoest Quentin, and Rush Alexander. 2020. Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 38–45, Online. Association for Computational Linguistics. [Google Scholar]
- Yang Kevin and Klein Dan. 2021. FUDGE: Controlled text generation with future discriminators. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 3511–3535, Online. Association for Computational Linguistics. [Google Scholar]
- Yuan Hongyi, Yuan Zheng, Gan Ruyi, Zhang Jiaxing, Xie Yutao, and Yu Sheng. 2022. Biobart: Pretraining and evaluation of a biomedical generative language model.
- Zhang Hanqing and Song Dawei. 2022. Discup: Discriminator cooperative unlikelihood prompt tuning for controllable text generation. In The 2022 Conference on Empirical Methods in Natural Language Processing, Abu Dhabi. [Google Scholar]