Abstract
Developing high-performance entity normalization algorithms that can alleviate the term variation problem is of great interest to the biomedical community. Although deep learning-based methods have been successfully applied to biomedical entity normalization, they often depend on traditional context-independent word embeddings. Bidirectional Encoder Representations from Transformers (BERT), BERT for Biomedical Text Mining (BioBERT) and BERT for Clinical Text Mining (ClinicalBERT) were recently introduced to pre-train contextualized word representation models using bidirectional Transformers, advancing the state-of-the-art for many natural language processing tasks. In this study, we proposed an entity normalization architecture by fine-tuning the pre-trained BERT / BioBERT / ClinicalBERT models and conducted extensive experiments to evaluate the effectiveness of the pre-trained models for biomedical entity normalization using three different types of datasets. Our experimental results show that the best fine-tuned models consistently outperformed previous methods and advanced the state-of-the-art for biomedical entity normalization, with up to 1.17% increase in accuracy.
Introduction
Entity linking, which aims to link entity mentions detected in a document to their corresponding concepts in a given knowledge base (KB) or an ontology1, is one of the fundamental tasks in information extraction. The main challenges of this task are (1) ambiguity – the same entity mention may be linked to multiple concepts, (2) variation – the same concept can be linked by different entity mentions, and (3) absence – entity mentions may not be linked to any concept in the given KB. In the biomedical domain, this task is also known as entity normalization or encoding. Unlike in the general domain where ambiguity is the primary challenge, variation is much more common than ambiguity in the biomedical domain2,3. Therefore, developing high-performance entity normalization algorithms that can alleviate the variation problem is of great interest to the biomedical community.
Many studies have focused on solving the variation challenge in the biomedical domain, resulting in development of rule-based methods3–5, machine learning-based methods6,7, and deep learning-based methods2,8. Kang et al.5 developed a rule-based natural language processing (NLP) module containing 5 types of rules, to improve disease normalization in biomedical text. Ghiasvand and Kate4 first automatically learned 554 edit distance patterns of term variations between all the synonyms of disorder concepts in the Unified Medical Language System (UMLS)9 as well as between the entity mentions in the training data and their corresponding concepts in the UMLS. They then normalized the entity mentions in the test data by performing exact match between the variations generated by the learned patterns and an entity mention in the training data or a concept name in the given KB. Their system named UWM was the best system for the disease and disorder mention normalization task of the SemEval 2014 challenge4,10. D’Souza and Ng3 proposed a multi-pass sieve system by defining 10 types of rules at different priority levels to measure morphological similarity between entity mentions and candidate concepts in the given KB. Leaman et al.7 proposed a pairwise learning-to-rank method by adopting vector space model to represent entity mentions and concepts, and using a similarity matrix to measure the similarities between entity mentions and candidate concepts. Xu et al.6 also proposed a pairwise learning-to-rank method by defining 3 kinds of features and employing the linear RankSVM11 to normalize each positive adverse reaction mention to an entry in MedDRA. Their system achieved the best performance in the TAC 2017 ADR challenge12. Li et al.2 proposed a convolutional neural network (CNN) architecture that regarded biomedical entity normalization as a ranking problem, which takes advantage of CNN in modeling semantic similarities between entity mentions and candidate concepts. The method outperformed traditional rule-based methods, achieving the state-of-the-art performance. Luo et al.13 proposed a multi-view CNN with multi-task shared structure to normalize diagnostic and procedure names simultaneously in Chinese discharge summaries to standard concepts.
Although deep learning-based methods2,13 have been successfully applied to biomedical entity normalization, they required pre-trained word embeddings that were often learned from a large corpus of unannotated texts. Word2vec14 has been widely adopted to pre-train word embeddings from large corpora and was also used in the work of Li et al.2 and Luo et al.13. Recently, ELMo15 generalized traditional word embeddings to contextual word embeddings and advanced the state-of-the-art for several major NLP benchmarks when integrating contextual word embeddings with existing task-specific architectures. The Generative Pre-trained Transformer (GPT)16 introduced minimal task-specific parameters and could be trained on the downstream tasks by simply fine-tuning the pre-trained parameters. Unlike ELMo and GPT, which used unidirectional language models for pre-training, Bidirectional Encoder Representations from Transformers (BERT) introduced masked language models to enable pre-training deep bidirectional representations and advanced the state-of-the-art for eleven NLP tasks17. Based on the BERT architecture, BioBERT18 (BERT for Biomedical Text Mining) and ClinicalBERT19–21 (BERT for Clinical Text Mining), which were domain-specific language representation models pre-trained on large-scale biomedical articles and clinical notes, were introduced to advance the state-of-the-art performance on many biomedical and clinical NLP tasks.
Despite promising work on the pre-trained BERT / BioBERT / ClinicalBERT models for many NLP tasks such as named entity recognition (NER), relation classification (RC) and question answering (QA) in both the general domain17 and biomedical domain18–22, no existing work has investigated the models for biomedical entity normalization. This task is very different from the above NLP tasks in that NER and QA are token-level tagging tasks and RC is single sentence classification task while biomedical entity normalization can be seen as sentence pair classification task, where we decide whether a candidate concept can be linked by a given entity mention. As a preliminary study, here we proposed an entity normalization architecture by fine-tuning the pre-trained BERT / BioBERT / ClinicalBERT models and conducted extensive experiments to evaluate the effectiveness of the pre-trained models for the entity normalization task using three different types of datasets in the biomedical domain.
Methods
Datasets
We used three different types of datasets in this study, namely ShARe/CLEF - the ShARe/CLEF eHealth 2013 Challenge corpus23, NCBI - the NCBI disease corpus24, and TAC2017ADR - the TAC 2017 ADR corpus12. Table 1 shows the statistics of the three datasets.
Table 1:
Statistics of the three types of datasets used in this study.
ShARe/CLEF (Clinical Notes) | NCBI (PubMed Abstracts) | TAC2017ADR (Drug Labels) | ||||
---|---|---|---|---|---|---|
train | test | train | test | train | test | |
#documents | 199 | 99 | 692 | 100 | 101 | 99 |
#mentions | 5,816 | 5,351 | 5,921 | 960 | 7,038 | 6,343 |
#mentions that are linkable | 4,175 | 3,601 | 5,921 | 960 | 6,991 | 6,325 |
#mentions that are unlinkable | 1,641 | 1,750 | 0 | 0 | 47 | 18 |
#concepts | 88,150 | 9,664 | 23,668 |
ShARe/CLEF: This dataset contains 298 de-identified clinical notes collected from a US intensive care data repository including discharge summaries, electrocardiograms, echocardiograms, and radiology reports, which was partitioned into 199 notes for training and development and 99 notes for testing. Based on a pre-defined annotation guideline, a disorder mention in each clinical note was manually annotated with its mapping concept unique identifier (CUI) within the SNOMED-CT subset of the UMLS9. If there was no mapping concept for a disorder mention, a CUI-less label (i.e., unlinkable) was assigned. We followed the guideline to construct the SNOMED-CT subset from the UMLS 2012AB, which contains 88,150 disorder concepts. Table 1 shows that 28.2% of the training mentions and 32.7% of the testing mentions were unlinkable, which illustrates the absence challenge of entity normalization.
NCBI: This dataset contains 792 PubMed abstracts, which was split into 692 abstracts for training and development, and 100 abstracts for testing. A disorder mention in each PubMed abstract was manually annotated with its mapping concept identifier in the MEDIC lexicon25. In this study, we used the July 6, 2012 version of MEDIC, which contains 7,827 MeSH identifiers and 4,004 OMIM identifiers, grouped into 9,664 disease concepts. Different from the ShARe/CLEF dataset, only those disorder mentions that can be mapped to a concept in MEDIC were annotated in NCBI. As a result, all the annotated disorder mentions have their corresponding concept identifiers.
TAC2017ADR: This dataset contains 200 drug labels, which was split into 101 labels for training and development, and 99 labels for testing. An adverse reaction in each drug label was manually annotated with its mapping MedDRA Lower Level Term (LLT) and the corresponding Preferred Term (PT). If there was no ideal PT mapped for an adverse reaction mention, a High Level Term (HLT) or a High Level Group Term (HLGT) was provided if appropriate, otherwise an “unmapped” tag (i.e., unlinkable) was assigned to the mention. In this study, we constructed a KB from MedDRA v18.1, which contains 21,612 PTs, 1,721 HLTs, and 335 HLGTs, grouped into 23,668 unique concepts. Note that only 0.7% of the training mentions and 0.3% of the testing mentions were unlinkable in this dataset.
Entity Normalization - Problem Definition
Given an entity mention m recognized from a sentence x within a document d, and a KB which consists of a set of concepts, the task of entity normalization is to link m to the corresponding concept c in KB, m → c. If there is no mapping concept in KB for m, then m → NIL, where NIL denotes that m is unlinkable.
Entity Normalization – System Architecture
Figure 1 shows the system architecture for entity normalization used in this study, which consists of four modules: preprocessing, candidate concept generation, candidate concept ranking and unlinkable mention prediction.
Figure 1:
System architecture for entity normalization used in this study.
- Preprocessing: We preprocessed each mention and each concept in KB with the following strategies.
- - Abbreviation Resolution – We used Ab3p26 toolkit to detect the abbreviations within each document, and then replaced each mention in short-form abbreviation with its corresponding long form. (e.g., WT → Wilms tumor) Specifically, for the ShARe/CLEF and NCBI datasets, we also expanded all possible abbreviated disorder mentions using Schwartz and Hearst’s algorithm27 and a list of disorder abbreviations collected from Wikipedia as in previous work2,3.
- - Numeric Synonyms Resolution – We replaced all the numerical words in the mentions and concepts to their corresponding Arabic numerals as in previous work2,3,7. (e.g., one / first / i / single → 1)
- - Other Preprocessing – Finally, we tokenized all the mentions and concepts by whitespace, removed all the punctuations, stemmed the tokens with the Porter stemmer and converted all the tokens into lower case ASCII. All of these were implemented using the CLAMP28 toolkit.
Candidate Concept Generation: We generated candidate concepts for each mention with the commonly used information retrieval (IR) based method6,29–31, which included the following two steps. We first indexed all the concept names and training mentions with their concept ids. Then, we employed the traditional IR model of BM2532 provided by Lucene to retrieve the top 10 candidate concepts for each mention m.
Candidate Concept Ranking: We reranked the candidate concepts by fine-tuning the pre-trained BERT / BioBERT / ClinicalBERT models, where we transformed the ranking task as a sentence-pair classification task. Specifically, for each mention m and a candidate concept c, we constructed a sequence [CLS] m [SEP] c as the input of the fine-tuning procedure, where [CLS] was the special word used for the classification output, and [SEP] was the special word used for separating m and c. The output of the fine-tuning procedure was the final hidden state of the first word [CLS] of the input sequence, which was a fixed-dimensional word embedding C ϵ ℝH. The only new parameters added during the fine-tuning procedure were W ϵ ℝK×H. which was used for the final classifier layer. Here K = 2 was the number of classifier labels. If c is the mapping concept for m, the classifier label is 1, otherwise the label is 0. The probability of label=1 was computed with a softmax function, which was used as the ranking score of each pair (m, c): score (m, c) = P(label = 1|m, c) = softmax(CWT).
Unlinkable Mention Prediction: Because some entity mentions may not have any mapping concepts in KB, it is necessary to predict unlinkable mentions. If there were no candidate concepts returned from Lucene BM25, we predicted m as an unlinkable mention and return m → NIL undoubtedly. Otherwise, we chose the top ranking concept c * = arg max score (m, c"). Here, we validated whether m → c * holds by adopting a simple and widely used method to learn a NIL-threshold m. Namely, if score (m, c *) > m, then m → c *, otherwise m → NIL. We learned the threshold m from the training data with a small held-out development set.
BERT Models
In this study, we used the pre-trained BERT33, BioBERT34, and ClinicalBERT19 models for the fine-tuning procedure. BERT models were trained on Wikipedia and BooksCorpus. BioBERT models were initialized with BERTBase_Cased model and pre-trained with additional biomedical corpus including PubMed abstracts (PubMed), PubMed Central full-text articles (PMC), or PubMed+PMC. There were three types of publicly available ClinicalBERT19–21 models trained with clinical notes from MIMIC-III (Medical Information Mart for Intensive Care III) critical care database35. Huang et al.20 pre-trained the ClinicalBERT model from scratch with randomly sampled 100,000 clinical notes from MIMIC-III. Si et al.19 pre-trained two ClinicalBERT models initialized from BERTBase_Cased and BERTLarge_Cased with all the clinical notes from MIMIC-III. Alsentzer et al.21 pre-trained two ClinicalBERT models initialized from BioBERT with all the clinical notes and all the discharge summaries from MIMIC-III. In this study, we investigated the two ClinicalBERT models at 300K training steps released by Si et al.19. More specifically, we investigated four different versions of BERT models (i.e., BERTBase_Cased, BERTBase_Uncased, BERTLarge_Cased, BERTLarge_Uncased), three different versions of BioBERT models (i.e., BioBERTBase_Cased+PubMed, BioBERTBase_Cased+PMC, BioBERTBase_Cased+PubMed+PMC), and two different versions of ClinicalBERT models (i.e., ClinicalBERTBase_Cased+MIMIC, ClinicalBERTLarge_Cased+MIMIC).
Parameters Settings
For fine-tuning, most model hyperparameters were the same as those saved in the pre-trained model, with the exception of the batch size, learning rate, and number of training epochs33. In this study, we fixed the learning rate at 2e-5, tuned the batch size with 16 and 32, tuned the number of training epochs from 1 to 10, and saved the model with the best performance.
Evaluation Metrics
Following previous work2,3, we evaluated the performance of different entity normalization algorithms in terms of accuracy, which was the percentage of entity mentions that were correctly normalized.
Results
Comparisons of different pre-trained models
Table 2 shows the performance comparisons of different pre-trained models with the BM25 baseline for biomedical entity normalization. From the table, we see that (1) All the BERT / BioBERT / ClinicalBERT models outperformed the BM25 model by at least 5.44% (90.58 vs. 85.14) and 1.53% (92.62 vs. 91.09) on both the ShARe/CLEF and TAC2017ADR datasets. Most of them outperformed the BM25 model for the NCBI dataset by up to 0.83% (89.06 vs. 88.23) except BERTLarge_Uncased, BioBERTBase_Cased+PubMed and ClinicalBERTLarge_Cased+MIMIC. (2) The BERT models with cased version were better than that with uncased version in most cases for biomedical entity normalization. (3) For the ShARe/CLEF and TAC2017ADR datasets, all the three BioBERT models outperformed the BERTBase_Cased model and both the two ClinicalBERT models outperformed the corresponding BERTBase_Cased and BERTLarge_Cased models. However, for the NCBI dataset, only BioBERTBase_Cased+PubMed+PMC and ClinicalBERTBase_Cased+MIMIC were better than BERTBase_Cased. (4) BioBERTBase_Cased+PubMed achieved the best performance on both the ShARe/CLEF and TAC2017ADR datasets, while BioBERTBase_Cased+PubMed+PMC achieved the best performance for the NCBI dataset.
Table 2:
Comparisons of different pre-trained models. The bold score denotes the best performance of each dataset.
ShARe/CLEF | NCBI | TAC2017ADR | |
---|---|---|---|
BM25 | 85.14 | 88.23 | 91.09 |
BERTBase_Cased | 90.62 | 88.85 | 92.62 |
BERTBase_Uncased | 90.58 | 88.65 | 92.97 |
BERTLarge_Cased | 90.73 | 88.85 | 92.87 |
BERTLarge_Uncased | 90.66 | 88.13 | 92.87 |
BioBERTBase_Cased+PubMed | 91.10 | 88.23 | 93.22 |
BioBERTBase_Cased+PMC | 90.99 | 88.65 | 92.97 |
BioBERTBase_Cased+PubMed+PMC | 91.09 | 89.06 | 93.17 |
ClinicalBERTBase_Cased+MIMIC | 90.62 | 88.96 | 92.70 |
ClinicalBERTLarge_Cased+MIMIC | 90.88 | 88.13 | 92.94 |
Comparisons with existing work
We compared the following state-of-the-art methods with our best fine-tuned BERT-based ranking model.
UWM4: the best challenge system on the ShARe/CLEF dataset, which is a rule-based system.
TaggerOne36: the best machine learning-based system up to date on the NCBI dataset. It performs named entity recognition and normalization jointly, which is significantly different from our problem definition.
Xu et al.’system6: the best challenge system on the TAC2017ADR dataset, which is a machine learning-based system.
D’Souza & Ng’s system3: the best rule-based system up to date on both the ShARe/CLEF and NCBI datasets.
CNN-based ranking2: the best deep learning-based system up to date on both the ShARe/CLEF and NCBI datasets. Since we cannot completely reconstructed the KBs as used but not released in Li et al.’s work2, we reimplemented the system and used the same settings as described in their paper. In addition, we employed word2vec14 to train the word embeddings with a dimension size of 50 from all the clinical notes in MIMIC-III19, the PubMed biomedical abstracts as used in Li et al.’s work2, and the drug labels as used in Xu et al.’s work6 for the ShARe/CLEF, NCBI, and TAC2017ADR datasets, respectively.
Table 3 shows the performance comparisons of the state-of-the-art methods with our best fine-tuned BERT-based ranking models for biomedical entity normalization. The table shows that our best BERT-based ranking models consistently outperformed previous methods and achieved the state-of-the-art performance in terms of accuracy by 0.35%, 0.26% and 1.17% on the ShARe/CLEF, NCBI, TAC2017ADR datasets, respectively. Note that, due to we used different KBs, the results of our reimplemented CNN-based ranking on the ShARe/CLEF and NCBI datasets were different from that reported in Li et al.’s work2.
Table 3:
Comparisons with existing work. The bold score denotes the best performance of each dataset.
The impact of different batch sizes
Table 4 shows the impact of different batch sizes on the three datasets. We compared batch sizes of 16 and 32 as suggested by Devlin et al.17. From the table, we observe that (1) For the NCBI and TAC2017ADR datasets, setting batch size as 16 achieved better performance than as 32. For the ShARe/CLEF dataset, there was no obvious difference between different batch size settings. (2) The best performance was achieved when batch size was set as 16 on all the three datasets.
Table 4:
The impact of different batch sizes. The underlined score denotes that the performance of the model with the current batch size was better than the other choice. The bold score denotes the best performance of each dataset.
ShARe/CLEF | NCBI | TAC2017ADR | ||||
---|---|---|---|---|---|---|
batch size | 16 | 32 | 16 | 32 | 16 | 32 |
BERTBase_Cased | 90.56 | 90.62 | 88.85 | 88.65 | 92.62 | 92.56 |
BERTBase_Uncased | 90.56 | 90.58 | 88.65 | 88.13 | 92.97 | 92.65 |
BERTLarge_Cased | 90.73 | 90.71 | 88.85 | 88.33 | 92.42 | 92.87 |
BERTLarge_Uncased | 90.66 | 90.66 | 88.13 | 88.13 | 92.87 | 92.70 |
BioBERTBase_Cased+PubMed | 91.10 | 91.01 | 88.23 | 88.02 | 93.22 | 92.98 |
BioBERTBase_Cased+PMC | 90.81 | 90.99 | 88.65 | 88.65 | 92.97 | 92.89 |
BioBERTBase_Cased+PubMed+PMC | 91.01 | 91.09 | 89.06 | 88.85 | 93.17 | 92.89 |
ClinicalBERTBase_Cased+MIMIC | 90.62 | 90.54 | 88.96 | 88.44 | 92.70 | 92.67 |
ClinicalBERTLarge_Cased+MIMIC | 90.88 | 90.73 | 88.13 | 88.02 | 92.94 | 92.80 |
Discussion
In this study, we developed an entity normalization architecture by fine-tuning the pre-trained BERT / BioBERT / ClinicalBERT models and conducted extensive experiments to evaluate the effectiveness of the pre-trained models for the entity normalization task using biomedical datasets of three different types. Our best fine-tuned models consistently outperformed previous methods and advanced the state-of-the-art on biomedical entity normalization by up to 1.17% increase in accuracy. To the best of our knowledge, this is the first study to apply and evaluate the pre-trained BERT / BioBERT / ClinicalBERT models for biomedical entity normalization.
From Table 2, we notice that although all the best fine-tuned models outperformed BM25 on the three datasets, it did not improve too much on the NCBI dataset (i.e., by up to 0.83%). BERTLarge_Uncased and ClinicalBERTLarge_Cased+MIMIC performed even worse than BM25. This indicates the difficulty of this dataset. Choosing an appropriate pre-trained model for this dataset is necessary. In the future, we will further investigate better methods for this dataset, e.g., tuning different learning rates to find a better fine-tuned model.
The BERT models with cased version were better than that with uncased version in most cases for biomedical entity normalization. This indicates that the BERT models with cased version could capture more precise contextualized word representations than that with uncased version, and they are benefit for the entity normalization task.
The three BioBERT models were initialized with BERTBase_Cased and pre-trained with biomedical corpora34. The two ClinicalBERT models were initialized with BERTBase_Cased and BERTLarge_Cased, and pre-trained with clinical notes from MIMIC-III19. For the ShARe/CLEF and TAC2017ADR datasets, all the three BioBERT models outperformed the BERTBase_Cased model and both the two ClinicalBERT models outperformed the corresponding BERTBase_Cased and BERTLarge_Cased models. For the NCBI dataset, BioBERTBase_Cased+PubMed+PMC and ClinicalBERTBase_Cased+MIMIC were better than BERTBase_Cased. These indicate that the domain-specific BioBERT and ClinicalBERT are more appropriate than BERT for biomedical entity normalization. It would be interesting to pre-train a new bidirectional language representation model from scratch (or initialized with BERTBase or BERTLarge) using a large amount of drug labels from dailymed37 and evaluate their effects on the TAC2017ADR dataset. We plan to conduct these studies in future.
The best performance was achieved when fine-tuning BioBERTBase_Cased+PubMed for both the ShARe/CLEF and TAC2017ADR datasets, and when fine-tuning BioBERTBase_Cased+PubMed+PMC for the NCBI dataset. This indicates that the model (i.e., BioBERTBase_Cased+PubMed+PMC) initialized with BERTBase_Cased and pre-trained with both PubMed abstracts and PubMed Central full-text articles is effective for the NCBI dataset, and the pre-trained model (i.e., BioBERTBase_Cased+PubMed) with only PubMed abstracts is useful for both the ShARe/CLEF and TAC2017ADR datasets as well. This also illustrates that PubMed Central full-text articles are helpful for the PubMed abstracts but not for the clinical text and drug labels.
From Table 3, we notice that our best fine-tuned BERT-based ranking consistently outperformed the CNN-based ranking on all the three datasets, which indicates that pre-trained contextualized word representation models using bidirectional Transformers are more effective than the traditional context-independent word embeddings for the entity normalization task. Although the best fine-tuned models consistently outperformed previous state-of-the-art methods on all the three datasets, the improvements on the ShARe/CLEF and NCBI datasets were 0.35% and 0.26%, which was less than that on the TAC2017ADR dataset (i.e., 1.17%). For the ShARe/CLEF dataset, the main reason may be that we may not have completely reconstructed the ontology used in previous work2–4, which was not released. For the NCBI dataset, the best performance was from TaggerOne (i.e., 88.80) which was reported by Leaman and Lu36. Their model was a joint model, which performed named entity recognition (with gold entity mentions as input) and normalization simultaneously. Such joint models could often leverage more contextual information to achieve better performance36,38. In the future, we will also investigate joint models to further improve entity normalization performance.
At this time, we applied and evaluated the pre-trained BERT / BioBERT / ClinicalBERT models for candidate concept ranking by transforming the ranking task as a sentence-pair classification task, which was a pointwise learning to rank method. We will further investigate pairwise learning to rank methods as used in previous work6,7. We are also planning to introduce the features used in Xu et al.’s system6 into the final classifier layer of the candidate concept ranking module.
Conclusion
In this study, we applied and evaluated pre-trained language representation models for entity normalization using three biomedical datasets of different types. Preliminary results show that fine-tuning the pre-trained language representation models effectively advanced the state-of-the-art for biomedical named entity normalization.
Acknowledgement
This work is supported by NLM 5R01LM010681, NCI U24 CA194215, and NIGMS 5U01TR002062. Part of this work is supported by NVIDIA Corporation with the donation of the Quadro P6000 GPU.
Conflicts of Interest
Dr. Xu and The University of Texas Health Science Center at Houston have research-related financial interests in Melax Technologies, Inc.
Figures & Table
References
- 1.Shen W, Wang J, Han J. Entity Linking with a Knowledge Base: Issues, Techniques, and Solutions. TKDE. 2014;99(1) doi: http://doi.ieeecomputersociety.org/10.1109/TKDE.2014.2327028. [Google Scholar]
- 2.Li H, Chen Q, Tang B, et al. CNN-based ranking for biomedical entity normalization. BMC Bioinformatics. 2017;18(11):385. doi: 10.1186/s12859-017-1805-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.D’Souza J, Ng V. Sieve-Based Entity Linking for the Biomedical Domain. ACL. 2015:297–302. [Google Scholar]
- 4.Ghiasvand O, Kate RJ. UWM: Disorder Mention Extraction from Clinical Text Using CRFs and Normalization Using Learned Edit Distance Patterns. SemEval@COLING. 2014:828–832. [Google Scholar]
- 5.Kang N, Singh B, Afzal Z, van Mulligen EM, Kors JA. Using rule-based natural language processing to improve disease normalization in biomedical text. JAMIA. 2012;20(5):876–881. doi: 10.1136/amiajnl-2012-001173. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Xu J, Lee H-J, Ji Z, Wang J, Wei Q, Xu H. UTH_CCB System for Adverse Drug Reaction Extraction from Drug Labels at TAC-ADR 2017. TAC. 2017 [Google Scholar]
- 7.Leaman R, Doǧan RI, Lu Z. DNorm: disease name normalization with pairwise learning to rank. Bioinformatics. 2013;29:2909–2917. doi: 10.1093/bioinformatics/btt474. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Luo Y, Song G, Li P, Qi Z. Multi-Task Medical Concept Normalization Using Multi-View Convolutional Neural Network. AAAI. 2018;Vol 1:5868–5875. [Google Scholar]
- 9.Bodenreider O. The Unified Medical Language System (UMLS): integrating biomedical terminology. Nucleic Acids Res. 2004;32(suppl_1):D267. doi: 10.1093/nar/gkh061. doi:10.1093/nar/gkh061. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Pradhan S, Elhadad N, Chapman WW, Manandhar S, Savova G. SemEval-2014 Task 7: Analysis of Clinical Text. SemEval. 2014:54–62. [Google Scholar]
- 11.Lee C-P, Lin C-J. Large-scale linear ranksvm. Neural Comput. 2014;26(4):781–817. doi: 10.1162/NECO_a_00571. [DOI] [PubMed] [Google Scholar]
- 12.Roberts K, Demner-Fushman D, Tonning JM. Overview of the TAC 2017 Adverse Reaction Extraction from Drug Labels Track. TAC. 2017 [Google Scholar]
- 13.Luo Y, Song G, Li P, Qi Z. Multi-Task Medical Concept Normalization Using Multi-View Convolutional Neural Network. AAAI. 2018 [Google Scholar]
- 14.Mikolov T, Sutskever I, Chen K, Corrado GS, Dean J. Distributed representations of words and phrases and their compositionality. In: Advances in Neural Information Processing Systems. 2013:3111–3119. [Google Scholar]
- 15.Peters ME, Neumann M, Iyyer M, et al. Deep Contextualized Word Representations. In: NAACL-HLT. 2018:2227–2237. https://aclanthology.info/papers/N18-1202/n18-1202. [Google Scholar]
- 16.Radford A, Narasimhan K, Salimans T, Sutskever I. Improving Language Understanding with Unsupervised Learning. 2018.
- 17.Devlin J, Chang M-W, Lee K, Toutanova K. BERT: Pre-training of deep bidirectional transformers for language understanding. arXiv Prepr arXiv181004805. 2018 [Google Scholar]
- 18.Lee J, Yoon W, Kim S, et al. BioBERT: a pre-trained biomedical language representation model for biomedical text mining. CoRR. 2019 doi: 10.1093/bioinformatics/btz682. abs/1901.0. http://arxiv.org/abs/1901.08746. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Si Y, Wang J, Xu H, Roberts K. Enhancing clinical concept extraction with contextual embeddings. JAMIA. July 2019:ocz096. doi: 10.1093/jamia/ocz096. doi:10.1093/jamia/ocz096. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Huang K, Altosaar J, Ranganath R. ClinicalBERT: Modeling Clinical Notes and Predicting Hospital Readmission. CoRR. 2019 abs/1904.0. http://arxiv.org/abs/1904.05342. [Google Scholar]
- 21.Alsentzer E, Murphy JR, Boag W, et al. Publicly Available Clinical BERT Embeddings. CoRR. 2019 abs/1904.0. http://arxiv.org/abs/1904.03323. [Google Scholar]
- 22.Wei Q, Ji Z, Si Y, et al. Relation Extraction from Clinical Narratives Using Pre-trained Language Models. AMIA. 2019 [PMC free article] [PubMed] [Google Scholar]
- 23.Pradhan S, Elhadad N, South BR, et al. Evaluating the state of the art in disorder recognition and normalization of the clinical narrative. JAMIA. 2015;22(1):143–154. doi: 10.1136/amiajnl-2013-002544. doi:10.1136/amiajnl-2013-002544. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Doǧan RI, Leaman R, Lu Z. NCBI disease corpus: a resource for disease name recognition and concept normalization. JBI. 2014;47:1–10. doi: 10.1016/j.jbi.2013.12.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Davis AP, Wiegers TC, Rosenstein MC, Mattingly CJ. MEDIC: a practical disease vocabulary used at the Comparative Toxicogenomics Database. Database. 2012;2012:bar065. doi: 10.1093/database/bar065. doi:10.1093/database/bar065. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Sohn S, Comeau DC, Kim W, Wilbur WJ. Abbreviation definition identification based on automatic precision estimates. BMC Bioinformatics. 2008;9 doi: 10.1186/1471-2105-9-402. doi:10.1186/1471-2105-9-402. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Schwartz AS, Hearst MA. A Simple Algorithm for Identifying Abbreviation Definitions in Biomedical Text; Proceedings of the 8th Pacific Symposium on Biocomputing; 2003. pp. 451–462. [PubMed] [Google Scholar]
- 28.Soysal E, Wang J, Jiang M, et al. CLAMP - a toolkit for efficiently building customized clinical natural language processing pipelines. JAMIA. 2017;25(3):331–336. doi: 10.1093/jamia/ocx132. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Ji Z, Lu Z, Li H. An Information Retrieval Approach to Short Text Conversation. 2014 http://arxiv.org/abs/14086988. [Google Scholar]
- 30.Xu J, Zhang Y, Wang J, et al. UTH-CCB: The Participation of the SemEval 2015 Challenge - Task 14. SemEval. 2015:311–314. http://www.aclweb.org/anthology/S15-2052. [Google Scholar]
- 31.Zhang Y, Wang J, Tang B, et al. UTH_CCB: a report for semeval 2014 - task 7 analysis of clinical text. SemEval. 2014:802. [Google Scholar]
- 32.Robertson SE, Walker S, Jones S, Hancock-Beaulieu M, Gatford M. Okapi at TREC-3; Proceedings of TREC; 1995. pp. 109–126. [Google Scholar]
- 33.Devlin J, Chang M-W, Lee K, Toutanova K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. CoRR. 2018 abs/1810.0. http://arxiv.org/abs/1810.04805. [Google Scholar]
- 34.Lee J, Yoon W, Kim S, et al. BioBERT: a pre-trained biomedical language representation model for biomedical text mining. arXiv Prepr arXiv190108746. 2019 doi: 10.1093/bioinformatics/btz682. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Johnson AEW, Pollard TJ, Shen L, et al. MIMIC-III, a freely accessible critical care database. Sci data. 2016;3:160035. doi: 10.1038/sdata.2016.35. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Leaman R, Lu Z. TaggerOne: joint named entity recognition and normalization with semi-Markov Models. Bioinformatics. 2016;32(18):2839–2846. doi: 10.1093/bioinformatics/btw343. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Health NI of. DailyMed. 2014 [Google Scholar]
- 38.Ji Z, Sun A, Cong G, Han J. Joint Recognition and Linking of Fine-Grained Locations from Tweets. WWW. 2016:1271–1281. doi:10.1145/2872427.2883067. [Google Scholar]