Abstract
Background Electronic medical records (EMRs) are revolutionizing health-related research. One key issue for study quality is the accurate identification of patients with the condition of interest. Information in EMRs can be entered as structured codes or unstructured free text. The majority of research studies have used only coded parts of EMRs for case-detection, which may bias findings, miss cases, and reduce study quality. This review examines whether incorporating information from text into case-detection algorithms can improve research quality.
Methods A systematic search returned 9659 papers, 67 of which reported on the extraction of information from free text of EMRs with the stated purpose of detecting cases of a named clinical condition. Methods for extracting information from text and the technical accuracy of case-detection algorithms were reviewed.
Results Studies mainly used US hospital-based EMRs, and extracted information from text for 41 conditions using keyword searches, rule-based algorithms, and machine learning methods. There was no clear difference in case-detection algorithm accuracy between rule-based and machine learning methods of extraction. Inclusion of information from text resulted in a significant improvement in algorithm sensitivity and area under the receiver operating characteristic in comparison to codes alone (median sensitivity 78% (codes + text) vs 62% (codes), P = .03; median area under the receiver operating characteristic 95% (codes + text) vs 88% (codes), P = .025).
Conclusions Text in EMRs is accessible, especially with open source information extraction algorithms, and significantly improves case detection when combined with codes. More harmonization of reporting within EMR studies is needed, particularly standardized reporting of algorithm accuracy metrics like positive predictive value (precision) and sensitivity (recall).
Keywords: electronic health records, review, text mining, data quality, case detection
INTRODUCTION
Information recorded in electronic medical records (EMRs), clinical reports, and summaries has the possibility of revolutionizing health-related research. EMR data can be used for disease registries, epidemiological studies, drug safety surveillance, clinical trials, and healthcare audits.
Information recording in EMRs
In most EMRs there is the possibility for the clinician both to code their findings in a structured format and also to enter information in narrative free text. There are various nomenclatures for structuring or coding information; the most widely used are International Classification of Diseases version 10, 1 Systematized Nomenclature of Medicine – Clinical Terms, 2 and the International Classification of Primary Care. 3 Within multi-modal EMRs there are also laboratory, pathology, and radiology reports, admission and discharge summaries, and chief complaints fields, which are in unstructured or semi-structured text. The balance of recording by the clinician, between codes and narrative text, is likely to vary by institution, EMR system, department, disease type, and component of the record.
Why do EMRs contain free text instead of being completely structured?
Clinicians experience a tension between choosing to code information and expressing it in text. 4 Among the main motivators for clinicians to code rather than use text is the increased ease of search, access, and retrieval. 5,6 A coded record allows the clinician to readily demonstrate that appropriate care has been provided, accurate diagnoses are made, and targets met. 7 This is especially important for billing after episodes of care, or for incentive based systems such as the National Health Service (NHS) Quality and Outcomes Framework in UK primary care. 8
Coded data can be analyzed and summarized easily and on a large scale, whereas free text cannot. In contrast to structured data, narrative text is highly variable, 9 but is more engaging, captures the patient’s narrative, can be told from different perspectives, and allows expression of feelings. 10 It is a better reminder for the clinician of the human encounter. 7
Additionally, clinicians have given a number of reasons why they find coding onerous; the choices available in coded data may be too limiting, and may not allow for the expression of nuances. 11 The process of finding and entering codes on the computer represents an additional cognitive load, 5 and may take longer than summarizing the consultation in text. 6 Free text may be chosen when no code precisely describes clinical findings, or when there is a need to give supporting evidence for a diagnosis or suspicion. 12 Clinicians use free text as a pragmatic solution to recording vague diagnoses or strange collections of symptoms, when diagnoses need qualification, and for psycho-social problems. 7 Text can summarize processes of deduction, and modal language can be used to convey a range of possible outcomes. Codes do not easily accommodate diagnostic uncertainty, so a patient may be labeled with a diagnosis prematurely or incorrectly. Similarly, a clinician may have a range of possible differential diagnoses, but only code the one that supports the choice of treatment. 7
Why case-detection is central to EMR research
One key quality issue in research using data from EMRs is the precision of case-detection. Studies have shown that classification errors in the case identification process can considerably bias study findings. 13 If cases of the disease of interest are not well defined, then the conclusions drawn from the study will be of poor quality. Case-detection algorithms are created from several structured pieces of information, such as sets of diagnostic and prescription codes; existing examples include dementia, 14 stroke, 15 diabetes, 13,16 depression, 17 hypertension, 18 and rheumatoid arthritis. 19–23
To date, research using EMRs has mainly relied on coded information to define cases. Abstraction and analysis of the coded information is straightforward in comparison to abstraction of the text, which also requires anonymizing and annotating. As yet we have little understanding of how much information, and what type, is contained within unstructured sections of the record, and therefore how biases may arise from ignoring the content of the text. Adding in text may markedly improve rates and accuracy of case-detection when using EMRs for research. UK studies have shown that our understanding of the date of diagnosis, 24 and the number of symptoms prior to diagnosis 12,25 can change substantially when information extracted from free text is added to the coded information.
Methods for extracting information from text
The volume of EMRs available means that human review of text is too time-consuming and labor intensive to be achievable in most studies. However, the automation of extraction of information from text makes the clinical information contained therein more accessible. Natural language processing (NLP) is a subfield of computer science concerned with intelligent processing of human language. For over 50 years computer scientists have developed algorithms to analyze natural language text, using either sets of hand-written rules or machine learning techniques. 26 However, adapting such algorithms to medical text has proved difficult, for two main reasons: 1) patient privacy and confidentiality issues, which create difficulties in obtaining suitable data to develop and test algorithms on 27,28 and 2) the nuances of medical text, which make it difficult to obtain reliable clinical results using standard processing techniques. 29,30 The majority of tools for analysis of text are trained on edited text genres such as newspaper articles or scientific papers. 31 While medical discharge summaries, diagnostic test reports, and letters may be written in standard English, consultation notes are hastily written, and do not go through an editing process. These notes are terse, with a telegraphic style and limited use of full sentence syntax; in particular, sentential subjects are very rare, and even finite verbs are uncommon. 31 Standard NLP tools make many errors when applied to clinical notes. It has often been necessary for a new NLP tool to be developed or adapted for each medical database, and even for each clinical question, when processing EMR free text. This is labor intensive, as it requires the tools to be tested on significant amounts of text already annotated by human experts.
Aims
It is not clear how successful researchers have been in incorporating information extracted from EMR text into their case-detection algorithms, or how much of an improvement the addition of this information gives in comparison to codes alone. In the present study we aimed to review information extraction from EMR text for the stated purpose of case-detection for named clinical conditions.
In particular, we aimed to 1) systematically describe the methods of information extraction from text, 2) evaluate the current technical accuracy of information extraction algorithms, and 3) understand the additional benefits of using text for case-detection rather than structured data alone.
METHODS
Systematic search
Searches were conducted between July 2014 and July 2015 on PubMed and Web of Science (WoS), using search terms derived from Medical Subject Headings vocabulary (US National Library of Medicine): 1) “electronic health records” or “electronic medical records” or “electronic patient records” or “hospital records” or “personal health records” or “computerized patient records” or “computerized medical records” or “automated medical records” combined with 2) “free text” or “narrative” or “text mining” or “natural language processing.” No date constraints were placed on papers retrieved. These searches returned 641 articles from PubMed and 9018 from WoS, of which 6975 were duplications within WoS search results or between WoS and PubMed results. Following review of titles and abstracts, 249 papers were retained to examine in full text.
Eligibility
To be eligible for this review, published research had to meet all of the following four criteria:
Primary research with full text published in English.
Information extracted from the text of EMR, medical letter, or medical report by any method.
Information extracted from text for stated clinical condition.
Stated purpose of information extraction was case-detection.
Exclusion of papers
Papers were excluded in two stages; a title/abstract review (2735 excluded) and then a full text review of 249 papers (182 excluded). In total 67 papers met the eligibility criteria ( Figure 1 ). Major reasons for rejection were because papers focused on:
Describing or comparing EMR systems.
Generating text from EMR data.
Problem list or decision support development, clinical interventions delivered through EMRs.
Acceptability, satisfaction, barriers, or facilitators with EMR systems.
Technical details of information extraction with no stated clinical condition.
Cause of injury or event detected rather than clinical condition.
Extraction of information from text for purposes other than case-detection.
Extraction of characteristics of a defined population, rather than pure case-detection.
Text used only for case validation not detection.
Extraction of information from studies
The full text of all studies was scrutinized and details were abstracted into a table ( Supplementary Data ). Wilcoxon signed rank tests were performed to compare extracted values of median accuracy of algorithms between studies, using IBM SPSS statistics 22.
Assessment of algorithms
The most rigorous method for assessing the accuracy of an algorithm is to compare its results against a gold standard. Most studies reported in this review assessed performance by means of manual review, unless noted otherwise. Method of assessment was not an eligibility criterion for inclusion of a study; studies were included even if they reported no assessment of algorithms. If any of the following measures was stated in the study it was extracted and reported here, and studies reporting any of these measures were included in the technical accuracy section of the results. Measures included:
1) Sensitivity (Recall) measures the proportion of actual positives that are correctly identified as such (e.g., the percentage of sick people who are correctly identified as having the condition).
2) Specificity measures the proportion of negatives that are correctly identified as such (e.g., the percentage of healthy people who are correctly identified as not having the condition), and is complementary to the false positive rate (1 – False Positive Rate).
3) Positive predictive value (precision) and negative predictive value (PPV and NPV, respectively) are the proportions of positive and negative results in tests that are true positive and true negative results. These values are dependent on the prevalence of the condition in the population, so a low prevalence condition may give rise to a low PPV despite high sensitivity and specificity.
- 4) F-measure : In informatics, the positive predictive value is called precision, and sensitivity is called recall. The F-score can be used as a single measure of performance of the test and is the harmonic mean of precision and recall:
5) Area under the ROC (AUROC) : The trade-off between sensitivity and specificity can be represented graphically as a receiver operating characteristic (ROC). The ROC curve illustrates the performance of a binary classifier system as its discrimination threshold is varied. When using normalized units, the area under the curve (often referred to as simply the AUC, or AUROC) is equal to the probability that a classifier will rank a randomly chosen positive instance higher than a randomly chosen negative one.
RESULTS
The 67 studies included in this review were published between 2000 and 2015, with the majority from 2010 to 2015 (41 studies, 61%). The majority of studies used data that originated in the United States (US) (and were conducted by US teams; 57 studies, 85%). Data from the Netherlands was used in eight studies (12%; including two that incorporated data from Italy and Denmark). One study was conducted using data from Canada (1%), and one using data from Sweden (1%).
Reasons for case-detection
The majority of studies (87%) gave a reason for wanting to detect cases from medical records. The most common reasons were to use them in further medical records research (29 studies, 43%). Other studies stated the purpose was for epidemic surveillance of infectious diseases (12 studies, 18%); for surveillance of indicators of cancer, diabetes, or hospital acquired infection to assist prevention (4 studies, 6%); for estimation of incidence of conditions in the population (5 studies, 7%); or for clinical trial recruitment (5 studies, 7%). Three studies stated they were seeking improvements in clinical decision-making (4%), and one study was populating a cancer registry (1%). Nine studies (13%) did not identify the purpose of case-detection.
Conditions studied
Forty-one conditions were studied in the 67 studies, and four studies each ascertained two conditions. Conditions could be divided into four categories: chronic or noncommunicable diseases, infectious diseases, psychological disorders, and injuries or events ( Table 1 ).
Table 1:
Type of condition | No. of studies | Conditions included ( N studies if >1) |
---|---|---|
Chronic or noncommunicable conditions | 42 (59%) | Obesity (7), cancer (4), rheumatoid and psoriatic arthritis (5), diabetes (3), inflammatory bowel disease (incl. celiac) (3), asthma (3), COPD (2), pancreatic cyst (2), heart failure (2) hypertension, angina pectoris, atrial fibrillation, disorders of sex development, multiple sclerosis, hepatobiliary disease, cataract, priapism, facial pain, peripheral arterial disease, coronary artery disease |
Infectious diseases | 18 (25%) | Acute respiratory infection (2), pneumonia (4) influenza or influenza-like illness (5) MRSA (2), gastrointestinal infection, genital chlamydia, chicken pox, fever, hospital acquired urinary tract infection |
Psychological disorders | 4 (6%) | Depression (2), binge eating disorder, bipolar disorder |
Injuries and events | 7 (10%) | Venous thromboembolism (2) acute myocardial infarction, upper GI bleeding, ischemic stroke, acute renal failure, acute orbital fracture |
Types of medical records
Thirty-seven studies drew on multiple sections of hospital EMRs, such as codes, prescriptions, laboratory or pathology reports, and clinical notes (55%). Other studies used a focused part of the hospital EMR: nine studies (13%) used hospital discharge summaries, five studies (7%) used imaging reports (X-ray or CT scans), three (4%) used the narrative portion of emergency department records, two (3%) used laboratory reports only, and one study used pathology reports (1%). Ten studies (15%) used primary care records that contained a mixture of structured fields (codes and prescriptions) and free text.
Information extraction from text
There were three main types of information extraction: keyword search, rule-based algorithm, and machine learning algorithms. Sixteen studies (24%) used only a keyword search to extract information. Forty-five studies (67%) reported a rule-based NLP algorithm to extract information from text. An algorithm was categorized as rule-based if it combined a keyword search with any negation or context modifying module, although many algorithms were more sophisticated than this. Six studies (9%) used machine learning, Bayesian, or hybrid (rule-based + machine learning) approaches.
Several information extraction algorithms were used in more than one study. Studies used established NLP algorithms such as MedLEE (9 studies), 32,33 HITEx (4 studies), 34 cTAKES (5 studies), 35 Unstructured information management architecture (3 studies), 36–38 Topaz (2 studies), 39,40 Regenstrief extraction tool (REX; 2 studies), 37,41 and the KnowledgeMap concept identifier (2 studies). 42,43 Keyword search tools reported in more than one study included EMERSE (2 studies) 44 and the Unified Medical Language System (UMLS) search tool (2 studies). The most common structured output format of algorithms was the National Library of Medicine UMLS Metathesaurus of Concept Unique Identifiers, 45 which was used in 23 studies. NLP algorithms also output to the Systematized Nomenclature of Medicine Clinical Terms, Medical Subject Headings, and Hospital International Classification of Disease Adaptation codes.
Context modifiers and negation were assessed by several add-on algorithms, notably NegEx (5 studies), 46 and ConText (2 studies). 47 Medication information was extracted using MedEx (2 studies), 48 which produced RxNorm encoded medications, and FreePharma NLP (1 study). 49
Case-detection algorithms (CDA)
After information from text was extracted, there were several different methods for reaching ascertainment of cases. Three studies manually reviewed the results of a keyword search, and four studies considered the presence of a single code or keyword to be sufficient for a case. The remaining 60 studies (90%) used an algorithm to detect cases. In 15 studies (23%) the same algorithm performed NLP and detected cases. In 16 studies (24%), a new algorithm was used to combine outputs of NLP using only textual information. In 29 studies (43%), the information from text was combined with codes, lab results, or medications to detect cases, using rule based, logistic regression, Bayesian, or machine learning models. The breakdown of algorithm types is shown in Table 2 .
Table 2:
Type of case-detection | No. of studies (%) | Detail |
---|---|---|
No additional algorithm (manual review of information) | 3 (4) | |
Single keyword or code sufficient to define case | 4 (6) | |
Same NLP algorithm as extracted info also detected cases (text only) | 15 (23) | |
New rule-based CDA (text only) | 11 (16) | |
Logistic regression or machine learning CDA (text only) | 5 (4) | Logistic regression 50 ; decision tree 51 ; Bayesian network vs rule-based 39 ; naïve Bayes vs perception neural network 52 ; naïve Bayes 53 |
New rule-based CDA (combining text with codes, labs, or medication) | 12 (18) | |
Logistic regression CDA (combining text with codes, labs or medication) | 8 (12) | |
Machine learning algorithm (combining text with codes, labs, or medication) | 6 (9) | Ripper 54 |
Support vector machines (SVM) 55 | ||
Decision tree, vs SVM vs Ripper vs metacost 56 | ||
Naïve Bayes vs SVM vs random forest vs logistic regression 57 | ||
Bayesian network model vs EM-MAP model 40 | ||
Random forest 58 | ||
Comparison of rule based CDA with machine learning and logistic regression CDAs (combining text with codes, labs, or medication) | 3 (4) | Rule based vs SVM vs random forest vs Ripper vs logistic regression 59 |
Rule based vs logistic regression 60 | ||
Rule based vs decision tree 61 |
Technical Accuracy
Table 3 summarizes technical accuracy by type of case-detection algorithm and by medical condition. Fifty-six studies reported accuracy metrics for their algorithms and were grouped into three sets: 15 studies reported no additional CDA on top of the NLP algorithm extracting information from text ( Supplementary Table A ); 20 studies reported secondary rule-based case detection algorithms using combinations of different sources of text, or combining text, codes and medication ( Supplementary Table B ); and 21 studies reported probabilistic case detection algorithms (regression, Bayesian, or machine learning) combining different sources of text, or combining text, codes, and medication ( Supplementary Table C ). In order to ascertain whether technical accuracy was influenced by the type of condition, the median accuracy of algorithms is shown for conditions that are examined in more than one study (full details in Supplementary Table D with references). Table 3 shows no clear pattern of difference in accuracy by type of algorithm, nor much variability in performance by condition, with the exception of obesity, the ascertainment for which had lower than average performance, and for which the majority of studies were using a single source of data (hospital discharge letters in the i2b2 challenge 62 ).
Table 3:
No. of Studies | Sensitivity (Recall) | Specificity | PPV (Precision) | Negative predictive value | F measure | AUROC | |
---|---|---|---|---|---|---|---|
Algorithm type | |||||||
Single algorithm for NLP and case detection | 15 | 96.2 | 97.4 | 85.35 | 96.6 | 49 | – |
Rule-based secondary case detection algorithm | 20 | 91.2 | 95.45 | 77.5 | 98.95 | 97.57 | 94.4 |
Probabilistic secondary case detection algorithm (Logistic Regression; Bayesian; machine learning) | 21 | 80 | 95 | 86 | 95.4 | 77 | 94 |
Condition | |||||||
Respiratory infections | 11 | 92.9 | 95.45 | 54 | 99.9 | – | 95.85 |
Bowel disease | 4 | 79.45 | 94.45 | 57.5 | 100 | – | 87.5 |
Inflammatory arthritis | 5 | 70 | 96 | 93.7 | – | – | 94.4 |
Cancer | 3 | 93 | 92.9 | 95 | – | 93.5 | – |
Diabetes | 2 | 96.2 | 98 | – | – | 98.65 | – |
Obesity | 5 | 48.4 | – | 76.3 | – | 49 | – |
Mental health | 3 | 73.1 | 90 | 87.85 | 96.6 | – | 80 |
MRSA | 2 | 99.2 | 99.4 | 97.9 | – | 99 | – |
Cardiovascular | 7 | 82 | 96 | 84.7 | 93 | 74.85 | 92.9 |
Additional benefit of information extraction from text
The main benefit of extracting information from text was that case-detection was significantly improved. Table 4 shows selected accuracy metrics for 19 studies that reported direct comparisons of case-detection algorithms using codes only, text only, and/or a combination of codes and text. Medians were significantly higher in code/text-combined algorithms compared to codes-only algorithms for sensitivity (recall) ( P = .028) and AUROC ( P = .025), but not for PPV (precision) ( P = .066). There were no significant differences between accuracy of algorithms using codes only and text only.
Table 4:
Codes only
|
Text only
|
Combination of codes + text
|
||||||||
---|---|---|---|---|---|---|---|---|---|---|
Study | Condition | Sensitivity (Recall) | PPV (precision) | AUROC | Sensitivity (Recall) | PPV (precision) | AUROC | Sensitivity (Recall) | PPV (Precision) | AUROC |
Gundlapalli (2008) 68 | Inflammatory bowel disease | 27 | 50 | 64 | 86 | 43 | 90 | 100 | 40 | 99 |
Graiser (2007) 69 | Lymphoma | 42.9 | 90.0 | 81.2 | ||||||
Valkhoff (2014) 70 | Upper GI bleed | |||||||||
ICD-9 (ARS) | 72 | |||||||||
ICD-9 (HSD) | 78 | |||||||||
ICD-10 (Aarhus) | 77 | 47 | ||||||||
ICPC codes | 21 | 22 | ||||||||
DeLisle (2013) 71 | Pneumonia | 52 | 52.8 | 74.8 | 63.6 | |||||
Li (2008) 63 | Ischemic stroke | 90 | 56 | |||||||
Ludvigsson (2014) 72 | Celiac | 53.8 | 78.1 | |||||||
Pakhomov (2007) 53 | Angina | 88 | 88 | |||||||
Ananthakrishnan (2013) 66 | Inflammatory bowel disease: Crohn’s | 89 | 95 | |||||||
Ulcerative colitis | 86 | 94 | ||||||||
Carroll (2012) 42 | Rheumatoid arthritis | 49 | 80 | 88 | 71 | 86 | 97 | |||
Liao (2010) 73 | Rheumatoid arthritis | 51 | 88 | 56 | 89 | 63 | 94 | |||
Xia (2014) 74 | Multiple sclerosis | 76.4 | 91.6 | 93.7 | 75.8 | 91.4 | 94.1 | 82.7 | 92.1 | 95.8 |
DeLisle (2010) 75 | Acute respiratory infection | 79 | 31.5 | 88 | 88 | 18 | 94 | 73 | 52 | 86 |
Zheng (2014) 76 | Acute respiratory infection | 79 | 31 | 78 | 88 | 18 | 90 | 75 | 49 | 87 |
Carroll (2011) 55 | Rheumatoid arthritis | 78.1 | 93.2 | 95.5 | 68.8 | 91.8 | 89.5 | 85.8 | 93.7 | 96.6 |
Karnik (2012) 57 | Atrial fibrillation | 61.7 | 59.8 | 62.7 | 58 | 60 | 60 | |||
Castro (2015) 60 | Bipolar disorder | 79 | 85 | |||||||
McPeek (2013) 43 | Venous thromboembolism | 69 | 95 | 90 | ||||||
Wu (2013) 61 | Asthma | 30.8 | 57.1 | 84.6 | 88.0 | |||||
Zeng 2006 77 | Asthma and COPD | 72.5 | 82.3 | 76.7 | 82.3 | 92.4 | 87.4 | |||
Median | 61.7 | 72.0 | 88.0 | 78.1 | 73.0 | 90.0 | 78.1 | 86.0 | 95.4 |
Some other studies reported other improvements in case finding with the addition of text. Friedlin et al. 41 reported that their NLP algorithm accurately identified three times as many methicillin-resistant Staphylococcus aureus (MRSA) positive blood cultures as their current electronic laboratory reporting system. Li et al. 63 reported that of 2609 cases detected, MedLEE found 1253 (48%) that were not retrieved by searching International Classification of Diseases version-9 codes. A further five studies reported an increase in the number of cases found by using text, including for cancer, 64 hypertension, 65 inflammatory bowel disease, 66 ischemic stroke, 63 and disorders of sex development in children. 67 These studies reported a statistically significant increase in cases ( P = .003), 64 a 7–12% increase in cases, 65,66 226 patients being found using keyword search compared to 14 with manual search, 67 and 702 more patients found using text than with codes alone. 63
DISCUSSION
This review of extracting information from the text of EMRs for case-detection has shown that text can contribute to case-detection of a wide range of conditions including infectious diseases, noncommunicable diseases, and acute events, as well as psychological conditions. However, differences in accuracy of case-detection using information from text compared to codes alone are not always reported explicitly or in a useful form.
The eligible studies suggest that the majority of work so far has been conducted in the United States; very few other countries are represented. The majority of data sources used in these studies were full multi-modal electronic hospital record systems and parts of these records, such as discharge summaries or pathology reports. The source of information is important to note as it affects the portability of the method of information extraction. Documents such as reports and clinical notes use a terse, telegraphic style where the grammatical rules of standard English are discarded in favor of concise information presentation, and where the recipient or reader already has good knowledge of common abbreviations and contractions. Discharge summaries or letters may use more standard English structures and therefore algorithms developed for non-medical text sources may be suitable.
The technical accuracy of algorithms extracting information from text, or combining text information with codes, was generally good but with some variability. Because of the wide range of possible measures of algorithm accuracy, many algorithms were not directly comparable to one another. Many studies reported algorithms with sensitivity and specificity (and related values) of over 90%. Different methods of information extraction were reported, ranging from manual review of records to both rule-based algorithms and probabilistic or statistically driven models using machine learning methods. No particular type of algorithm stood out as particularly better than any other. Accuracy also varied by condition, but no clear pattern was evident.
Some studies reported statistically on the additional benefit to case-detection of extracting information from text compared to codes alone and found there was a significant improvement in case-detection accuracy by incorporating information extracted from text. Given that only 19 studies reported these comparisons, with a large proportion of missing data due to inconsistencies in reporting, this finding needs to be confirmed in a larger pool of studies.
What are the future directions for information extraction from EMR text?
There is no consensus in the literature of what is “good enough” for case-detection models or how much error is acceptable when ascertaining cases. If these algorithms were to be used for identifying patients for clinical trials, or for estimating service needs, a high standard of accuracy would be required. While sensitivities and specificities over 95% sound impressive, if we are looking for cases of a disease with a 1% prevalence, using a case-detection algorithm with a 98% sensitivity and a 97% specificity, the probability that a patient identified as a case really is a case (PPV) is only 25%, because of the high number of false positives. 78 Work is needed to understand better what constitutes appropriate and safe standards for identifying patients or outcomes for research by these methods.
Additionally, there appear to be two cultures of reporting: The consensus within the field of informatics is to report the measures precision , recall , and F-measure , whereas in medicine, the practice is usually to use sensitivity and specificity . Researchers in biomedical informatics understand that sensitivity is equal to recall and positive predictive value is equal to precision. Specificity on the other hand is not used in informatics outside the biomedical domain. To make it easier to compare results and draw conclusions from them, the two cultures must become more integrated. The aim should be for more standardized ways of reporting the accuracy of both information extraction and case-detection algorithms.
These studies used many different algorithms for information extraction from text, and in around half of studies, algorithms were specific to the individual study. Such algorithms take significant human effort and time to develop, requiring domain expertise, programming skills, and iterative evaluation and development. 42 Re-using existing algorithms and nomenclatures minimizes effort and ensures comparability with other studies. MedLEE was utilized across nine studies, HITEx in five studies, and CTAKES in seven. Two studies reported specifically on the portability of an NLP algorithm keyword search tool, 42,67 suggesting the porting was successful with minor moderations to the algorithms. Future research could also investigate which approaches give the best chance of portability of CDAs to different settings, conditions, and purposes.
Another approach for reducing the effort associated with extracting information from text would be to develop generalizable estimates of context effects. Some studies reported keyword searches, which do not require complex algorithms, and can be a cheap and quick method of extracting information from text. However, these searches pick up all incidences of keywords, not taking account of negation, uncertainty, or other contextual effects. If we were able to estimate that the influence of context effects or modifiers were small, we would be reassured that keyword searches were an adequate and pragmatic approach to extracting information from text.
Achieving anonymity or de-identification is another barrier to the use of text from EMRs. In general, data protection regulations state that only de-identified data can be released to researchers without the patient’s explicit consent. De-identification of structured records is fairly straightforward, but anonymizing free text is a much more difficult task, as patient identifiers may be located in any part of the text. Algorithms that automate the process of de-identification of text have been developed and are reviewed elsewhere. 28,79 If these algorithms perform well enough, they could be run at source—for example, within the clinical institution where identifiers are not a problem—and anonymize the text before EMRs are extracted for secondary purposes. A set of standards for safe and secure de-identification to protect patient privacy is needed, therefore, so that the accuracy of de-identification algorithms can be compared against these standards.
Strengths and limitations of the current study
This study identified a good range of published papers on extraction of information from text in EMRs. We used two sensitive databases that covered both medical and informatics fields to pick up as many articles as possible. Once studies were retrieved from the search they were then scrutinized and chosen in a rigorous fashion. This means the selection process was likely to have favored specificity over sensitivity in terms of studies meeting eligibility criteria, but whether this will have affected our conclusions is not clear. It may have reduced the power to find differences between types of algorithms. Additionally, only studies published in English were used. We are aware that some NLP groups also publish in French and German, so future work may seek to incorporate these studies by searching in other languages.
As this is a wide literature, it was not possible to also include studies reporting on extracting characteristics of defined populations, although the methodologies used in these studies would have considerably overlapped with the studies reported. A further review may want to scrutinize other reasons for information extraction from EMR text, such as medication usage and adverse events, or the quality of care given. It is possible that text may contain more valuable information for some research purposes than others, and so the value of extracting information from text should be reviewed for a range of purposes and compared.
One limitation of the literature identified in the current study was the small number of studies explicitly comparing algorithms containing information extracted from text to other algorithms using structured data only—less than one-third of studies identified for this review. If consensus on reporting within studies can be achieved, especially within those comparing case detection methods, this review could usefully be repeated in a few years’ time and further results ascertained to support the inclusion of information from text in EMR research.
Conclusions
A wide range of studies showed that information extracted from EMR text has been used to identify varied conditions with variable degrees of success. Most of the research has so far come from research groups in the United States using hospital-based EMRs. There is likely to be benefit gained from adding information extracted from text to case-detection algorithms in terms of improved sensitivity and specificity, although numbers of studies are too small to make firm conclusions. There is no standardization in the reporting of the performance of the algorithms, which makes comparison of studies difficult. Researchers in the field would benefit from more standardized reporting of algorithm performance, such as always reporting sensitivity (recall) and PPV (precision), and from working towards making information extraction methods and their outputs more compatible and comparable between studies.
FUNDING
This work was supported in part by the Wellcome Trust, grant number 086105/Z/08/Z.
COMPETING INTERESTS
The authors have no competing interests to declare.
Supplementary Material
ACKNOWLEDGEMENTS
We gratefully acknowledge the contribution of Dr Tim Williams, Director of Research at Clinical Practice Research Datalink, London, UK, who read and commented on copies of the manuscript.
SUPPLEMENTARY MATERIAL
Supplementary material is available online at http://jamia.oxfordjournals.org/ .
CONTRIBUTORS
E.F., J.C., and J.C. conceived and designed the study;. E.F., H.S., J.C., J.C., and D.S. analyzed and interpreted the data; E.F., H.S., J.C., J.C., and D.S. drafted and revised the manuscript; and E.F., J.C., H.S., D.S., and J.C. approved the final manuscript.
REFERENCES
- 1. World Health Organization. International statistical classification of diseases and related health problems 10th revision, edition 2010 . Geneva, Switzerland, 2010.
- 2. International Health Terminology Standards Development Organisation (IHTSDO). SNOMED CT Starter Guide . Copenhagen, Denmark. Accessed from: www.snomed.org/starterguide.pdf . 2014 .
- 3. Soler J-K, Okkes I, Wood M, Lamberts H . The coming of age of ICPC: celebrating the 21st birthday of the International Classification of Primary Care . Fam Pract. 2008. ; 25 ( 4 ): 312 – 317 . [DOI] [PubMed] [Google Scholar]
- 4. Lovis C, Baud RH, Planche P . Power of expression in the electronic patient record: structured data or narrative text? Int J Med Inform. 2000. ; 58-59:101-110 . [DOI] [PubMed] [Google Scholar]
- 5. Walsh SH . The clinician's perspective on electronic health records and how they can affect patient care . Br Med J. 2004. ; 328 : 1184 – 1187 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6. Powsner SM, Wyatt JC, Wright P . Opportunities for and challenges of computerisation . Lancet. 1998. ; 352 : 1617 – 1622 . [DOI] [PubMed] [Google Scholar]
- 7. de Lusignan S, Wells SE, Hague NJ, Thiru K . Managers see the problems associated with coding clinical data as a technical issue whilst clinicians also see cultural barriers . Methods Inf Med. 2003. ; 42 : 416 – 422 . [PubMed] [Google Scholar]
- 8. NHS England. Quality and Outcomes Framework guidance for GMS contract 2013/14. London, 2013.
- 9. Tange HJ, Hasman A, de Vries Robbe PF, Schouten HC . Medical narratives in electronic medical records . Int J Med Inform. 1997. ; 46 ( 1 ): 7 – 29 . [DOI] [PubMed] [Google Scholar]
- 10. Greenhalgh T, Hurwitz B . Narrative based medicine: why study narrative . Br Med J. 1999. ; 318 ( 7175 ): 48 – 50 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11. Stein HD, Nadkarni P, Erdos J, Miller PL . Exploring the degree of concordance of coded and textual data in answering clinical queries from a clinical data repository . J Am Med Inform Assoc. 2000. ; 7 ( 1 ): 42 – 54 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12. Ford E, Nicholson A, Koeling R, et al. . Optimising the use of electronic health records to estimate the incidence of rheumatoid arthritis in primary care: what information is hidden in free text? BMC Med Res Methodol . 2013 ; 13 ( 1 ): 1 – 12 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13. Manuel DG, Rosella LC, Stukel TA . Importance of accurately identifying disease in studies using electronic health records . Br Med J. 2010. ; 341 : c4226 . [DOI] [PubMed] [Google Scholar]
- 14. Imfeld P, Pernus YBB, Jick SS, Meier CR . Epidemiology, co-morbidities, and medication use of patients with alzheimer's disease or vascular dementia in the UK . J Alzheimers Dis. 2013. ; 35 : 565 – 573 . [DOI] [PubMed] [Google Scholar]
- 15. Gulliford MC, Charlton J, Ashworth M, Rudd A, Toschke AM , The eCRT research team. Selection of medical diagnostic codes for analysis of electronic patient records. Application to stroke in a primary care database . PLoS ONE. 2009. ; 4 ( 9 ): e7168 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16. Currie CJ, Poole CD, Evans M, Peters JR, Morgan CL . Mortality and other important diabetes-related outcomes with insulin vs other antihyperglycemic therapies in type 2 diabetes . J Clin Endocrinol Metab. 2013. ; 98 ( 2 ): 668 – 677 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17. Rait G, Walters K, Griffin M, Buszewicz M, Nazareth I . Recent trends in the incidence of recorded depression and depressve symptoms in primary care . Br J Psych. 2009. ; 195 ( 6 ): 520 – 524 . [DOI] [PubMed] [Google Scholar]
- 18. Ryan R, Majeed A . Prevalence of treated hypertension in general practice in England and Wales, 1994 to 1998 . Health Stat Q. 2002. ( 16 ): 14 – 18 . [Google Scholar]
- 19. Chung CP, Rohan P, Krishnaswami S, McPheeters ML . A systematic review of validated methods for identifying patients with rheumatoid arthritis using administrative or claims data . Vaccine. 2013. ; 31 : K41 – K61 . [DOI] [PubMed] [Google Scholar]
- 20. Lin C, Karlson EW, Canhao H, et al. . Automatic prediction of rheumatoid arthritis disease activity from the electronic medical records . PLoS One. 2013. ; 8 ( 8 ): e69932 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21. Nielen MMJ, Ursum J, Schellevis FG, Korevaar JC . The validity of the diagnosis of inflammatory arthritis in a large population based primary care database . BMC Fam Pract. 2013. ; 14 : 79 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22. Chen Y, Carroll RJ, Hinz ERM, et al. . Applying active learning to high-throughput phenotyping algorithms for electronic health records data . J Am Med Inform Assoc. 2013. ; 20 ( E2 ): E253 – E259 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23. Widdifield J, Bombardier C, Bernatsky S, et al. . An administrative data validation study of the accuracy of algorithms for identifying rheumatoid arthritis: the influence of the reference standard on algorithm performance . BMC Musculoskelet Disord. 2014. ; 15 ( 1 ): 216 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24. Tate AR, Martin AGR, Murray-Thomas T, Anderson SR, Cassell JA . Determining the date of diagnosis - is it a simple matter? The impact of different approaches to dating diagnosis on estimates of delayed care for ovarian cancer in UK primary care . BMC Med Res Methodol. 2009. ; 9 ( 1 ): 42 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25. Koeling R, Tate AR, Carroll JA . Automatically estimating the incidence of symptoms recorded in GP free text notes . Proceedings of the First International Workshop on Managing Interoperability and Complexity in Health Systems. ACM; 2011: 43-50 . [Google Scholar]
- 26. Manning CD, Schütze H . Foundations of Statistical Natural Language Processing . MIT press; , Cambridge MA: ; 1999. . [Google Scholar]
- 27. Dorr DA, Phillips WF, Phansalkar S, Sims SA, Hurdle JF . Assessing the difficulty and time cost of de-identification in clinical narratives . Methods Inform Med. 2006. ; 45 ( 3 ): 246 – 252 . [PubMed] [Google Scholar]
- 28. Meystre SM, Friedlin FJ, South BR, Shen S, Samore MH . Automatic de-identification of textual documents in the electronic health record: a review of recent research . BMC Med Res Methodol. 2010. ; 10 : 70 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29. Kalra D, Ingram D . Electronic health records. Information Technology Solutions for Healthcare . Springer, London; 2006: 135-181 . [Google Scholar]
- 30. Resnik P, Niv M, Nossal M, Kapit A, Toren R . Communication of clinically relevant information in electronic health records: a comparison between structured data and unrestricted physician language . Perspectives in Health Information Management, CAC Proceedings . 2008. ; Fall . [Google Scholar]
- 31. Carroll J, Koeling R, Puri S . Lexical acquisition for clinical text mining using distributional similarity . Computational Linguistics and Intelligent Text Processing. Springer, Berlin Heidelberg; 2012: 232-246 . [Google Scholar]
- 32. Friedman C, Alderson PO, Austin JH, Cimino JJ, Johnson SB . A general natural-language text processor for clinical radiology . J Am Med Inform Assoc. 1994. ; 1 ( 2 ): 161 – 174 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33. Friedman C, Shagina L, Lussier Y, Hripcsak G . Automated encoding of clinical documents based on natural language processing . J Am Med Inform Assoc. 2004. ; 11 ( 5 ): 392 – 402 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34. Zeng QT, Goryachev S, Weiss S, Sordo M, Murphy SN, Lazarus R . Extracting principal diagnosis, co-morbidity and smoking status for asthma research: evaluation of a natural language processing system . BMC Med Inform Decis Mak. 2006. ; 6 ( 1 ): 30 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35. Savova GK, Masanz JJ, Ogren PV, et al. . Mayo clinical Text Analysis and Knowledge Extraction System (cTAKES): architecture, component evaluation and applications . J Am Med Inform Assoc. 2010. ; 17 ( 5 ): 507 – 513 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36. Jones M, DuVall SL, Spuhl J, Samore MH, Nielson C, Rubin M . Identification of methicillin-resistant Staphylococcus aureus within the Nation's Veterans Affairs Medical Centers using natural language processing . BMC Med Inform Decis Mak. 2012. ; 12 ( 1 ): 34 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37. Mehrabi S, Schmidt CM, Waters JA, et al. . An efficient pancreatic cyst identification methodology using natural language processing . In: Lehmann CU, Ammenwerth E, Nohr C , eds. Medinfo 2013: Proceedings of the 14th World Congress on Medical and Health Informatics, Pts 1 and 2 . 2013. : 192 : 822 – 826 . [PubMed] [Google Scholar]
- 38. Roch AM, Mehrabi S, Krishnan A, et al. . Automated pancreatic cyst screening using natural language processing: a new tool in the early detection of pancreatic cancer . HPB. 2015. ; 17 ( 5 ): 447 – 453 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39. Ye Y, Tsui F, Wagner M, Espino JU, Li Q . Influenza detection from emergency department reports using natural language processing and Bayesian network classifiers . J Am Med Inform Assoc. 2014. ; 21 ( 5 ): 815 – 823 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40. Tsui F, Wagner M, Cooper G, et al. . Probabilistic case detection for disease surveillance using data in electronic medical records . Online J Public Health Inform. 2011. ; 3 ( 3 ). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41. Friedlin J, Grannis S, Overhage JM . Using natural language processing to improve accuracy of automated notifiable disease reporting . AMIA Annu Symp Proc. 2008. : 207 – 211 . [PMC free article] [PubMed] [Google Scholar]
- 42. Carroll RJ, Thompson WK, Eyler AE, et al. . Portability of an algorithm to identify rheumatoid arthritis in electronic health records . J Am Med Inform Assoc. 2012. ; 19 ( e1 ): e162 – e169 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43. McPeek Hinz ER, Bastarache L, Denny JC . A natural language processing algorithm to define a venous thromboembolism phenotype . AMIA Annu Symp Proc. 2013. ; 2013 : 975 – 983 . [PMC free article] [PubMed] [Google Scholar]
- 44. Hanauer DA . EMERSE: The Electronic Medical Record Search Engine . AMIA Annu Symp Proc. 2006. : 941 . [PMC free article] [PubMed] [Google Scholar]
- 45. Lindberg DA, Humphreys BL, McCray AT . The Unified Medical Language System . Methods Inf Med. 1993. ; 32 ( 4 ): 281 – 291 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46. Chapman WW, Bridewell W, Hanbury P, Cooper GF, Buchanan BG . A simple algorithm for identifying negated findings and diseases in discharge summaries . J Biomed Inform. 2001. ; 34 ( 5 ): 301 – 310 . [DOI] [PubMed] [Google Scholar]
- 47. Harkema H, Dowling JN, Thornblade T, Chapman WW . ConText: an algorithm for determining negation, experiencer, and temporal status from clinical reports . J Biomed Inform. 2009. ; 42 ( 5 ): 839 – 851 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48. Xu H, Stenner SP, Doan S, Johnson KB, Waitman LR, Denny JC . MedEx: a medication information extraction system for clinical narratives . J Am Med Inform Assoc. 2010. ; 17 ( 1 ): 19 – 24 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49. Wilke RA, Berg RL, Peissig P, et al. . Use of an electronic medical record for the identification of research subjects with diabetes mellitus . Clin Med Res. 2007. ; 5 ( 1 ): 1 – 7 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50. Cano C, Blanco A, Peshkin L . Automated identification of diagnosis and co-morbidity in clinical records . Methods Inf Med. 2009. ; 48 ( 6 ): 546 – 551 . [DOI] [PubMed] [Google Scholar]
- 51. Yadav K, Sarioglu E, Smith M, Choi H-A . Automated outcome classification of emergency department computed tomography imaging reports . Acad Emerg Med. 2013. ; 20 ( 8 ): 848 – 854 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52. Pakhomov SV, Buntrock J, Chute CG . Prospective recruitment of patients with congestive heart failure using an ad-hoc binary classifier . J Biomed Inform. 2005. ; 38 ( 2 ): 145 – 153 . [DOI] [PubMed] [Google Scholar]
- 53. Pakhomov S, Weston SA, Jacobsen SJ, Chute CG, Meverden R, Roger VL . Electronic medical records for clinical research: application to the identification of heart failure . Am J Manag Care. 2007. ; 13 ( 6 Part 1 ): 281 – 288 . [PubMed] [Google Scholar]
- 54. Afzal Z, Schuemie MJ, van Blijderveen JC, Sen EF, Sturkenboom MCJM, Kors JA . Improving sensitivity of machine learning methods for automated case identification from free-text electronic medical records . BMC Med Inform Decis Mak. 2013. ; 13 : 30. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55. Carroll RJ, Eyler AE, Denny JC . Naïve electronic health record phenotype identification for rheumatoid arthritis . AMIA Annu Symp Proc. 2011. ; 2011 : 189 – 196 . [PMC free article] [PubMed] [Google Scholar]
- 56. Afzal Z, Engelkes M, Verhamme KM, et al. . Automatic generation of case-detection algorithms to identify children with asthma from large electronic health record databases . Pharmacoepidemiol Drug Saf. 2013. ; 22 ( 8 ): 826 – 833 . [DOI] [PubMed] [Google Scholar]
- 57. Karnik S, Tan SL, Berg B, et al. . Predicting atrial fibrillation and flutter using electronic health records . 2012 Annual International Conference of the IEEE Engineering in Medicine and Biology Society. 2012. ; 2012 : 5562 – 5565 . [DOI] [PubMed] [Google Scholar]
- 58. Love TJ, Cai T, Karlson EW . Validation of psoriatic arthritis diagnoses in electronic medical records using natural language processing . Semin Arthritis Rheum. 2011. ; 40 ( 5 ): 413 – 420 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59. Xu H, Fu Z, Shah A, et al. . Extracting and integrating data from entire electronic health records for detecting colorectal cancer cases . AMIA Annu Symp Proc. 2011. ; 2011 : 1564 . [PMC free article] [PubMed] [Google Scholar]
- 60. Castro VM, Minnier J, Murphy SN, et al. . Validation of electronic health record phenotyping of bipolar disorder cases and controls . Am J Psychiatry. 2015. ; 172 ( 4 ): 363 – 372 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61. Wu ST, Sohn S, Ravikumar KE, et al. . Automated chart review for asthma cohort identification using natural language processing: an exploratory study . Ann Allergy Asthma Immunol. 2013. ; 111 ( 5 ): 364 – 369 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62. Uzuner O . Second i2b2 workshop on natural language processing challenges for clinical records . AMIA Annu Symp Proc. 2007. ; 2007 : 1252 – 1253 . [PubMed] [Google Scholar]
- 63. Li L, Chase HS, Patel CO, Friedman C, Weng C . Comparing ICD9-encoded diagnoses and NLP-processed discharge summaries for clinical trials pre-screening: a case study . AMIA Annu Symp Proc. 2008. : 404 – 408 . [PMC free article] [PubMed] [Google Scholar]
- 64. Hanauer DA, Miela G, Chinnaiyan AM, Chang AE, Blayney DW . The registry case finding engine: an automated tool to identify cancer cases from unstructured, free-text pathology reports and clinical notes . J Am Coll Surg. 2007. ; 205 ( 5 ): 690 – 697 . [DOI] [PubMed] [Google Scholar]
- 65. Baus A, Hendryx M, Pollard C . Identifying patients with hypertension: a case for auditing electronic health record data . Perspect Health Inf Manag. 2012. ; 9 : 1e . [PMC free article] [PubMed] [Google Scholar]
- 66. Ananthakrishnan AN, Cai T, Savova G, et al. . Improving case definition of Crohn's disease and ulcerative colitis in electronic medical records using natural language processing: a novel informatics approach . Inflamm Bowel Dis. 2013. ; 19 ( 7 ): 1411 – 1420 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67. Hanauer DA, Gardner M, Sandberg DE . Unbiased identification of patients with disorders of sex development . PLoS One. 2014. ; 9 ( 9 ): e108702 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 68. Gundlapalli AV, South BR, Phansalkar S, et al. . Application of natural language processing to VA electronic health records to identify phenotypic characteristics for clinical and research purposes . Summit on Translat Bioinforma. 2008. : 36 – 40 . [PMC free article] [PubMed] [Google Scholar]
- 69. Graiser M, Moore SG, Victor R, et al. . Development of query strategies to identify a histologic lymphoma subtype in a large linked database system . Cancer Inform. 2007. ; 3 : 149 – 158 . [PMC free article] [PubMed] [Google Scholar]
- 70. Valkhoff VE, Coloma PM, Masclee GMC, et al. . Validation study in four health-care databases: upper gastrointestinal bleeding misclassification affects precision but not magnitude of drug-related upper gastrointestinal bleeding risk . J Clin Epidemiol. 2014. ; 67 ( 8 ): 921 – 931 . [DOI] [PubMed] [Google Scholar]
- 71. DeLisle S, Kim B, Deepak J, et al. . Using the electronic medical record to identify community-acquired pneumonia: toward a replicable automated strategy . PLoS One. 2013. ; 8 ( 8 ): e70944 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 72. Ludvigsson JF, Pathak J, Murphy S, et al. . Use of computerized algorithm to identify individuals in need of testing for celiac disease . J Am Med Inform Assoc. 2013. ; 20 ( e2 ): e306 – e310 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 73. Liao KP, Cai T, Gainer V, et al. . Electronic medical records for discovery research in rheumatoid arthritis . Arthritis Care Res. 2010. ; 62 ( 8 ): 1120 – 1127 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 74. Xia Z, Secor E, Chibnik LB, et al. . Modeling disease severity in multiple sclerosis using electronic health records . PLoS One. 2013. ; 8 ( 11 ): e78927 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 75. DeLisle S, South B, Anthony JA, et al. . Combining free text and structured electronic medical record entries to detect acute respiratory infections . PLoS One. 2010. ; 5 ( 10 ): e13377 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 76. Zheng H, Gaff H, Smith G, DeLisle S . Epidemic surveillance using an electronic medical record: an empiric approach to performance improvement . PLoS One. 2014. ; 9 ( 7 ): e100845 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 77. Zeng QT, Goryachev S, Weiss S, Sordo M, Murphy SN, Lazarus R . Extracting principal diagnosis, co-morbidity and smoking status for asthma research: evaluation of a natural language processing system . BMC Med Inform Decis Mak. 2006. ; 6 : 30 . [DOI] [PMC free article] [PubMed] [Google Scholar]
- 78. Stone JV . Bayes' Rule: A Tutorial Introduction to Bayesian Analysis . Sebtel Press; ; Sheffield, United Kingdom: , 2013. . [Google Scholar]
- 79. Kushida CA, Nichols DA, Jadrnicek R, Miller R, Walsh JK, Griffin K . Strategies for de-identification and anonymization of electronic health record data for use in multicenter research studies . Med Care. 2012. ; 50 ( Suppl ): S82 – S101 . [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.