Skip to main content
Scientific Reports logoLink to Scientific Reports
. 2018 Apr 20;8:6329. doi: 10.1038/s41598-018-24389-w

Clinical Assistant Diagnosis for Electronic Medical Record Based on Convolutional Neural Network

Zhongliang Yang 1,2, Yongfeng Huang 1,2,, Yiran Jiang 3, Yuxi Sun 3, Yu-Jin Zhang 1, Pengcheng Luo 4
PMCID: PMC5910396  PMID: 29679019

Abstract

Automatically extracting useful information from electronic medical records along with conducting disease diagnoses is a promising task for both clinical decision support(CDS) and neural language processing(NLP). Most of the existing systems are based on artificially constructed knowledge bases, and then auxiliary diagnosis is done by rule matching. In this study, we present a clinical intelligent decision approach based on Convolutional Neural Networks(CNN), which can automatically extract high-level semantic information of electronic medical records and then perform automatic diagnosis without artificial construction of rules or knowledge bases. We use collected 18,590 copies of the real-world clinical electronic medical records to train and test the proposed model. Experimental results show that the proposed model can achieve 98.67% accuracy and 96.02% recall, which strongly supports that using convolutional neural network to automatically learn high-level semantic features of electronic medical records and then conduct assist diagnosis is feasible and effective.

Introduction

Automatically extracting useful information from electronic medical records (EMRs) with implementing apropriate diagnosis is one ultimate goal of intelligent medical construction1. It is such a meaningful and promising task that it can not only effectively improve working efficiency but also reduce misdiagnosis rate of doctors for making a diagnosis24. At the same time it can help us to better understand the clinical manifestations of various diseases58, and even the relation between the various diseases9,10. Previous works show that such models can even sometimes outperforming experienced doctors11 in improving teaching practice12,13 and assisting diagnosis1417.

However, it is extremely challenging for the following reasons: firstly, medical records contain complex reports of patients which include patient’s statements, vital signs, history of treatment, history of allergies and so on. The varieties of which make it arduous to filter and represent complicated information. Secondly, it is possible that different patients with the same diseases may have disparate performances or symptoms, which makes it difficult to establish a unified rule to predict the possibility of incidence and analyze information of diseases. Thirdly, notes and reports in EMRs are likely to be written by different doctors from diverse departments. Thus, even if some words or records seem to have the same meaning, there would still exist dissimilar writing expressions due to different writing habits, which increases the difficulty of identifying the actual symtoms as well as reaching an accurate conclusion.

Most of existing platforms are rule-based methods, which can also be called expert systems. Expert systems are designed to solve complex problems by reasoning through bodies of knowledge, represented mainly as hand-crafted rules. They carry out rule matching on each input electronic medical record in order to chase down the disease which fits these diagnosis rules best and make a diagnose for the disease. This kind of methods have made great achievements in the field of medical auxiliary diagnosis1820. These models hope to imitate the logical reasoning process in the diagnosis procedure of a doctor which makes its diagnosis more logical. However, it weakens the effects in viable utilization owing to some deficiencies of the model itself. Firstly, it cannot solve the problem like semantic ambiguity. Hence, it should contain all potential descriptions as far as possible while setting up the knowledge base, which may cause redundancies in the knowledge base and result in low efficiencies. For example, it was estimated that about fifteen person-years were spent building the Internist-1/QMR knowledge base for internal medicine21. It may also lead to decline in matching accuracy because of not considering all possible conditions. Secondly, a variety of hospitals and departments encounter a wide range of cases which may have great disparities. It is fairly complicated and adverse to manage and maintain the knowledge base constructed by hundreds and thousands of diseases from such various departments, which leads to low efficiencies.

High level semantic understanding for medical record texts has always been hard because of its high coding degree22,23. In recent years, with the development of natural language processing, there has been an increasing number of auxiliary diagnostic methods based on semantic analysis algorithm16,22,24. These kinds of methods try to conduct a high level semantic understanding on EMRs, which mainly draw on natural language processing related technology22. They hope to help the computer better understand the semantics of electronic medical records, and then make a diagnosis accordingly. The ultimate goal they hope to achieve is far from easy to achieve. During these years, with the extensive adoption of deep neural network technology in the field of natural language processing, the application of deep neural network on semantic understanding with analyzing texts has become a popular research23,25,26.

To achieve the ultimate goal, in this study, we applied a multi-layer convolutional neural network for high level semantic understanding for electronic medical records, which can then be used for disease diagnoses. In the past few years, convolutional neural network has made notable progress in fields such as computer vision2729 and natural language processing23,26. The incremental advancement of CNN is likely to benefit the development of new technology and inventions in other fields. A large number of researches and applications have shown that the convolutional neural network has a powerful ability in feature extractions and expressions27,30, which does not require hand-designed features but carries out self-learning through plenty of data. Previous studies have shown that neural network can represent the words in the texts into a dense vector through learning and mapping them into a continuous vector space3135. In this vector space, semantically similar words are distributed in the same region33. Thus, even if the two sections of the text are not the same, as long as the expressions are of the same meaning, they will have similar mathematical expressions, reflected in the semantic space very close32. This can greatly alleviate the problem of semantic ambiguity, and is more efficient than the model based on knowledge base. So we don’t need to build a large number of complex rules or knowledge base to guide how the model decides, but the model itself can automatically extract useful information from the electronic medical records by self-learning, and then conduct disease diagnoses based on these information. This makes our model lighter and more efficient than the knowledge base-based model. The overall framework of our model is shown in Fig. 1. The input of our model is an electronic medical record and the output is the probability of diseases we predicted.

Figure 1.

Figure 1

The overall framework of the proposed model. We use the convolutional neural network to extract the semantic feature vectors of unstructured electronic medical records and map them to the feature space, finally we use the classifier to calculate the probable probability of each disease and select the highest probability of the disease as the auxiliary diagnosis of our model.

Results

Data Preparing

To promote the development of the related fields, in this study, we collected and released a large real-world electronic medical records dataset (C-EMRs) collected from Huangshi Central Hospital in China. It has a total number of 18,590 EMRs and contains the most common diseases of each department, which are Hypertension, Diabetes, Chronic Obstructive Pulmonary Disease (COPD), Gout, Arrhythmia, Asthma, Gastritis, Stomach Polyps. After expunging personal information, each electronic medical record includes thirteen items: chief complaint, physical examination, history of present illness and so forth. Each electronic medical record corresponds to a result of doctor’s diagnosis, which will be used as the label for each EMR samples during the training process. Due to the possibility of a patient with multiple diseases, it is possible that the two electronic medical records have the same content, but the diagnostic results are different. In our dataset, there are altogether 447 patients whose situations are consistent with what is mentioned above. The number and proportion of each disease are shown in Fig. 2.

Figure 2.

Figure 2

The number and proportion of each disease in C-EMRs.

The electronic medical records number of different disease in C-EMRs are imbalance. For diabetes, there are 5642 medical records, but for gout there are only 657. In order to avoid biases and to ensure that there is enough training data, we choose the diseases that has more than 1,000 records to form the training set, which are hypertension, diabetes, COPD, arrhythmia, asthma and gastritis. Also in order to prevent that the training dataset has too much biases, we randomly selected almost the same number records of these diseases as training and testing data. So finally the training data for our model are 7000 EMRs, and another 400 EMRs for testing, which are distributed as Table 1.

Table 1.

The number of electronic medical records of each disease in the training set and test set, and the percentage of test data relative to training data of each disease.

Hypertension Diabetes COPD Arrhythmia Asthma Gastritis Total
Trianing set 1250 1350 1250 1200 1000 950 7000
Test set 68 68 68 62 69 65 400

Experiment Results

We use stochastic gradient descent with momentum 0.9 to train parameters of our network. Our model can quickly converge during the training processing, training after about 20 epochs (one epoch means that all the training samples finish one training session) can reach a steady state with high accuracy and the loss curve is very smooth, which can be seen in Fig. 3(a). From Fig. 3(b) we can see that the prediction time of each electronic medical record is mainly between 10 and 20 milliseconds, which can be predicted in real time.

Figure 3.

Figure 3

The processing of training. (a) Shows the accuracy of train/test set and the loss of the training set varies with the number of epochs. (b) Shows the prediction time of each EMR in the test set.

In Table 2, the Precision, Recall, F1-score and Accuracy of four machine learning algorithms, which are Support Vector Machine (SVM), Multinomial Naïve Bayes(MultinomialNB), Logistic Regression and k-NearestNeighbor, as well as our proposed model are reported. These four machine learning algorithms that we compared to have been applied to the auxiliary diagnosis of electronic medical records in some previous related works and have achieved good results3638. From the results shown in Table 2 we can see that our model has achieved the best effect on each evaluation method. On the test set, our model achieves a 98.67% accuracy and a recall rate of 96.02%, which strongly proves that CNN do have stronger capability of information extraction from texts than other algorithms. Table 3 shows the average prediction time of different methods for each EMR in test set. From Table 3 we can see, the average diagnostic time of our model for each electronic medical record is only 13.82 milliseconds, which indicates that our model can be very efficient in the diagnosis process.

Table 2.

Results of different methods, where “CNN” indicates the performance of the proposed model.

Method Training set Testing set
Precision Recall F1-score Accuracy Precision Recall F1-score Accuracy
SVM 0.96 0.95 0.96 0.9549 0.93 0.93 0.93 0.9315
MultinomialNB 0.93 0.92 0.92 0.9236 0.87 0.86 0.86 0.8600
LogisticRegression 0.93 0.93 0.93 0.9293 0.92 0.92 0.92 0.9175
KNeighborsClassifier 0.89 0.89 0.89 0.8911 0.90 0.89 0.89 0.8925
CNN 0.9947 0.9946 0.9946 0.9982 0.9594 0.9602 0.9596 0.9867

Table 3.

The average prediction time of different methods for each EMR in test set, where “CNN” indicates the performance of the proposed model.

Method SVM MultinomialNB LogisticRegression KNeighborsClassifier CNN
Time(ms) 180.5 ± 2.92 172.5 ± 2.55 167.5 ± 1.58 205.0 ± 1.0 13.82 ± 1.83

As we have mentioned before, our model can automatically extract high-level semantic features from electronic medical records and map them to a high-dimensional feature space (usually hundreds to thousands of dimensions). We can use t-Distributed Stochastic Neighbor Embedding (t-SNE)39,40 technique for the dimensionality reduction and visualization of these high-dimensional feature vectors, which can be find in Fig. 4. In this feature space, each point represents an electronic medical record and different colors indicate different diseases. At the beginning of training (Epoch = 0), since the model parameters are randomly initialized, all the electronic medical records in the feature space are randomly distributed and indivisible. After 5 epoch, electronic medical records of different diseases began to have a trend of separation. After 10 epoch, the electronic medical records of all kinds of diseases have been separated, except for some areas and the edge of each category. When the training reaches 100 epoch, we can clearly see that the samples of each disease have been completely separated, and the electronic medical records of the same disease are also gathered together. After training, electronic medical records belong to the same kind of disease distribution in the same area. Considering that some patients may also suffer from a variety of diseases, there will be individual records mixed with other categories. For each inputted electronic medical record, we mapped it to the feature space, and by analyzing its location distribution in the feature space, we can calculate the possibility of which disease it belongs to.

Figure 4.

Figure 4

The change of feature space with the training process. In this feature space, each point represents an electronic medical record and different colors indicate different diseases. At the beginning of training (Epoch = 0), since the model parameters are randomly initialized, all the electronic medical records in the feature space are randomly distributed and indivisible. After 5 epoch, electronic medical records of different diseases began to have a trend of separation. After 10 epoch, the electronic medical records of all kinds of diseases have been separated, except for some areas and the edge of each category. When the training reaches 100 epoch, we can clearly see that the samples of each disease have been completely separated, and the electronic medical records of the same disease are also gathered together.

Discussion

Automatic extraction of useful information in electronic medical records is of great significance and value for the study of clinical treatment and related diseases1,57. The current clinical diagnosis model or system is mostly based on the large-scale medical knowledge base of human construction18,41,42. Through the association extraction of electronic medical records and the rule matching with the knowledge base, the electronic medical records are analyzed and the clinical auxiliary diagnosis is provided. This kind of method is usually of heavy workload21 and the actual effect is not very satisfactory. In this study, we propose a method of information extraction and analysis of electronic medical records using convolutional neural network, and finally conduct clinical auxiliary diagnosis. Comparing with other machine learning algorithms, our model is proved superior to other algorithms on various metrics (Table 2). The high precision (95.94%) achieved by our model means that the probability of misdiagnosis of our model is very low, which is extremely important for practical use. At the same time, the recall (96.02%) of our model is also high, which means that the probability of missed diagnosis is extremely low in our model. Combined with these test results, we can find that our model has significantly practical value.

It is worth noting that our model does not require human building large scale knowledge bases and complex rules, since all the model parameters and features are automatically learned from a large number of historical electronic medical records, which makes our model quite lightweight and fairly practical. At the same time, our model is very efficient, through testing we found that the average prediction time of each electronic medical record is 13.82 milliseconds, which outperforms other machine learning methods (SVM: 180.5 ms, MultinomialNB: 172.5 ms, LogisticRegression: 167.5 ms, KNeighborsClassifier: 205.0 ms), which has been shown in Table 3.

These results strongly support that it is feasible and effective to use the convolutional neural network to automatically learn high-level semantic features of electronic medical records and then conduct assist diagnosis. Based on these advantages, our model can effectively improve the clinical diagnostic efficiency of doctors. At the same time, because our model is affirmed by a large number of historical diagnostic medical records, it can effectively reduce the possibility of misdiagnosis.

As the results shown in Fig. 3, our model can effectively extract high-level semantic features of electronic medical records and map them into high-dimensional feature space. In this feature space, electronic medical records of different diseases have different distribution, and the electronic medical records of the same disease are gathered together. By analyzing this feature space (we can also call it “disease space”), we may even be able to help clinicians better understand the relation between various diseases and what tendencies are likely to occur in the same disease, which would be one of the most promising aspects of the proposed model. We hope that our research will not only help clinicians better make clinical diagnosis, on the other hand, we hope to help clinicians further understand the various clinical diseases from another perspective.

Although we have made gratifying achievements, we still should consider some limitations on the current exploratory reseach. Firstly, the sample types used for model training and testing are not enough as we used only several most common and medically different diseases. Therefore, in the future research, we’ll try to do more research, including more types of diseases, and more similar diseases, such as diabetes I and II. Secondly, in this study, we only consider three main contents of electronic medical records: chief complaint, history of present illnesses and physical examination. Although these three items are likely the most crucial ones, other contents corresponding to this record are equally important. In the future research, we will take into account comprehensive contents in the electronic medical record, even including other diagnostic information, such as medical images.

In summary, the major contributions of this paper are as follows. Firstly, we designed and implemented an auxiliary diagnosis model for electronic medical records based on convolution neural network. We hope this model can not only effectively improve working efficiency but also reduce misdiagnosis rate of doctors for making a diagnosis. Since our model can conduct high level semantic understanding of the electronic medical records, we also hope that it can help doctors to better understand the clinical manifestations of various diseases, and even the relation between the various diseases. Secondly, in order to promote the development of the related fields, we collected and released a large real-world electronic medical records dataset (C-EMRs). It has a total number of 18,590 EMRs and contains the most common diseases of each department, which are Hypertension, Diabetes, Chronic Obstructive Pulmonary Disease (COPD), Gout, Arrhythmia, Asthma, Gastritis, Stomach Polyps. Thirdly, we tested and evaluated the proposed auxiliary diagnosis model for electronic medical records based on CNN on this dataset. The test results show that the method has high diagnostic efficiency (13.82 milliseconds costs for each EMR prediction) and diagnostic accuracy (acc: 98.67%, recall: 96.02%). Although our model still has space for further improvement, it has shown significant and practical value for clinical research. We hope that our work will serve as a guide for future related work and help promote the further development of the auxiliary diagnosis of electronic medical records.

Method

Model Structure and Analysis

In this study, we propose a method using convolutional neural network to extract features from electronic medical records and conduct disease prediction. The input of the proposed model is an electronic medical record and the output is the prediction probability of diseases. The final structure of the convolutional neural network used in this study is as follows: an embedding layer, a convolutional layer with three different sizes of convolutional kernels, an average pooling layer and a fully connected layer following with a softmax classification. The embedding layer transforms the inputted EMR text into a two-dimensional matrix form which is suitable for the processing of convolution. The convolutional layer is used to extract features from the input matrix and convolution kernels of different sizes can learn different context related features. The pooling layer is served for down sampling the features, which can enhance the robustness of the model and significantly influence the performance27,43. The purpose of the fully connected layer is to fuse all these features and pass them to the softmax classifier for disease prediction. The softmax classifier, whose parameters have been learned during the training process, calculates the correlation between the input feature vector and the various diseases, and finally concludes the probability value of each disease. The practical parameters setting will be given in the next section “Experiment Setting”.

For each of the structured medical records inputted, we first make it unstructured by connecting each of its contents to form a whole passage. For each passage S, we illustrate it with a matrix XN×D, as shown in Equation (1), where the i-th row indicates the i-th word in passage S, each word is represented as a D-dimension vertor which is randomly initialized, that is

X=[Word1Word2WordN]=[x1,1x1,2x1,Dx2,1x2,2x2,DxN,1xN,2xN,D] 1

Generally, let Xi:j refer to the matrix which consists of the words vectors from the i-th word to the j-th word, that is:

Xi:j=[WordiWordi+1Wordj]=[xi,1xi,2xi,Dxi+1,1xi+1,2xi+1,Dxj,1xj,2xj,D] 2

The convolution layer contains convolution kernels of multiple sizes, and each size contains multiple number of convolution kernels. The width of each convolution kernel is the same as the width of the input matrix. Suppose that the height of the k-th convolution kernel is H, the convolutional kernel can be expressed as WkH×D, that is

Wk=[w1,1kw1,2kw1,Dkw2,1kw2,2kw2,DkwH,1kwH,2kwH,Dk] 3

Convolution operation is a feature extraction process for the elements in the local region of the input matrix. For example, when w1,1k and x1,1 coincide, then the feature c1k extracted from X1:H by the convolutional kernel can be:

c1k=f(i=1Hj=1Dwi,jkxi,j+bi,jk), 4

where the weight wi,jk denotes the importance of the j-th value in the i-th word vector, bi,jk is the bias term and f is a nonlinear function, here we follow previous works27 and use ReLu function as our nonlinear function, which is defined as

y=ReLu(x)=max(0,x). 5

The convolution process is that the convolution kernel Wk slids from the top to the bottom on the input matrix X with a certain step Tc, and calculates the features of each local region. Finally, the document feature extracted by convolution kernel Wk is:

Ck=[c1k,c2k,,cNH+1Tck]Τ. 6

The pooling layer can reduce the number of neural network parameters while maintaining the overall distribution of the data, which can effectively prevent the model over-fitting and improve the robustness of the model27,43. The pooling operation is very similar to the convolution operation, the only difference is that it only calculates the average or maximum value of the local area. We conduct a max pooling operation after each convolution operation on the feature Ck, suppose the height of a pooling kernel is Hp and the step size is Tp, then the output is:

Mk=[m1k,m2k,,mNpk]Τ, 7

where

mik=max(cik,ci+1k,,ci+Hp1k). 8
Np=NH+1TcHp+1Tp, 9

The process described above is a process in which one convolution kernel Wk produces one feature Mk. After all the convolution and pooling operations have been completed, all the extracted features are concatenated end to end to obtain the feature vector of the entire EMR, which can be indicated as

FΤ=[F1Τ;F2Τ;;FlΤ], 10

where Fi = Mi, l indicates the number of the features.

Fully connected layer is used to further blend features and extract higher-level features. By defining a weight matrix WF, we compute the weighted sum of each feature element and obtain the final feature representation of the inputted text S:

y=WFF+bf, 11

where WF and bf are learned weight matrix and bias, the values in weight matrix WF reflect the importance of each feature. The dimension of the output vector y is L, which is the number of labels. In our realization, L is the number of diseases required to be predicted. We then pass the vector y through the softmax classifier to get the predicted probability of each disease:

Pi=exp(yi)j=1Lexp(yj), 12

where Pi indicates the prevalence of the i-th disease corresponding to the input medical records.

In the process of training, we update network parameters through applying backpropagation algorithm, and the loss function of the whole network consists of two parts, one is the error term and the other is the regularization term, which can be described as:

LOSS=num(PT)Τ(PT)+WF2=numi=0L(PiTi)2+WF2, 13

where num is the batch size of EMRs, P indicates the output of the classifier, and each element as Pi represents the prevalence of the i-th disease. T is the target value which corresponds to the doctor’s diagnosis result. For instance, if this medical records corresponds to the t-th disease, the value of the t-th element in vector T is 1, and the remaining values are 0. The error term in the loss function calculates the mean square error (MSE) between the prediction vector and the actual label. We hope that through the self-learning of the model, the mean square error gets smaller and smaller, that is, the prediction results are getting closer to the real values. In order to strengthen the regularization and prevent overfitting, we adopted the dropout mechanism and a constraint on l2-norms of the weight vectors during the training process. Dropout mechanism means that in the training process of deep learning network, the neural network unit is temporarily discarded from the network, i.e. set to zero, according to a certain probability. This mechanism has been proved to effectively prevent neural network overfitting, and significantly improve the model’s performance27,44,45.

We train our model by minimizing the LOSS function over a batch size number of samples. We use stochastic gradient descent with momentum 0.9 to train the parameters of our network. The update rule for weight w is:

wi+1=wi+αViλLwi|Di, 14

where i is the iteration idex, α ∈ (0, 1] is the momentum factor, V is the momentum variable, λ is the learning rate, and Lwi|Di is the average over the i-th batch Di of the derivative of the LOSS function with respect to w, evaluated at wi.

Experiment Setting

After removing the patient’s private information from EMRs, combined with the doctor’s advice, we mainly used chief complaint, history of present illnesses and physical examination in EMRs as our input text by just connecting each of its contents to form a whole passage. Then we made use of the most commonly used Chinese text segmentation tool, which is Jieba46, to do the word segmentation on the input texts. After that, we built the dictionary based on the dataset and counted the length of each passage. We only consider the words that appear more than five times and the others will be remarked as a character “〈unk〉”. So finally we get 17,274 unique words in our dictionary.

Since our model requires the input matrix be of a certain size, that is, the length of the input text should be constant. We design multiple sets of comparision experiments to choose the best value of this super-parameters. According to the experiment results, we finally fix each input electronic medical record text into 130 words. Less than 130 words will be padding with zero, and more than 130 words will be discarded.

For the input of our model, we map each word to a vector of 300 dimensions which are randomly initialized, so the dimension of input matrix will be 130 × 300. The width of the convolutional kernel is the same as the input matrix, thus 300. However, the height of the convolutional kernel is not fixed, we set the kernel heights to be 4, 5, 6 by comparing the results of different kernel sizes, and each of the different heights has 128 convolution kernels. The dimension of the feature extracted for each EMR is 3 × 128 = 384 and the dimension of the output vector is six, corresponding to six diseases that require diagnosis, so the weight matrix WF of fully connected layer would be WF384×6.

Evaluation Method

We use several evaluation indicators commonly used in classification tasks to evaluate the performance of our model, which are precision, recall, F1-score and accuracy. Their conceptions and formulas are described as follows:

  • Precision measures the proportion of positive samples in the classified samples.
    Precision=TPTP+FP. 15
  • Recall measures the proportion of positives that are correctly identified as such.
    Recall=TPTP+FN. 16
  • F1-score is a measure of a test’s accuracy. It considers both the precision and the recall of the test. The F1 score is the harmonic average of the precision and recall, where an F1 score reaches its best value at 1 and worst at 0.
    F1score=2×Precision×RecallPrecision+Recall. 17
  • Accu3racy measures the proportion of true results (both true positives and true negatives) among the total number of cases examined

Accuracy=TP+TNTP+FN+FP+TN. 18

where TP (True Positive) represents the number of positive samples that are predicted to be positive by the model, FP (False Positive) indicates the number of negative samples predicted to be positive, FN (False Negative) illustrates the number of positive samples predicted to be negative and TN (True Negative) represents the number of negative samples predicted to be negative.

Data availability

The dataset analysed during the current study is available in the Github repository, https://github.com/YangzlTHU/C-EMRs.

Acknowledgements

This research is supported by the National Natural Science Foundation of China (No. U1536201, No. U1536207 and No. U1636113).

Author Contributions

Project design and implementation were conceived by Z.Y., Y.J., Y.S. and Y.H. Data collection was performed by Z.Y. and P.L. Data analysis was performed by Z.Y., Y.H. and Y.Z. The experiments and programming were performed by Z.Y. and Y.J. Manuscript drafting and editing was performed by Y.S. and Z.Y. revised the manuscript. All authors reviewed the manuscript in its final form.

Competing Interests

The authors declare no competing interests.

Footnotes

Publisher's note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Jensen PB, Jensen LJ, Brunak S. Mining electronic health records: towards better research applications and clinical care. Nat. Rev. Genet. 2012;13:395. doi: 10.1038/nrg3208. [DOI] [PubMed] [Google Scholar]
  • 2.Stewart WF, Shah NR, Selna MJ, Paulus RA, Walker JM. Bridging the inferential gap: The electronic health record and clinical evidence. Heal. Aff. 2007;26:w181–91. doi: 10.1377/hlthaff.26.2.w181. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Hillestad R, et al. Can electronic medical record systems transform health care? Potential health benefits, savings, and costs. Heal. Aff. 2005;24:1103. doi: 10.1377/hlthaff.24.5.1103. [DOI] [PubMed] [Google Scholar]
  • 4.Groopman, J. E. How doctors think (Houghton Mifflin, 2007).
  • 5.Gann B. Giving patients choice and control: health informatics on the patient journey. Yearb Med Inf. 2012;7:70–73. [PubMed] [Google Scholar]
  • 6.Paparrizos, J., White, R. W. & Horvitz, E. Screening for pancreatic adenocarcinoma using signals from web search logs: Feasibility study and results. J Oncol Pract (2016). [DOI] [PubMed]
  • 7.Tang H, Ng JHK. Googling for a diagnosis use of google as a diagnostic aid: internet based study. Bmj. 2006;333:1143. doi: 10.1136/bmj.39003.640567.AE. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.White RW, Horvitz E. Cyberchondria:studies of the escalation of medical concerns in web search. Acm Transactions on Inf. Syst. 2009;27:1–37. doi: 10.1145/1629096.1629101. [DOI] [Google Scholar]
  • 9.Prokosch HU, Ganslandt T. Perspectives for medical informatics. reusing the electronic medical record for clinical research. Methods Inf. Medicine. 2009;48:38–44. [PubMed] [Google Scholar]
  • 10.Kohane IS. Using electronic health records to drive discovery in disease genomics. Nat. Rev. Genet. 2011;12:417–28. doi: 10.1038/nrg2999. [DOI] [PubMed] [Google Scholar]
  • 11.de Dombal FT, Leaper DJ, Staniland JR, Mccann AP, Horrocks JC. Computer-aided diagnosis of acute abdominal pain. Br Med J. 1972;2:9–13. doi: 10.1136/bmj.2.5804.9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Miller RA, Jr MF. Use of the quick medical reference (qmr) program as a tool for medical education. Methods Inf. Medicine. 1989;28:340. doi: 10.1055/s-0038-1636814. [DOI] [PubMed] [Google Scholar]
  • 13.Warner, H. R. Iliad as an expert consultant to teach differential diagnosis. In Symposium on Computer Application, 371–376 (1988).
  • 14.Barnett GO, Cimino JJ, Hupp JA, Hoffer EP. Dxplain. an evolving diagnostic decision-support system. Jama. 1987;258:67. doi: 10.1001/jama.1987.03400010071030. [DOI] [PubMed] [Google Scholar]
  • 15.Bisson LJ, et al. Accuracy of a computer-based diagnostic program for ambulatory patients with knee pain. Am. J. Sports Medicine. 2014;42:2371–6. doi: 10.1177/0363546514541654. [DOI] [PubMed] [Google Scholar]
  • 16.Ramnarayan, P., Kulkarni, G., Tomlinson, A. & Britto, J. Isabel: a novel internet-delivered clinical decision support system. (2004).
  • 17.Melle WV. Mycin: a knowledge-based consultation program for infectious disease diagnosis. Int. J. Man-Machine Stud. 1978;10:313–322. doi: 10.1016/S0020-7373(78)80049-2. [DOI] [Google Scholar]
  • 18.Pestian, J. P. et al. A shared task involving multi-label classification of clinical free text. In Proceedings of the Workshop on BioNLP 2007: Biological, Translational, and Clinical Language Processing, 97–104 (Association for Computational Linguistics, 2007).
  • 19.Salvaneschi, P., Masera, A., Lazzari, M. & Lancini, S. Diagnosing ancient monuments with expert software. Struct. Eng. Int. 7, – (1997).
  • 20.Salvaneschi P, Cadei M, Lazzari M. Applying ai to structural safety monitoring and evaluation. IEEE Expert. 2002;11:24–34. doi: 10.1109/64.511774. [DOI] [Google Scholar]
  • 21.Middleton B, et al. Probabilistic diagnosis using a reformulation of the internist-1/qmr knowledge base. ii. evaluation of diagnostic performance. Methods Inf. Medicine. 1991;30:256–267. doi: 10.1055/s-0038-1634847. [DOI] [PubMed] [Google Scholar]
  • 22.Chen, P., Barrera, A. & Rhodes, C. Semantic analysis of free text and its application on automatically assigning icd-9-cm codes to patient records. In Cognitive Informatics (ICCI), 2010 9th IEEE International Conference on, 68–74 (IEEE, 2010).
  • 23.Kalchbrenner, N., Grefenstette, E. & Blunsom, P. A convolutional neural network for modelling sentences. arXiv preprint arXiv:1404.2188 (2014).
  • 24.Lally A, et al. Watsonpaths: Scenario-based question answering and inference over unstructured information. Ibm Corp. 2014;38:59–76. [Google Scholar]
  • 25.Vincent P, Larochelle H, Lajoie I, Bengio Y, Manzagol PA. Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion. J. Mach. Learn. Res. 2010;11:3371–3408. [Google Scholar]
  • 26.Kim, Y. Convolutional neural networks for sentence classification. arXiv preprint arXiv:1408.5882 (2014).
  • 27.Krizhevsky, A., Sutskever, I. & Hinton, G. E. Imagenet classification with deep convolutional neural networks. In International Conference on Neural Information Processing Systems, 1097–1105 (2012).
  • 28.Simonyan, K. & Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014).
  • 29.Szegedy, C. et al. Going deeper with convolutions. In Proceedings of the IEEE conference on computer vision and pattern recognition, 1–9 (2015).
  • 30.Zeiler, M. D. & Fergus, R. Visualizing and understanding convolutional networks. In European conference on computer vision, 818–833 (Springer, 2014).
  • 31.Bengio Y, Vincent P, Janvin C. A neural probabilistic language model. J. Mach. Learn. Res. 2003;3:1137–1155. [Google Scholar]
  • 32.Mikolov, T., Karafit, M., Burget, L., Cernock, J. & Khudanpur, S. Recurrent neural network based language model. In Interspeech 2010, Conference of the International Speech Communication Association, Makuhari, Chiba, Japan, September, 1045–1048 (2010).
  • 33.Mikolov, T., Yih, W. T. & Zweig, G. Linguistic regularities in continuous space word representations. In HLT-NAACL (2013).
  • 34.Collobert R, Weston J, Karlen M, Kavukcuoglu K, Kuksa P. Natural language processing (almost) from scratch. J. Mach. Learn. Res. 2011;12:2493–2537. [Google Scholar]
  • 35.Le, Q. V. & Mikolov, T. Distributed representations of sentences and documents. 4, II–1188 (2014).
  • 36.Ribeiro-Neto B, Laender AHF, Lima LRSD. An experimental study in automatically categorizing medical documents. J. Assoc. for Inf. Sci. Technol. 2001;52:391401. [Google Scholar]
  • 37.Medori, J. Machine learning and features selection for semi-automatic icd-9-cm encoding. In NAACL Hlt 2010 Second Louhi Workshop on Text and Data Mining of Health Documents, 84–89 (2010).
  • 38.Lita, L. V., Yu, S., Niculescu, S. & Bi, J. Large scale diagnostic code classification for medical pati ent records (2008).
  • 39.Maaten, L. V. D. Accelerating t-SNE using tree-based algorithms (JMLR.org, 2014).
  • 40.Maaten L, Hinton G. Visualizing non-metric similarities in multiple maps. Mach. Learn. 2012;87:33–55. doi: 10.1007/s10994-011-5273-4. [DOI] [Google Scholar]
  • 41.Peleg M, Keren S, Denekamp Y. Mapping computerized clinical guidelines to electronic medical records: Knowledge-data ontological mapper (kdom) J. biomedical informatics. 2008;41:180–201. doi: 10.1016/j.jbi.2007.05.003. [DOI] [PubMed] [Google Scholar]
  • 42.Kang, S. M. & Wagacha, P. W. Extracting diagnosis patterns in electronic medical records using association rule mining. Int. J. Comput. Appl. 108 (2014).
  • 43.Boureau, Y. L., Ponce, J. & Lecun, Y. A theoretical analysis of feature pooling in visual recognition. In International Conference on Machine Learning, 111–118 (2010).
  • 44.Srivastava N, Hinton G, Krizhevsky A, Sutskever I, Salakhutdinov R. Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014;15:1929–1958. [Google Scholar]
  • 45.Bouthillier, X., Konda, K., Vincent, P. & Memisevic, R. Dropout as data augmentation. Comput. Sci. (2015).
  • 46.Junyi-Sun. Jieba. https://github.com/fxsjy/jieba.

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The dataset analysed during the current study is available in the Github repository, https://github.com/YangzlTHU/C-EMRs.


Articles from Scientific Reports are provided here courtesy of Nature Publishing Group

RESOURCES