Skip to main content
International Journal of Environmental Research and Public Health logoLink to International Journal of Environmental Research and Public Health
. 2018 Jul 27;15(8):1596. doi: 10.3390/ijerph15081596

Predicting Infectious Disease Using Deep Learning and Big Data

Sangwon Chae 1, Sungjun Kwon 1, Donghyun Lee 1,*
PMCID: PMC6121625  PMID: 30060525

Abstract

Infectious disease occurs when a person is infected by a pathogen from another person or an animal. It is a problem that causes harm at both individual and macro scales. The Korea Center for Disease Control (KCDC) operates a surveillance system to minimize infectious disease contagions. However, in this system, it is difficult to immediately act against infectious disease because of missing and delayed reports. Moreover, infectious disease trends are not known, which means prediction is not easy. This study predicts infectious diseases by optimizing the parameters of deep learning algorithms while considering big data including social media data. The performance of the deep neural network (DNN) and long-short term memory (LSTM) learning models were compared with the autoregressive integrated moving average (ARIMA) when predicting three infectious diseases one week into the future. The results show that the DNN and LSTM models perform better than ARIMA. When predicting chickenpox, the top-10 DNN and LSTM models improved average performance by 24% and 19%, respectively. The DNN model performed stably and the LSTM model was more accurate when infectious disease was spreading. We believe that this study’s models can help eliminate reporting delays in existing surveillance systems and, therefore, minimize costs to society.

Keywords: infectious disease prediction, deep neural network, long short-term memory, deep learning, social media big data

1. Introduction

Infectious disease occurs when a person is infected by a pathogen from another person or an animal. It not only harms individuals, but also causes harm on a macro scale and, therefore, is regarded as a social problem [1]. At the Korea Center for Disease Control (KCDC), infectious disease surveillance is a comprehensive process in which information on infectious disease outbreaks and vectors are continuously and systematically collected, analyzed, and interpreted. Moreover, the results are distributed quickly to people who need them to prevent and control infectious disease. The KCDC operates a mandatory surveillance system in which mandatory reports are made without delay to the relevant health center when an infectious disease occurs and it operates a sentinel surveillance system in which the medical organization that has been designated as the sentinel reports to the relevant health center within seven days. The targets of mandatory surveillance consist of a total of 59 infectious diseases from Groups 1 to 4 by the KCDC. The targets of sentinel surveillance include influenza from Group 3 along with 21 infectious diseases from Group 5. Overall, a total of 80 infectious diseases in six groups are monitored. In the current Korean infectious disease reporting system, if there is a legally defined infectious disease patient at a medical organization, a report is made to the managing health center through the infectious disease web reporting system. The managing health center reports to the city and province health offices through another system and the city and province health offices report to the KCDC.

In the conventional reporting system, some medical organizations’ infectious disease reports are incomplete and delays can occur in the reporting system. For instance, in the traditional influenza surveillance system, around two weeks elapses between when a report is made and when it is disseminated [2]. The KCDC has been running an automated infectious disease reporting system as a pilot project since 2015. However, by 2017, only 2.3% of all medical organizations were participating in the pilot project. In medical organizations using the conventional infectious disease reporting system, a large number of missing and delayed reports can occur, which hinders a prompt response to infectious disease. As such, it is necessary to create a data-based infectious disease prediction model to handle situations in real time. Furthermore, if this model can understand the extent of infectious disease trends, the costs to society from infectious disease can be minimized.

An increasing number of researchers recognize these facts and are performing data-based infectious disease surveillance studies to supplement existing systems and design new models [3,4,5,6,7,8,9]. Among these, studies are currently being performed on detecting infectious disease using big data such as Internet search queries [10,11,12,13,14,15]. The Internet search data can be gathered and processed at a speed that is close to real time. According to Towers et al., Internet search data can create surveillance data faster than conventional surveillance systems [16]. For example, when Huang et al. predicted hand, foot, and mouth disease using the generalized additive model (GAM), the model that included search query data obtained the best results. As such, it has been reported that new big data surveillance tools have the advantage of being easy to access and can identify infectious disease trends before official organizations [17]. In addition to Internet search data, social media big data is also being considered. Tenkanen et al. report that social media big data is relatively easy to collect and can be used freely, which means accessibility is satisfactory and the data is created continuously in real time with rich content [18]. As such, studies have used Twitter data to predict the occurrences of mental illness [19] and infectious disease [20,21,22,23] in addition to predictions in a variety of other scientific fields [24,25,26,27]. In particular, a study by Shin et al. reported that infectious diseases and Twitter data are highly correlated. There is the possibility of using digital surveillance systems to monitor infectious disease in the future [20]. When these points are considered, using search query data and social media big data should have a positive effect on infectious disease predictions.

In addition to these studies, there are also studies that have used techniques from the field of deep learning to predict infectious disease [22,23,28,29]. Deep learning is an analysis method and, like big data, it is being actively used in a variety of fields [30]. Deep learning yields satisfactory results when it is used to perform tasks that are difficult for conventional analysis methods [31,32,33]. In a study by Xu et al., a model that used deep learning yielded better prediction performance than the generalized linear model (GLM), the least absolute shrinkage and selection operator (LASSO) model, and the autoregressive integrated moving average (ARIMA) model [28]. As such, methods of predicting infectious disease that use deep learning are helpful for designing effective models.

There are also examples of infectious disease prediction based on environmental factors such as weather [34,35,36,37]. Previous studies have confirmed that weather data comprises a factor that has a great influence on the occurrence of infectious diseases [38,39,40]. Liang et al. showed that rainfall and humidity are risk factors for a hemorrhagic fever with a renal syndrome [41]. In addition, a study by Huang et al. reported that trends in dengue fever show a strong correlation with temperature and humidity [42]. Previous studies indicate that infectious disease can be predicted more effectively if weather variables, Internet big data, and deep learning are used.

Most previous research has attempted to predict infectious disease using Internet search query data alone. However, as discussed above, it is necessary to also consider various big data and environmental factors such as weather when predicting infectious disease. In addition, in the case of models that use deep learning, it is possible to improve prediction performance by optimizing the deep learning model by optimizing its parameters. Therefore, the aim of this study is to design a model that uses the infectious disease occurrence data provided by the KCDC, search query data from search engines that are specialized for South Korea, Twitter social media big data, and weather data such as temperature and humidity. According to a study by Kwon et al., a model that considers the time difference between clinical and non-clinical data can detect infectious disease outbreaks one to two weeks before current surveillance systems [43]. Therefore, this study adds lag to the collected dataset to take temporal characteristics into account. In addition, in the design process, a thorough testing of all the input variable combinations is performed to examine the effects of each resulting dataset on infectious disease outbreaks and select the optimal model with the most explanatory power. The model’s prediction performance is verified by comparing it with an infectious disease prediction model that uses a deep learning method and an infectious disease prediction model that uses time series analysis.

Ultimately, using the results obtained by this study, it should be possible to create a model that can predict trends about the occurrence of infectious disease in real time. Such a model can not only eliminate the reporting time differences in conventional surveillance systems but also minimize the societal costs and economic losses caused by infectious disease.

The remainder of this paper is organized as follows. Section 2 describes the data sources and standards used in this study and introduces the analysis methodology used to design the prediction model. In Section 3, the analysis results are described and their implications are discussed. Section 4 discusses the results. Section 5 concludes the paper.

2. Data and Methods

2.1. Research Data

As mentioned above, this study uses four kinds of data to predict infectious disease, which includes search query data, social media big data, temperature, and humidity. The standards for the non-clinical data are as follows. Data from 576 days between 1 January, 2016 and 29 July, 2017 was used. The infectious diseases selected for this study are subject to mandatory reporting. Unlike those diseases subject to mandatory reporting, diseases subject to sentinel reporting aggregate data on a weekly basis. Since prediction is also performed on a weekly basis, it is difficult to cope with infectious diseases in real time. Therefore, diseases that are subject to sentinel reporting were excluded from the study. Moreover, the study excluded infectious diseases with an annual occurrence rate of less than 100 as well as infectious diseases that have a statistically insignificant model with an adjusted R-squared value of less than 0.25 when regression analysis is performed using all variables. Three infectious diseases satisfied all conditions, which include malaria, chickenpox, and scarlet fever. The search data was collected from the Naver Data Lab (https://datalab.naver.com/keyword/trendSearch.naver). The usage share data provided by InternetTrend (http://internettrend.co.kr/trendForward.tsp) on search engines in the health/medicine field in the first half of 2017 shows that the Naver search engine had the highest usage share (86.1%) in South Korea. Therefore, it was chosen as the search engine for extracting search data. Note that the collected search data consists of only Korean terms because the search engine is specific to South Korea. The search queries used in this study consisted of the infectious disease’s proper name and symptoms (e.g., “chickenpox” and “chickenpox symptoms” in Korea). The frequency of inquiries using these search queries were used as the search data. The number of searches were normalized with respect to the largest number of searches within the study period.

Weather data (temperature and humidity) were collected from the Korea Meteorological Administration’s weather information open portal (https://data.kma.go.kr). Hourly data collected from weather stations nationwide was converted into daily average data for each station. In Gyeonggi-do province, where around half of South Korea’s population lives, there are many weather stations crowded together. There was a concern that simply finding the averages of the daily data for each station would cause errors to occur, so the following process was performed. First, the averages of the data from each station were collected for the eight provinces in South Korea (Gyeonggi-do, Gangwon-do, Chungcheongnam-do, Chungcheongbuk-do, Jeollanam-do, Jeollabuk-do, Gyeongsangbuk-do, and Gyeongsangnam-do). Next, the averages of the data for each of the eight provinces were found to obtain South Korea’s national average weather data. Average temperature (degrees Celsius) and average humidity (percentage) were recorded.

Social media big data was collected for each infectious disease from Twitter through a web crawler that used the Python Selenium library. For the Twitter data, the daily number of tweets mentioning infectious disease was recorded.

Lastly, infectious disease data was collected from the infectious disease web statistics system (https://is.cdc.go.kr/dstat/index.jsp). This data consists of the daily number of people who were infected throughout South Korea. Table 1 shows the sources and descriptions of the data.

Table 1.

Description of data types. KCDC: Korea Center for Disease Control; KMA: Korea Meteorological Administration’s weather information open portal.

Variable Source Description Number of Observations
Occurrences KCDC Daily number of confirmed infectious disease diagnoses 576
Naver Naver Data Lab Daily Naver search frequency
Twitter Twitter Daily number of Twitter mentions
Temperature KMA Average daily temperature for all of South Korea
Humidity Average daily humidity for all of South Korea

Table 2 shows the statistics for each of the infectious disease variables used in this study. In the case of temperature and humidity, the same conditions were used, which means they were put in a shared category. The data in Table 2 shows that an average of 166.76 people are infected with chickenpox daily with a standard deviation of 98.37 and the daily Naver frequency average is 33.94 with a standard deviation of 15.50. We observed that all the statistics for chickenpox are higher than those for other infectious diseases.

Table 2.

Data statistics.

Disease Variable Min. Median Mean Max. Var. SD
Chicken Pox Occurrences 16 146.5 166.76 562 9676.29 98.37
Naver 10.56 29.01 33.94 100 239.97 15.50
Twitter 1 10 12.81 194 220.39 14.85
Scarlet fever Occurrences 3 37 46.33 252 1346.85 36.70
Naver 0.32 2.72 4.99 100 72.63 8.52
Twitter 0 0 0.27 15 1.06 1.03
Malaria Occurrences 0 0 1.65 14 6.65 2.58
Naver 5.52 17.24 23.13 100 193.02 13.89
Twitter 0 3 4.23 34 15.08 3.88
Environ-mental variables Temperature (°C) −10.82 14.19 13.27 29.54 94.07 9.70
Humidity (%) 30.87 67.31 66.70 94.31 172.81 13.15

2.2. Analysis Method

Figure 1 shows the overall framework of the model used in this study including the data collection process and the comparison of models designed using the deep neural network (DNN) method, the long-short term memory (LSTM) method, the autoregressive integrated moving average (ARIMA) method, and the ordinary least squares (OLS) method.

Figure 1.

Figure 1

Infectious disease prediction model.

This study constructed an infectious disease surveillance model that uses non-clinical search data, twitter data, and weather data. To design the optimal prediction model, the OLS models that use all possible combinations of variables in the dataset were created. The adjusted R-squared values of each model were compared. In addition, lags of 1–14 days were added to each infectious disease and their adjusted R-squared values were compared in a preliminary analysis. A lag of seven days, which had high explanatory power for all infectious diseases, was selected as the optimal lag parameter. The optimal parameters were used to create the OLS, ARIMA, DNN, and LSTM models.

Before analysis, this study applied a lag of seven days between the input variables (optimal variable combination) and their associated output variable (disease occurrence). The OLS dataset was divided into a training data subset and a test data subset using a ratio of 8:2. This means all 569 rows of collected data were divided such that there were 455 rows for the training data subset and 114 rows for the test data subset. The training data subset was only used for model training. The test data subset was only used for prediction and performance evaluation in the model after training. The ARIMA dataset was also divided into a training data subset and test data subset using a ratio of 8:2, but only the disease occurrences were required for ARIMA. Similarly to the data above, the 569 rows of disease occurrence data were divided into 455 rows for the training data subset and 114 rows for the test data subset.

In the DNN and LSTM models, the whole dataset was divided into training, validation, and test data subsets at a ratio of 6:2:2 and training was performed. This means all 569 rows of collected data were divided into 341 rows for the training data subset, 114 rows for the validation data subset, and 114 rows for the test data subset. The training data subset was used for model training. The validation data subset was only used for performance evaluation during training. The final model after training was the model that yielded the best performance when the validation data subset was used in training. The test data subset was only used for the prediction and performance evaluation.

To compare the models, the root mean squared error (RMSE) was used to evaluate the prediction rates. RMSE is a common measurement for the difference between predicted and actual values. It is usually used in the other fields as well as in the prediction of infectious diseases [28,44,45]. RMSE is calculated using the equation below.

RMSE=t=1n(XtX^t)2n  (1)

2.2.1. Selecting Optimal Variable Combinations

The optimal variable combinations for the model were selected by considering all possible models in the regression analysis. The models are combinations of the four types of data in the dataset (Naver searches (n), Twitter searches (tw), temperature (t), and humidity (h)).

Figure 2 shows the adjusted R-squared values of 15 regression models for each infectious disease. Among the observed regression models, the models that are combinations of all variables had the best explanatory power. Therefore, this combination was chosen as the optimal variable combination.

Figure 2.

Figure 2

Variable optimization parameters. (a) Adjusted R-squared values of 15 regression models in chickenpox, (b) Adjusted R-squared values of 15 regression models in scarlet fever, (c) Adjusted R-squared values of 15 regression models in malaria.

2.2.2. Selecting the Optimal Prediction Time Difference

Previous results [43] have shown that it is possible to predict infectious disease at an early stage if a model is designed to consider the time difference between clinical data and non-clinical data. Based on this observation, our model was designed to consider the time difference in each data set. In this situation, “lag” refers to the time delay between the date the data is collected and the date at which the effects actually occur. This means analysis was performed by establishing the time difference between the four input variables used in this study and the output variable that is actually affected. For example, a lag of 1 means that the output variable of 2 January 2016 is calculated using the input variables of 1 January 2016.

Figure 3 shows the adjusted R-squared values of regression models when 1–14 days of lag were tested for each of the infectious diseases in order to select the optimal lag. In the case of chickenpox, it was found that lags of 1, 7, and 14 days yielded the highest explanatory power. For scarlet fever, it was found that lags of 4, 7, and 11 days yielded the highest explanatory power. In the case of malaria, it was found that lags of 1, 2, and 7 days yielded the highest explanatory power. For chickenpox and malaria, the lag with the highest explanatory power was one day. However, it was decided that this lag was not suitable for the ultimate goal of reducing the length of delay from reporting to dissemination. In the observed regression models, the explanatory power of a lag of seven days was high for all infectious diseases. Therefore, it was decided that this lag was the most suitable and was used for later predictions.

Figure 3.

Figure 3

Adjusted R-squared for each lag. (a) Adjusted R-squared values of regression models that applied lag of 1–14 days in chickenpox, (b) Adjusted R-squared values of regression models that applied lag of 1–14 days in scarlet fever, (c) Adjusted R-squared values of regression models that applied lag of 1–14 days in malaria.

2.2.3. OLS

In this study, the OLS model was used to select the optimal parameter values. It was also used as a comparison model to evaluate the prediction performance of the deep learning models.

Linear regression is a regression analysis technique that models the linear correlation between the output variable y and one or more input variables x in the collected data. The model has the following form.

yi=β1xi1 +  +βpxip+εi= xiT β+ εi, i=1, , n   (2)

OLS is the most simple and commonly used form of linear regression. It is a technique that minimizes the sum of squared errors and can solve the mathematical expression for ß, which is the parameter to be predicted, by using the equation below.

β^=(XTX)1XTy= (xixiT)1(xiyi)   (3)

OLS analyses were performed by R version 3.3.3 (https://www.r-project.org/).

2.2.4. ARIMA

Because OLS is the simplest form of linear regression analysis, it is not sufficient for comparison with deep learning models. Therefore, we also compare the ARIMA model, which is often used for the prediction of infectious diseases [44,45,46]. This will more clearly compare traditional analysis methods (OLS and ARIMA) with deep learning (DNN and LSTM). The ARIMA model is a method for analyzing non-stationary time series data. One characteristic of ARIMA analysis is that it can be applied to any time series. In particular, it shows the detailed changes when the data fluctuates rapidly over time.

In this study, we used seasonal ARIMA because the collected data is seasonal. The seasonal ARIMA model is denoted as ARIMA(p, d, q)(P, D, Q)S. where p is the order of the autoregressive part, d is the order of the differencing, q is the order of the moving-average process, and S is the length of the seasonal cycle. (P, D, Q) is the seasonal part of the model. The seasonal ARIMA model is written below.

(1B)d(1BS)DYt= μ+ θ(B)θS(BS)ϕ(B)ϕS(BS)at  (4)

where Yt refers to the value of the time series at time t, μ is the mean term, at is the independent disturbance, B is the backshift operator, ϕ(B) is the autoregressive operator, and θ(B) is the moving average operator. ϕS(BS)  and θS(BS) are the seasonal operators of the model.

The ARIMA analyses were carried out using the R version 3.3.3.

2.2.5. DNN

The DNN model is a feedforward analysis method that is a basic model for deep learning. DNN is composed of a minimum of three node layers and, with the exception of the input node, each node uses a nonlinear activation function. DNN uses a supervised learning technique called backpropagation. In this study, an infectious disease prediction model that uses DNN was designed and the basic DNN model was compared with this more advanced deep learning model.

The variables used in DNN are bias b, input x, output y, weight w, calculation function σ and activation function f(σ). Each neuron in DNN uses the following equation.

σ :Sum=w·x+b  (5)
y :f(σ)=f(w·x+b)  (6)

Figure 4 shows the structure of a neuron in the DNN model. The DNN analyses were carried out using the “Dense layer” option of the Keras package in the Python version 3.5.3 (https://keras.io/). There are 10 parameters available in the Dense layer. We only modified the units, activation function, and dropout. The rest of the parameters used the default values (e.g., use_bias = True and kernel_regularizer = None).

Figure 4.

Figure 4

DNN structure. The variables used are bias b, input x, output y, weight w, calculation function σ and activation function f(σ).

2.2.6. LSTM

The LSTM model is suitable for predicting time series data when there is a time step with a random size [47]. It was thought that prediction performance could be improved by creating an infectious disease prediction model using LSTM and the time series data collected in this study.

An important advantage of recurrent neural networks (RNNs) is that contextual information is available when mapping IO sequences. However, there is a gradient problem in that the effect of a given input on the hidden layer can be increased or decreased significantly during the circular connection. As new inputs are overwritten, the sensitivity of the first input decreases over time. Therefore, the network is “forgotten”. The input gate, output gate, and forget gate are non-linear summation units that control the activation of the cell. The forget gate multiplies the previous state of the cell while the input and output gates multiply the IO of the cell. The activation function f of the gate is a logistic sigmoid. The IO activation functions g and h of the cell usually use hyperbolic tangents or logistic sigmoids. However, in some cases, h uses the identity function. As long as the forget gate is open and the input gate is closed, the memory cell continues to remember the first input. In this way, LSTM is an algorithm that resolves a problem in traditional RNNs [48].

The equations for forgetting, storing, renewing, and outputting information in the cell are shown below, respectively.

ft=σ(wf[ht1, xt]+bf)  (7)
it=σ(Wi[ht1, xt]+bi)  (8)
C˜t=tanh(WC[ht1, xt]+bC)  (9)
Ct=ft× Ct1 + it×C˜t   (10)
ot=σ(Wo[ht1, xt]+bo)  (11)
ht=ot×tanh(Ct )  (12)

When data (xt) is input to the LSTM cell in Equation (7), function ft determines the information to be forgotten in the cell layer. In Equations (8) and (9), information that will be newly saved in the cell layer is created in it and C˜t In Equation (10), the cell layer Ct  is renewed using ft, it, and C˜t In Equation (11), the cell layer’s information is used and ht is the output. In Equation (12), the cell state gets a value between −1 and 1 through the tanh function. The values of Ct  and ht are kept for the next iteration of LSTM. LSTM analyses were carried out using the “LSTM layer” of the Keras package in the Python version 3.5.3. There are 23 parameters available in the LSTM layer. We only set the units, activation function, return sequence, and dropout. The rest of the parameters used the default values (e.g., use_bias = True, recurrent regularizer = None, recurrent_constraint = None, and unit_forget_bias = None).

2.2.7. Determining the Optimal Deep Learning Parameters

Figure 5 shows the parameter selection method for the deep learning approach used in this study. The Adadelta, Adagrad, Adam, Adamax, Nadam, RMSprop, and stochastic gradient descent (SGD) optimizers were compared. All parameters of each optimizer used the default values of the Keras package. For instance, in SGD, the learning late is 0.01, the momentum is 0, the decay is 0, and the Nesterov momentum is false. In addition, the following activation functions were evaluated: exponential linear unit (ELU), rectified linear unit (ReLU), scaled ELU (SELU), and Softplus. Lastly, various numbers of epochs (400, 600, 800, and 1000) were evaluated. The other parameters were fixed as follows: number of hidden layers = 4, number of units in each hidden layer = 32, batch size = 32, and drop out = 0. Prediction models with variable and fixed parameters were trained on the data and the resulting models were compared to determine the optimal prediction model. To ensure the amount of DNN model data was the same as that of the LSTM model, previous data from the same time period as the LSTM was inserted. All deep learning models were implemented using the Keras package in the Python version 3.5.3.

Figure 5.

Figure 5

Deep learning parameter selection.

3. Results

3.1. OLS

The regression model was formed based on 569 days of data in which a lag of seven days was applied to each infectious disease dataset. The dataset was divided up in an 8:2 ratio and each part was used for constructing the regression model and prediction. Table 3 presents the OLS results.

Table 3.

OLS Results.

Disease R2 Adjusted R2 F p Variable Coefficient T p
Chickenpox 0.4077 0.4035 97.0659 <0.001 Naver 4.4569 18.2096 <0.001
Twitter 0.2162 0.9769 0.3290
Temperature −3.8421 −8.7717 <0.001
Humidity −0.8919 −3.1027 0.0020
Intercept 121.9521 6.6302 <0.001
Scarlet fever 0.2867 0.2817 56.6851 <0.001 Naver 2.1956 12.6929 <0.001
Twitter −1.981 −1.3940 0.1639
Temperature 0.2559 1.5660 0.1179
Humidity −0.5369 −4.4766 0.0491
Intercept 68.5623 9.4170 <0.001
Malaria 0.3863 0.3819 88.7462 <0.001 Naver 0.0649 7.1140 <0.001
Twitter 0.0369 1.6594 0.0976
Temperature 0.0770 5.7623 <0.001
Humidity 0.0129 1.6421 0.1011
Intercept −1.8257 −3.7883 <0.001

Each regression model had results that were below the level of significance (p < 0.05). The adjusted R-squared value was greater than 0.25 for all three infectious diseases, which means the models can be said to have significant explanatory power. Of the infectious disease regression models, the chickenpox model yielded significant results for the Naver search queries, temperature, and humidity. The scarlet fever model yielded significant results for the Naver search queries and humidity. Additionally, the malaria model yielded significant results for the Naver search queries and temperature (p < 0.05). Looking at these results together, the Naver search query data was significant for all three infectious diseases and the Twitter data was not significant for any of the three. It can be seen that the Internet search query data can be used to design an infectious disease prediction model, which was reported by previous studies. However, the results for the Twitter data differ from the results of previous studies. This is believed to be because Naver accounted for the largest share (86.2%) of Korean search engine use in the health/medicine field for the first half of 2017 while Twitter accounted for the smallest share (0.5%) of social media use in the health/medicine field for the same time period (http://internettrend.co.kr/trendForward.tsp). However, the Twitter data had an effect on the process of finding the model with the highest adjusted R-squared value. Therefore, it is expected to have an effect on future analysis as well. The temperature had a significant relationship with all infectious diseases except for scarlet fever and humidity had a significant relationship with all infectious diseases except for malaria. The values of the coefficients show that the most significant variables for chickenpox and scarlet fever was the Naver search query data (4.4589 and 2.1956, respectively) and, for malaria, it was the temperature values (0.0770). The effect of Naver search query data in particular was significant for all three infectious diseases, which confirms that it can be suitable for predicting infectious disease.

3.2. ARIMA

The seasonal ARIMA model was evaluated using the same data used for OLS. The autocorrelation function and the partial autocorrelation function were checked for the seasonality of infectious diseases and seasonality was observed. It was considered inappropriate to select the parameters (e.g., p, d, q) because cuts off and tails off are unclear. Therefore, the optimal model for each infectious disease was selected based on the Akaike information criterion (AIC) and RMSE. The AIC and RMSE were used to compare the ARIMA models.

Table 4 shows the AIC and RMSE of the seasonal ARIMA model for each infectious disease, which allows us to identify the top three ARIMA model. In addition, the choice of parameter values did not substantially affect the AIC and RMSE of the model for a single infectious disease.

Table 4.

Results of seasonal ARIMA. AIC: Akaike information criterion; RMSE: root mean squared error.

Disease Models AIC RMSE
Malaria ARIMA(1, 1, 1)(1, 0, 1)7 1740.807 3.179
ARIMA(2, 1, 2)(1, 0, 1)7 1741.823 3.174
ARIMA(1, 1, 1)(2, 0, 1)7 1742.746 3.179
Scarlet fever ARIMA(1, 0, 1)(0, 1, 2)7 3513.540 44.884
ARIMA(1, 0, 2)(0, 1, 2)7 3513.891 45.340
ARIMA(2, 0, 2)(1, 1, 1)7 3515.127 46.026
Chickenpox ARIMA(1, 0, 4)(0, 1, 2)7 4576.544 96.398
ARIMA(2, 0, 4)(1, 1, 2)7 4577.808 97.256
ARIMA(2, 0, 2)(1, 1, 1)7 4583.896 96.835

3.3. Top-Ranked DNN and LSTM Prediction Models

To compare the performance of each model, Figure 6 shows the 10 models with the lowest RMSE of the test data subset. The numbers inside the parentheses in each model’s name represent the optimizer, the activation function, and the number of epochs used in the models (e.g., DNN (1, 2, 3) indicates that the optimizer, the activation function, and the number of epoch are Adadelta, ReLU, and 800, respectively). The metric used to compare the models is RMSE, which shows the difference between the actual and predicted values. A smaller RMSE value indicates a smaller difference between the actual and predicted values and indicates a higher prediction performance.

Figure 6.

Figure 6

Ten models with the lowest RMSEs for each infectious disease. The numbers in parentheses indicate each model’s optimizer, activation, and number of epochs, respectively. (optimizer) 1: Adadelta, 2: Adagrad, 3: Adam, 4: Adamax, 5: Nadam, 6: RMSProp, and 7: SGD, (activation function) 1: ELU, 2: ReLU, 3: SELU, and 4: SoftPlus, (number of epochs) 1: 400, 2: 600, 3: 800, and 4: 1000. (a) The top-10 deep learning models and traditional analysis models with the lowest RMSEs in chickenpox, (b) The top-10 deep learning models and traditional analysis models with the lowest RMSEs in scarlet fever, (c) The top-10 deep learning models and traditional analysis models with the lowest RMSEs in malaria.

Table S1 shows the RMSE and prediction graphs of the DNN and LSTM models with the lowest RMSE for chickenpox. It can be seen that the prediction graphs for each analysis method have similar shapes overall. The 10 DNN models for chickenpox had a mean RMSE of 72.8215 and a standard deviation of 1.28, which shows stable model performance. When the prediction performances of each model were compared based on RMSE, the top 10 DNN models showed a 24.45% performance improvement compared to the ARIMA model. The mean RMSE of the 10 LSTM models was 78.2850, which is higher than the DNN models. The standard deviation was 3.64, which shows that the difference among LSTM models was more marked than the difference among DNN models. Despite this, the top 10 LSTM models achieved an 18.78% performance improvement over the ARIMA model on average. There was a difference between the DNN and LSTM models’ average figures and standard deviations. However, in the models with the lowest RMSE for each analysis method, there was not a big difference, which indicates that there was not a large difference in performance when the optimal parameters for each analysis method were used. Table S2 shows the RMSE and the prediction graphs of the DNN and LSTM models with the lowest RMSE for scarlet fever. The shapes of the graphs for the DNN models are similar. For the LSTM models, the shapes of the graphs are similar except for the model with the lowest RMSE. Unlike the graphs of the other LSTM models, the graph LSTM model with the lowest RMSE showed a strong tendency to follow the actual trend. This result infers that a prediction model that is better than existing prediction models can be designed by changing the deep learning parameters to achieve optimization. As is the case with chickenpox, the mean RMSE of the DNN model (34.4347) for scarlet fever was lower than that of the LSTM model (36.8140). The standard deviation of the DNN model (0.80) was also lower than that of the LSTM model (1.37). When comparing each model based on RMSE, the top 10 DNN models showed a 23.28% performance improvement over the ARIMA model and the LSTM models showed a 17.97% performance improvement over the ARIMA model. Table S3 shows the RMSE and the prediction graphs of the DNN and LSTM models with the lowest RMSE for malaria. Like other infectious diseases, the DNN model prediction graphs have a similar shape. However, the shapes of the LSTM model prediction graphs have a tendency to not follow the trend. The RMSEs of each prediction model, excluding the ARIMA model, showed little difference. This is believed to be because the number of malaria occurrences is fewer than those of the other infectious diseases. Therefore, adequate predictions could not be formed.

3.4. Optimal Performance Model

It is difficult to understand the special characteristics of each analysis method by simply comparing RMSE figures alone. Therefore, a detailed comparison was performed on the basic comparison models (OLS and ARIMA) and the analysis methods that use deep learning (DNN and LSTM). The deep learning models used for comparison were the models with the optimal performance and lowest RMSE so that they could best represent each analysis method. With the best performance.

Figure 7 shows the chickenpox predictions of the model with the lowest RMSE out of the 10 models with the lowest RMSE for each analysis method. The DNN model with the best performance had the following specifications, which include optimizer = Adadelta, activation function = ReLU, and number of epochs = 400 (DNN (1, 2, 1)). The LSTM model with the best performance had the following specifications, which include optimizer = Nadam, activation function = SoftPlus, epochs = 800 (LSTM (5, 4, 3)). The OLS model’s predictions had a smaller range of fluctuation than the deep learning models. From day 480, it seems to follow the trend, but it does not follow the small changes. After day 550, it cannot predict the downward shape even within a stable graph model. In short, the OLS model is not suitable as a prediction model. The ARIMA model’s prediction graph has a very simple shape. This model is cyclic and there is a slight increasing trend in which the predicted value per cycle increases by a factor of about 2.5 each time. This model cannot predict the trend at all. It can only predict a stable cyclic behavior. In contrast, the DNN (1, 2, 1) predictions followed the actual occurrence trend well. Moreover, it had a large range of fluctuation, which means it made accurate predictions overall. However, when the number of occurrences rose rapidly in days 510–520, it was unable to follow these values. The LSTM (5, 4, 3) predictions had a smaller range of fluctuation than the DNN (1, 2, 1) model. Its range of variance was small, which means it had a stable shape and it performed better than DNN (1, 2, 1) when the number of occurrences rose rapidly.

Figure 7.

Figure 7

Top performing chickenpox predictions for each analysis method. (a) Prediction graph of OLS model in chickenpox, (b) Prediction graph of ARIMA model in chickenpox, (c) Prediction graph of DNN (1, 2, 1) model with the best performance in chickenpox, (d) Prediction graph of LSTM (5, 4, 3) model with the best performance in chickenpox.

Figure 8 shows the scarlet fever predictions of the models with the best performance for each analysis method. The DNN model with the best performance had the following specifications: optimizer = Adadelta, activation function = ELU, and number of epochs = 600 (DNN (1, 1, 2)). The LSTM model with the best performance had the following specifications: optimizer = Adamax, activation function = ELU, number of epochs = 400 (LSTM (4, 1, 1)). The OLS model’s predictions were completely unable to follow the trend, which was similar with the chickenpox case. The ARIMA model’s prediction has no particular merit because it also predicts a simple cycle. Much like its chickenpox prediction, it can only predict a stable cyclic behavior. The DNN (1, 1, 2) predictions were relatively close when the number of occurrences was low, but they were too low when the number of occurrences was high. LSTM (4, 1, 1) had a larger range of variance than DNN (1, 1, 2) and its predictions were close when the number of occurrences was high. In the prediction graphs for all of the top performing scarlet fever models, none of the models were able to follow the trend on days 480–500 when there was a severe variance in the number of actual occurrences. Looking at the mean of each model’s predicted value for the number of occurrences, the LSTM model (88.7568) was larger than the mean of the DNN model (77.096). These same results were also seen in the case of chickenpox (DNN model mean = 237.5318, LSTM model mean = 241.5186). This showed that more suitable results can be obtained if the LSTM model is used to predict the maximum value for the number of occurrences and the DNN model is used to predict the minimum value.

Figure 8.

Figure 8

Top performing scarlet fever predictions for each analysis method. (a) Prediction graph of OLS model in scarlet fever, (b) Prediction graph of ARIMA model in scarlet fever, (c) Prediction graph of DNN (1, 1, 2) model with the best performance in scarlet fever, (d) Prediction graph of LSTM (4, 1, 1) model with the best performance in scarlet fever.

Figure 9 shows the malaria predictions of the models with the best performance for each analysis method. The DNN model with the best performance had the following specifications: optimizer = Adamax, activation function = SoftPlus, number of epochs = 800 (DNN (4, 4, 3)). The LSTM model with the best performance had the following specifications: optimizer = Adadelta, activation function = SoftPlus, number of epochs = 800 (LSTM (1, 4, 3)). The predictions of the analysis methods were not satisfactory, but the DNN (4, 4, 3) model’s predictions seemed to follow the trend relatively well. The ARIMA model predicts values close to 0. It is believed that the occurrences in the malaria data are less than those of other diseases and, therefore, not suited to time series analysis because occurrences are concentrated in the summer seasons. As seen in Section 3.3, the lowest RMSEs of each prediction model excluding the ARIMA model showed little difference. LSTM (1, 4, 3) had a lower range of variance than OLS and it seemed completely unable to make predictions. As mentioned before, the reason that predictions were inadequate for all of the models in addition to LSTM (1, 4, 3) was that the number of malaria occurrences was small and proper results could not be produced.

Figure 9.

Figure 9

Top performing malaria predictions for each analysis method. (a) Prediction graph of OLS model in malaria, (b) Prediction graph of ARIMA model in malaria, (c) Prediction graph of DNN (4, 4, 3) model with the best performance in malaria, (d) Prediction graph of LSTM (1, 4, 3) model with the best performance in malaria.

When the predictions were compared according to each analysis method, the DNN and LSTM deep learning models performed better than the OLS and ARIMA models by assuming that there was a sufficiently large number of occurrences. When comparing the DNN and LSTM models, the best models had similar performance, but the DNN models were better in terms of average performance. However, when the number of occurrences was large, the LSTM model made close predictions. It seems to be an analysis method that is suitable for circumstances when the number of occurrences is rapidly increasing and infectious disease is believed to be spreading.

4. Discussion

The deep learning model showed outstanding performance compared to the traditional ARIMA method. Of all the DNN and LSTM prediction models for chickenpox, the optimal models with the lowest RMSE yielded 27.22% and 27.33% better performance than the ARIMA model, respectively. The top 10 DNN models for chickenpox improved performance by an average of 24.45% and the LSTM models improved performance by an average of 18.78%. The lowest RMSEs of the DNN and LSTM prediction models for scarlet fever showed 26.25% and 23.79% improved performances compared to ARIMA models. The top 10 DNN models for scarlet fever improved performance by an average of 23.28%. The LSTM models improved performance by an average of 17.97%.

As noted in the previous sections, it was difficult to predict infectious diseases when the number of infections was small and concentrated in one season. In effect, we observed that the incidence of malaria was high over days 160–250 and after day 530. This period corresponds to the summer season in Korea. Predicting infectious diseases with this particular data set was difficult and it was not suitable for the ARIMA analysis. Even using this particular data set, when DNN was used, the trend of infectious diseases was followed comparatively (Figure 9). Moreover, there is a possibility that the performance would be improved in the DNN model if more diverse parameters were adjusted. This means using deep learning has the advantage of scalability and this can be further investigated in future studies.

The ARIMA model that was used in this study was observed to be effective if the number of incidences of infectious diseases was regular and had no increasing or decreasing trends. However, actual data can have trends and be irregular. Therefore, deep learning can be an excellent analytical method when analyzing such data and predicting future situations. According to the results of the previous analyses, the deep learning model follows increasing and decreasing trends sufficiently well. Moreover, the DNN and LSTM models were observed to be sensitive to decreasing trends and increasing trends, respectively.

5. Conclusions

Infectious disease is a social problem in that it can cause not only personal damage but also widespread harm. For this reason, research is being conducted to minimize social losses by predicting the spread of infectious diseases. The aim of this study was to design an infectious disease prediction model that is more suitable than existing models by using various input variables and deep learning techniques. Therefore, in this study, the optimal parameters were set using a variable selection method based on OLS. The relationship between actual instances of disease occurrence and the Internet search query data tends to have a time lag, which means a lag was added to each infectious disease’s dataset to find the future trend. Next, an analysis of ARIMA, DNN, and LSTM was performed with optimal parameters.

The results of OLS analysis using optimal parameters showed that the regression models for each infectious disease had significant results. Of the four input variables, the Naver search frequency had a significant relationship with all three infectious diseases. The performance of the OLS and ARIMA analysis was used to evaluate the deep learning models. Looking at the results for DNN and LSTM, both the deep learning models made much better predictions than the OLS and ARIMA models for all infectious diseases. Moreover, the DNN models had the best performance on average, but the LSTM models made more accurate predictions when infectious diseases were spreading. However, in the case of malaria, there were few occurrences of the disease compared to other infectious diseases, which means the predictions were not comparatively accurate.

This study was also able to reveal special characteristics of the DNN and LSTM models. The DNN model produced smaller values than the LSTM model on average when predicting infectious diseases. Suitable predictions can be made using the DNN model when predicting the minimum value for disease occurrence and using the LSTM model when predicting the maximum value.

In previous studies, deep learning algorithms were not used [10,11,12,13,14,15,17] or the amount of data considered was small [22,23,28,29]. This study used social media big data and weather data, which have not been sufficiently considered in existing studies. It also used deep learning analysis, which yields high prediction performance to increase the performance of infectious disease predictions. The results showed that, when selecting the optimal parameters, adding all input variables had the highest explanatory power. This means that, by adding various data, it was possible to design a model with higher explanatory power. Moreover, the LSTM model results for scarlet fever indicate that it is possible to optimize a deep learning model by changing its parameters in various ways and, therefore, design a prediction model that is better than existing prediction models.

This study has reviewed the factors involved with infectious disease occurrence using search query data and social media big data, which exist because of the development of the Internet as well as temperature and humidity weather data. It also constructed traditional prediction models such as OLS, ARIMA, and deep learning prediction models such as DNN and LSTM and compared their prediction performance to confirm that the models that use deep learning are the most suitable for infectious disease prediction. It is believed that infectious disease prediction models that employ deep learning can be used to supplement current infectious disease surveillance systems and, at the same time, predict trends in infectious disease. If this can reduce the time differences in reporting systems so that infectious disease trends can be known immediately, it is expected that immediate responses to infectious disease will become possible and costs to society can be minimized. According to a study by Shin et al., an emerging infectious disease known as the Middle East Respiratory Syndrome (MERS) has a deep correlation with Internet search data [20] and it will become possible to expand these methods to the real-time surveillance and prediction of emerging infectious diseases as well.

However, this study has three limitations, which include a relatively short data collection period, regionally combined predictions, and a consideration of a narrow range of parameters in the deep learning model. The search query data collection time period used in this research was relatively short extending from 1 January 2016 to 29 July 2017. The particular spatial ranges of data were averaged across the whole of South Korea. It is believed that, if the data is expanded and the spatial ranges are subdivided, the model’s performance will improve. In addition, an effort was made to change the DNN and LSTM model parameters and create a variety of prediction models, but the deep learning prediction models used in this study did not cover all the prediction models that could be implemented. Parameters such as hidden layers and batch size were not considered. Therefore, it is difficult to conclude that the most effective model was created. If more parameters are considered and more prediction models are made in future research, it is believed that prediction performance can be increased somewhat.

Supplementary Materials

The following are available online at http://www.mdpi.com/1660-4601/15/8/1596/s1, Table S1: The root mean squared error (RMSE) and prediction graphs of top 10 deep neural network (DNN) and long-short term memory (LSTM) models for chickenpox. The seasonal autoregressive integrated moving average (ARIMA) model is denoted as ARIMA(p, d, q)(P, D, Q)S. where p is the order of the autoregressive part, d is the order of the differencing, q is the order of the moving-average process, and S is the length of the seasonal cycle. (P, D, Q) is the seasonal part of the model. The numbers in parentheses indicate each deep learning model’s optimizer, activation, and number of epochs, respectively. (optimizer) 1: Adadelta, 2: Adagrad, 3: Adam, 4: Adamax, 5: Nadam, 6: RMSProp, and 7: SGD, (activation function) 1: ELU, 2: ReLU, 3: SELU, and 4: SoftPlus, (number of epochs) 1: 400, 2: 600, 3: 800, and 4: 1000, Table S2: The RMSE and prediction graphs of the top 10 DNN and LSTM models for scarlet fever, Table S3: The RMSE and prediction graphs of top 10 DNN and LSTM models for malaria.

Author Contributions

D.L. initiated the research and designed the experiments. S.C. and S.K. analyzed the data with the help of D.L. S.K. contributed to data collection with the help of D.L. S.C. and S.K. wrote the paper with the help of D.L.

Funding

This work was funded by the National Research Foundation of Korea grant number NRF-2016R1E1A2A01953984 and NRF-2018R1E1A2A02059027 and the Korea Ministry of Environment as Climate Change Correspondence R&D Program grant number 2014001300001.

Conflicts of Interest

The authors declare no conflict of interest.

References

  • 1.Jae C.M. Infectious disease, safety, state: History of infectious disease prevention and MERS situation. Crit. Stud. Mod. Korean Hist. 2015;34:517–542. [Google Scholar]
  • 2.Cheng C.K., Lau E.H., Ip D.K., Yeung A.S., Ho L.M., Cowling B.J. A profile of the online dissemination of national influenza surveillance data. BMC Public Health. 2009;9:339. doi: 10.1186/1471-2458-9-339. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Balcan D., Colizza V., Gonçalves B., Hu H., Ramasco J.J., Vespignani A. Multiscale mobility networks and the spatial spreading of infectious diseases. Proc. Natl. Acad. Sci. USA. 2009;106:21484. doi: 10.1073/pnas.0906910106. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Colizza V., Barrat A., Barthelemy M., Valleron A., Vespignani A. Modeling the Worldwide Spread of Pandemic Influenza: Baseline Case and Containment Interventions. PLOS Med. 2007;4:e13. doi: 10.1371/journal.pmed.0040013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Balcan D., Hu H., Goncalves B., Bajardi P., Poletto C., Ramasco J.J., Paolotti D., Perra N., Tizzoni M., van den Broeck W., et al. Seasonal transmission potential and activity peaks of the new influenza A(H1N1): A Monte Carlo likelihood analysis based on human mobility. BMC Med. 2009;1:45. doi: 10.1186/1741-7015-7-45. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Eubank S., Guclu H., Anil Kumar V.S., Marathe M.V., Srinivasan A., Toroczkai Z., Wang N. Modelling disease outbreaks in realistic urban social networks. Nature. 2004;429:180. doi: 10.1038/nature02541. [DOI] [PubMed] [Google Scholar]
  • 7.Ferguson N.M., Cummings D.A.T., Fraser C., Cajka J.C., Cooley P.C., Burke D.S. Strategies for mitigating an influenza pandemic. Nature. 2006;442:448. doi: 10.1038/nature04795. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Epstein J.M., Goedecke D.M., Yu F., Morris R.J., Wagener D.K., Bobashev G.V. Controlling Pandemic Flu: The Value of International Air Travel Restrictions. PLoS ONE. 2007;2:e401. doi: 10.1371/journal.pone.0000401. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Ciofi degli Atti M.L., Merler S., Rizzo C., Ajelli M., Massari M., Manfredi P., Furlanello C., Scalia Tomba G., Iannelli M. Mitigation Measures for Pandemic Influenza in Italy: An Individual Based Model Considering Different Scenarios. PLoS ONE. 2008;3:e1790. doi: 10.1371/journal.pone.0001790. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Zhang Y., Milinovich G., Xu Z., Bambrick H., Mengersen K., Tong S., Hu W. Monitoring Pertussis Infections Using Internet Search Queries. Sci. Rep. 2017;7:10437. doi: 10.1038/s41598-017-11195-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Rohart F., Milinovich G.J., Avril S.M.R., Lê Cao K., Tong S., Hu W. Disease surveillance based on Internet-based linear models: An Australian case study of previously unmodeled infection diseases. Sci. Rep. 2016;6:38522. doi: 10.1038/srep38522. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Lampos V., Miller A.C., Crossan S., Stefansen C. Advances in nowcasting influenza-like illness rates using search query logs. Sci. Rep. 2015;5:12760. doi: 10.1038/srep12760. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Cho S., Sohn C.H., Jo M.W., Shin S.Y., Lee J.H., Ryoo S.M., Kim W.Y., Seo D.W. Correlation between national influenza surveillance data and google trends in South Korea. PLoS ONE. 2013;8:e81422. doi: 10.1371/journal.pone.0081422. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Teng Y., Bi D., Xie G., Jin Y., Huang Y., Lin B., An X., Feng D., Tong Y. Dynamic Forecasting of Zika Epidemics Using Google Trends. PLoS ONE. 2017;12:e0165085. doi: 10.1371/journal.pone.0165085. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Dugas A.F., Jalalpour M., Gel Y., Levin S., Torcaso F., Igusa T., Rothman R.E. Influenza forecasting with Google Flu Trends. PLoS ONE. 2013;8:e56176. doi: 10.1371/journal.pone.0056176. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Towers S., Afzal S., Bernal G., Bliss N., Brown S., Espinoza B., Jackson J., Judson-Garcia J., Khan M., Lin M., et al. Mass Media and the Contagion of Fear: The Case of Ebola in America. PLoS ONE. 2015;10:e0129179. doi: 10.1371/journal.pone.0129179. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Huang D.C., Wang J.F. Monitoring hand, foot and mouth disease by combining search engine query data and meteorological factors. Sci. Total Environ. 2018;612:1293–1299. doi: 10.1016/j.scitotenv.2017.09.017. [DOI] [PubMed] [Google Scholar]
  • 18.Tenkanen H., di Minin E., Heikinheimo V., Hausmann A., Herbst M., Kajala L., Toivonen T. Instagram, Flickr, or Twitter: Assessing the usability of social media data for visitor monitoring in protected areas. Sci. Rep. 2017;7:17615. doi: 10.1038/s41598-017-18007-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Reece A.G., Reagan A.J., Lix K.L.M., Dodds P.S., Danforth C.M., Langer E.J. Forecasting the onset and course of mental illness with Twitter data. Sci. Rep. 2017;7:13006. doi: 10.1038/s41598-017-12961-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Shin S., Seo D., An J., Kwak H., Kim S., Gwack J., Jo M. High correlation of Middle East respiratory syndrome spread with Google search and Twitter trends in Korea. Sci. Rep. 2016;6:32920. doi: 10.1038/srep32920. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Thapen N., Simmie D., Hankin C., Gillard J. DEFENDER: Detecting and Forecasting Epidemics Using Novel Data-Analytics for Enhanced Response. PLoS ONE. 2016;11:e0155417. doi: 10.1371/journal.pone.0155417. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Allen C., Tsou M., Aslam A., Nagel A., Gawron J. Applying GIS and Machine Learning Methods to Twitter Data for Multiscale Surveillance of Influenza. PLoS ONE. 2016;11:e0157734. doi: 10.1371/journal.pone.0157734. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Volkova S., Ayton E., Porterfield K., Corley C.D. Forecasting influenza-like illness dynamics for military populations using neural networks and social media. PLoS ONE. 2017;12:e0188941. doi: 10.1371/journal.pone.0188941. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Simon T., Goldberg A., Aharonson-Daniel L., Leykin D., Adini B. Twitter in the Cross Fire—The Use of Social Media in the Westgate Mall Terror Attack in Kenya. PLoS ONE. 2014;9:e104136. doi: 10.1371/journal.pone.0104136. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Tafti A., Zotti R., Jank W. Real-Time Diffusion of Information on Twitter and the Financial Markets. PLoS ONE. 2016;11:e0159226. doi: 10.1371/journal.pone.0159226. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Xia F., Su X., Wang W., Zhang C., Ning Z., Lee I. Bibliographic Analysis of Nature Based on Twitter and Facebook Altmetrics Data. PLoS ONE. 2016;11:e0165997. doi: 10.1371/journal.pone.0165997. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Patel R., Belousov M., Jani M., Dasgupta N., Winakor C., Nenadic G., Dixon W.G. Frequent discussion of insomnia and weight gain with glucocorticoid therapy: An analysis of Twitter posts. Npj Digit. Med. 2018;1:7. doi: 10.1038/s41746-017-0007-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Xu Q., Gel Y.R., Ramirez Ramirez L.L., Nezafati K., Zhang Q., Tsui K.L. Forecasting influenza in Hong Kong with Google search Queries and statistical model fusion. PLoS ONE. 2017;12:e0176690. doi: 10.1371/journal.pone.0176690. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.He F., Hu Z., Zhang W., Cai L., Cai G., Aoyagi K. Construction and evaluation of two computational models for predicting the incidence of influenza in Nagasaki Prefecture, Japan. Sci. Rep. 2017;7:7192. doi: 10.1038/s41598-017-07475-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Najafabadi M.M., Villanustre F., Khoshgoftaar T.M., Seliya N., Wald R., Muharemagic E. Deep learning applications and challenges in big data analytics. J. Big Data. 2017;2 doi: 10.1186/s40537-014-0007-7. [DOI] [Google Scholar]
  • 31.Janowczyk A., Madabhushi A. Deep learning for digital pathology image analysis: A comprehensive tutorial with selected use cases. J. Pathol. Inform. 2016;7:29. doi: 10.4103/2153-3539.186902. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Esteva A., Kuprel B., Novoa R.A., Ko J., Swetter S.M., Blau H.M., Thrun S. Dermatologist-level classification of skin cancer with deep neural networks. Nature. 2017;542:115. doi: 10.1038/nature21056. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Bychkov D., Linder N., Turkki R., Nordling S., Kovanen P.E., Verrill C., Walliander M., Lundin M., Haglund C., Lundin J. Deep learning based tissue analysis predicts outcome in colorectal cancer. Sci. Rep. 2018;8:3395. doi: 10.1038/s41598-018-21758-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Song Y., Wang F., Wang B., Tao S., Zhang H., Liu S., Ramirez O., Zeng Q. Time Series Analyses of Hand, Foot and Mouth Disease Integrating Weather Variables. PLoS ONE. 2015;10:e0117296. doi: 10.1371/journal.pone.0117296. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Hii Y.L., Rocklöv J., Ng N. Short Term Effects of Weather on Hand, Foot and Mouth Disease. PLoS ONE. 2011;6:e16796. doi: 10.1371/journal.pone.0016796. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Lopman B., Armstrong B., Atchison C., Gray J.J. Host, Weather and Virological Factors Drive Norovirus Epidemiology: Time-Series Analysis of Laboratory Surveillance Data in England and Wales. PLoS ONE. 2009;4:e6671. doi: 10.1371/journal.pone.0006671. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Huang X., Williams G., Clements A.C.A., Hu W. Imported Dengue Cases, Weather Variation and Autochthonous Dengue Incidence in Cairns, Australia. PLoS ONE. 2013;8:e81887. doi: 10.1371/journal.pone.0081887. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Liu T., Zhang Y., Lin H., Lv X., Xiao J., Zeng W., Gu Y., Rutherford S., Tong S., Ma W. A large temperature fluctuation may trigger an epidemic erythromelalgia outbreak in China. Sci. Rep. 2015;5:9525. doi: 10.1038/srep09525. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Blanford J.I., Blanford S., Crane R.G., Mann M.E., Paaijmans K.P., Schreiber K.V., Thomas M.B. Implications of temperature variation for malaria parasite development across Africa. Sci. Rep. 2013;3:1300. doi: 10.1038/srep01300. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Noden B.H., Kent M.D., Beier J.C. The impact of variations in temperature on early Plasmodium falciparum development in Anopheles stephensi. Parasitology. 1995;111:539–545. doi: 10.1017/S0031182000077003. [DOI] [PubMed] [Google Scholar]
  • 41.Liang W., Gu X., Li X., Zhang K., Wu K., Pang M., Dong J., Merrill H.R., Hu T., Liu K., et al. Mapping the epidemic changes and risks of hemorrhagic fever with renal syndrome in Shaanxi Province, China, 2005–2016. Sci. Rep. 2018;8:749. doi: 10.1038/s41598-017-18819-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Huang X., Clements A.C.A., Williams G., Milinovich G., Hu W. A threshold analysis of dengue transmission in terms of weather variables and imported dengue cases in Australia. Emerg. Microbes Amp Infect. 2013;2:e87. doi: 10.1038/emi.2013.85. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Kwon C., Hwang S., Jung J. Monitoring Seasonal Influenza Epidemics in Korea through Query Search. J. Korea Soc. Simul. 2014;23:31–39. doi: 10.9709/JKSS.2014.23.4.031. [DOI] [Google Scholar]
  • 44.Zheng Y.L., Zhang L.P., Zhang X.L., Wank K., Zheng Y.J. Forecast Model Analysis for the Morbidity of Tuberculosis in Xinjiang, China. PLoS ONE. 2015;10:e0116832. doi: 10.1371/journal.pone.0116832. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Myriam G., Philippe Q., Joel G., Sylvie C., Guy L., Laurent G., Laurence M. Time series analysis of dengue incidence in Guadeloupe, French West Indies: Forecasting models using climate variables as predictors. BMC Infect. Dis. 2011;11:166. doi: 10.1186/1471-2334-11-166. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Li Q., Guo N.N., Han Z.Y., Zhang Y.B., Qi S.X., Xu Y.G., Wei Y.M., Han X., Liu Y.Y. Application of an autoregressive integrated moving average model for predicting the incidence of hemorrhagic fever with renal syndrome. Am. J. Trop. Med. Hyg. 2012;87:364–370. doi: 10.4269/ajtmh.2012.11-0472. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Bao W., Yue J., Rao Y. A deep learning framework for financial time series using stacked autoencoders and long-short term memory. PLoS ONE. 2017;12:e0180944. doi: 10.1371/journal.pone.0180944. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Alex G. Supervised Sequence Labelling with Recurrent Neural Networks. Springer; Berlin, Germany: 2012. Studies in Computational Intelligence. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials


Articles from International Journal of Environmental Research and Public Health are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES