Skip to main content
PLOS One logoLink to PLOS One
. 2021 Jul 1;16(7):e0253217. doi: 10.1371/journal.pone.0253217

Bayesian neural networks for stock price forecasting before and during COVID-19 pandemic

Rohitash Chandra 1,*,#, Yixuan He 1,#
Editor: Junhuan Zhang2
PMCID: PMC8248663  PMID: 34197473

Abstract

Recently, there has been much attention in the use of machine learning methods, particularly deep learning for stock price prediction. A major limitation of conventional deep learning is uncertainty quantification in predictions which affect investor confidence. Bayesian neural networks feature Bayesian inference for providing inference (training) of model parameters that provides a rigorous methodology for uncertainty quantification in predictions. Markov Chain Monte Carlo (MCMC) sampling methods have been prominent in implementing inference of Bayesian neural networks; however certain limitations existed due to a large number of parameters and the need for better computational resources. Recently, there has been much progress in the area of Bayesian neural networks given the use of Langevin gradients with parallel tempering MCMC that can be implemented in a parallel computing environment. The COVID-19 pandemic had a drastic impact in the world economy and stock markets given different levels of lockdowns due to rise and fall of daily infections. It is important to investigate the performance of related forecasting models during the COVID-19 pandemic given the volatility in stock markets. In this paper, we use novel Bayesian neural networks for multi-step-ahead stock price forecasting before and during COVID-19. We also investigate if the pre-COVID-19 datasets are useful of modelling stock price forecasting during COVID-19. Our results indicate due to high volatility in the stock-price during COVID-19, it is more challenging to provide forecasting. However, we found that Bayesian neural networks could provide reasonable predictions with uncertainty quantification despite high market volatility during the first peak of the COVID-19 pandemic.

1 Introduction

Stock price prediction is a challenging research area [1] due to multiple factors affecting the stock market that range from politics [2], weather and climate, and international and regional trade [3]. Machine learning methods such as neural networks have been widely used in stock forecasting [4]. Some studies show that neural networks outperforms statistical methods, such as multiple linear regression analysis [5], discriminant analysis [6] and related methods. Obtaining robust quantification of uncertainty with good prediction accuracy has been a major challenge for effective stock market prediction models.

Bayesian inference offers a methodology for robust estimation uncertainty quantification of parameters in prediction models. Bayesian inference enables models to feature uncertainty quantification in predictions using posterior distributions to represent the unknown parameters [7]. The probability of hypothesis in Bayesian inference is updated with evidence or data as it become available [8]. We can obtain the posterior distribution by sampling methods that take into account the prior distribution and the likelihood function to evaluate the model with given data [9]. Markov Chain Monte Carlo (MCMC) methods implement Bayesian inference by sampling from the posterior distribution [7, 10]. Bayesian neural networks use MCMC methods to estimate (train) neural network parameters (weights and biases) [11, 12]. MCMC methods face limitations as the size of model and data increase, due to the curse of dimensionality. Therefore, Hamiltonian MCMC [13] and Langevin dynamics based MCMC [14] that feature gradient-based proposal distributions were proposed to improve canonical MCMC methods. Parallel tempering MCMC with Langevin-gradients has been very promising for Bayesian neural networks [15]; therefore, it has the potential for forecasting the stock market.

The coronavirus disease 2019 (COVID-19) is an infectious disease [1618] which became a global pandemic [19]. The first confirmed or index case of COVID-19 was traced back to 17th November in Wuhan, Hubei, China that became known in December 2019 [20]. The COVID-19 pandemic forced many countries to close their borders and enforce a partial or full lock down which had a devastating impact on the world economy that can continue for years to follow [2123]. The lock downs and economic impact affected population depending on agriculture in rural areas, especially in developing countries [24, 25]. Several machine learning methods have been utilized for COVID-19 infection prediction in several countries [2633], and the impact of COVID-19 on the economy has also been studied.

In the literature, there is no work done using Bayesian neural networks for stock markets to provide uncertainty quantification in predictions. We can leverage these methods to harness power of neural networks in prediction accuracy and also to quantify uncertainty in predictions. Moreover, it is worthwhile to evaluate how neural network models perform during the COVID-19 pandemic given drastic changes in the international stock market with disruptions in international trade and prediction. Hence, it is important to investigate the performance of related forecasting models during the COVID-19 pandemic given volatility in stock markets.

In this paper, we use novel Bayesian neural networks for multi-step-ahead stock price prediction before and during COVID-19. We compare our forecasting results with state-of-art neural network training algorithms. Our training data features stocks from four different regions, that include Germany, China, Australia and the United States. We select stocks from these countries due to the effect of various types of lock downs during the course of the first phase of the COVID-19 pandemic that affected their gross domestic product (GDP). We restrict our study to selected stocks from these countries due to geopolitical status and diverse GDP forecasts during the pandemic [23]. Moreover, one of the selected stocks feature COVID-19 mask manufacturing company, while others feature regarding luxury goods for a diverse stock-market analysis. We investigate if the pre-COVID-19 datasets can provide any insights in modelling stock price forecasting during first phase of COVID-19. We compare the prediction performance pre-COVID-19 with results during COVID-19 to evaluate the ability of Bayesian neural networks given drastic changes in the stock price.

We note that although there are many studies in the literature regarding COVID-19 forecasting with machine learning methods, the use of Bayesian neural networks is limited. Moreover, most studies focus on the spread of the infections rather than stock market prediction. The novelty in this study lies in the use of Bayesian neural networks which provides better model uncertainty quantification when compared to classical neural networks.

The rest of the paper is organised as follows. Section 2 presents a review of related work. Section 3 presents the proposed methodology and Section 4 presents experiments and results. Section 5 provides a discussion and Section 6 concludes the paper with discussion of future work.

2 Related work

2.1 Stock market and price forecasting

There are two types of forecasting methods, which are fundamental analysis techniques [34] and technical analysis [35]. Fundamental analysis techniques measure intrinsic value of the stock by examining relevant data of the company such as audit reports, book value and price-to-earnings ratio (P/E ratio) [36]. Technical analysis attempts to predict stock markets using charts and quantitative indicators [37]. Highest and lowest values of stock price of a day, volume of stock, simple moving average, and exponential moving average can be considered as the financial indicators in the technical analysis. These are more suitable as the input for machine learning methods than fundamental analysis [38, 39].

Due to the substantial increase of the computational power, machine learning has become popular stock market forecasting method [40]. Some of the prominent machine learning methods include support vector machines (SVM) and neural networks. Schumaker and Chen [41] used SVM to classify the direction (rise or drop) of future stock prices. Lin et al. [42] proposed a quasi-linear SVM method for stock market prediction, which selected the subset of financial indexes as the model’s weighted inputs. Devi et al. [43] employed metaheuristic (cuckoo optimisation) method for training SVM parameters for stock market forecasting. Dase et al. [44] employed neural networks to improve stock market forecasting accuracy. Liao et al. [45] incorporated stochastic time effective function with neural networks and used different volatility parameters to assess the predictive performance of the model. Moghaddam et al. [46] employed neural networks for two type of input datasets in order to forecast the daily stocks of the NASDAQ stock exchange. Chopra et al. [47] subdivided nine stocks based on volatility and market capitalization and demonstrated that neural networks have good ability for stock price forecasting before and after demonetization in India.

Evolutionary optimisation methods such as genetic algorithms (GAs) have widely been been used to train neural networks and SVMs [48]. Khatibi et al. [49] presented combination of GAs with SVMs which used various financial indicators as input features that provided better performance when compared to neural networks alone. Qiu et al. [50] combined GAs with neural networks to enhance the accuracy of stock market forecasting index. Moreover, neural networks trained by hybrid of simulated annealing and GAs significantly enhanced the prediction accuracy over traditional backpropagation neural networks [51].

Furthermore, other hybrid methods that fall in the field of artificial intelligence and machine learning have also been used in stock market forecasting. Guresen et al. [52] presented a comparison amongst multilayer perceptron, dynamic neural networks, and hybrid neural networks that featured generalized autoregressive conditional heteroscedasticity (GARCH) model. The authors reported that multilayer perceptron provided best performance. Rathnayaka et al. [53] presented a hybrid model based on neural networks and autoregressive integrated moving average (ARIMA) for forecasting the Colombo stock exchange which provided better predictive ability under a high volatility than conventional time series forecasting methods. Zhong and Enke [54] applied three dimensionality reduction techniques which include principal component analysis (PCA), fuzzy robust principal component analysis (FRPCA), and kernel-based principal component analysis (KPCA) with neural networks to estimate the daily direction of the future stock market returns. The authors reported that the combination of neural networks and PCA provides more accurate results when compared to the other combinations.

2.2 Neural networks for forecasting

Deep learning refers to a special type neural networks which consists of multiple processing layers and enables high-level abstraction to model data [55]. In the literature, deep learning commonly refers to the outstanding models such as recurrent networks (RNNs), convolutional neural networks (CNNs), deep belief networks, and long short-term memory networks (LSTMs) [56, 57]. In recent years, with more computing power and massive datasets, deep learning models have demonstrated excellent performance in different fields, such as sentiment analysis [58], image analysis [59] and natural language processing [60].

The main advantage of deep learning models is the ability to automatically extract the good features of input data through the general-purpose learning procedure [61]. Therefore, deep learning has also been widely used in various forecasting applications. Amarasinghe et al. [62] investigated the effectiveness of CNNs for individual building level energy load forecasting. Huang and Kuo [63] combined CNNs and LSTMs for air pollution (PM 2.5) forecasting. Sudriani et al. [64] utilized LSTMs for forecasting discharge level of Cimandiri River which was beneficial for managing water resources.

The financial community has received a boost in developing solutions with deep learning models for financial forecasting research. Ding et al. [65] utilized CNNs to evaluate the impact of different events on stock price behavior in the short, middle and long term. Nelson et al. [66] used LSTMs to forecast the future trends of stock market based on the price history and technical analysis indicators. Apart from these, innovative approaches for training conventional neural networks have been utilised. Chandra et al. [67] used co-evolutionary RNNs for stock market forecasting and proposed framework for mobile application.

Bayesian neural networks have strength in forecasting due to promising prediction accuracy with uncertainty quantification. Different Bayesian neural networks such as recursive Bayesian recurrent neural networks [68] and evolutionary MCMC Bayesian neural networks [69] have been used for time series forecasting. Liang et al. [70] proposed an MCMC algorithm for neural networks for selected time series problems. Chandra et al. [15] presented Langevin gradient Bayesian neural networks with parallel tempering MCMC, which used high-performance computing for time series prediction. Bayesian neural networks have been applied to various fields such as railway passenger flow [71], a certain index of the national economy [72], and short-term commodity prices [73]; these applications have reported promising forecasting performance.

2.3 COVID-19 impact on world economy

As mentioned earlier, the first phase of COVID-19 has a devastating impact on the world economy and the stock market. Ahmar et al. [74] presented a study for forecasting effect of COVID-19 on stock market in Spain using ARIMA model. The authors combined the ARIMA model with α-Sutte indicator which uses 4 previous data points for forecasting and was more suitable when compared to ARIMA model alone. They also reported that the increase in the number of COVID-19 cases had direct effect on the stock market. Ali et al. [75] utilized the GARCH model to evaluate the volatility of the financial markets during the transfer of COVID-19 from China to Europe and then to the United States. The authors also performed a bivariate regression between the returns and volatility of various financial securities during the first phase of COVID-19. The results show that China gradually stabilized, while the global market has experienced a sharp drop of financial security with the spread of the epidemic. In another study, the researchers tried to build a predictive model to assess the relationship between health-related news and stock returns in the worst-hit countries by COVID-19 [76].

Moreover, Maliszewska et al. [23] evaluated the effect of COVID-19 on gross domestic product (GDP) and trade in four major areas that included the reduction in employment rate and capital demand, higher costs in international trade, sharp decrease in international tourism, and declining demand for services that require close human interaction. The economic model which is conceptually similar to the approach of modelling severe acute respiratory syndrome (SARS) outbreak in 2002 was applied to approximate the potential impact of COVID-19 on the global economy [77]. Nicola et al. [78] analyzed in detail the socio-economic impact of the COVID-19 on various industries and focused on the primary industry related to raw material mining, secondary industry related to finished product production, and tertiary industry including service-providing industries. McKinbbin and Fernando [79] predicted seven different scenarios of how COVID-19 might evolve in the coming year, which indicated that the outbreak can have a significant impact on the global economy in the short term. Guan et al. [80] utilized the latest global trade modelling framework to analyze the supply-chain effects on a number of idealized lock-down scenarios.

3 Methodology

3.1 State-space reconstruction

State-space reconstruction refers to embedding a time series into a vector so that it can be trained by machine learning models [81]. According to the Taken’s theorem, the embedding process must ensure that the original characteristics of the time series is retained [82]. Given a univariate time series, we can construct a multi-dimensional space vector by taking a point on the fixed delay of the original system. Using Taken’s embedding theorem [82], the state-Space reconstruction is given as follows.

Suppose the actual series of closing stock price is [x1, x2, …, xN], where N is the length of the series. First, we choose the embedding dimension m and a time lag T, and then capture windows of size m denoted by vector x¯ for every T delay until N is reached. Our problem is multi-step prediction where we have n prediction horizons denoted by vector y. The reconstructed vector by state-space embedding is denoted by [x¯,y]. Hence, for the first instance, we have

x¯1=[x1,x2,x3,x4,,xm]
y1=[xm+1,xm+2,xm+3,,xm+n]

In the same way, we can obtain the rest of the instances for the entire time series as given below.

x¯t=[x1+(t1)T,x2+(t1)T,x3+(t1)T,x4+(t1)T,,xm+(t1)T]
yt=[xm+(t1)T+1,xm+(t1)T+2,xm+(t1)T+3,,xm+(t1)T+n]

3.2 Neural networks

Canonical neural networks which are also known as multi-layer perception employs multiple layers in the model, where each layer features neurons that propagate information for layers ahead a shown in Fig 1. Each neuron can receive the signal of the neuron in the previous layer and generate the output to the next layer. The first layer is called the input layer, the last layer is called the output layer, and the other intermediate layers are called hidden layers. The hidden layer could be one layer or feature multiple layers.

Fig 1. The time series (shown in red circles) is used as input for the neural network which predicts 5 steps-ahead in time (shown by black circles).

Fig 1

A sliding window approach is used to reconstruct the dataset in this way using Taken’s theorem.

A neural network model f(x) can be defined as a composition of other functions. Given a series of input-output pairs {x¯t,yt}, the model is trained to approximate the function f such that f(x¯t)=yt for all pairs. In our setting,

f(x¯t)=g(δo+h=1Hvh×g(δh+i=1mwihx¯t,i)), (1)

where, m is the input number and H is the number of hidden layers. The function g(.) is the sigmoid activation function which is used in the hidden and output layers. The setup for multi-step ahead time series prediction problem using neural networks with one hidden layer is shown in Fig 1. The complete set of parameters for the neural network model is shown in Fig 1 θ=(w˜,v˜,δ), where δ = (δo, δh). w˜ is the weight of the input to hidden layer. v˜ is the weight of the hidden to output layer. δh is the bias for the hidden layer, and δo is the bias for the output layer.

Stochastic gradient descent (SGD) is one of the prominent methods of training neural networks. SGD is an iterative method to optimize a differentiable objective function with help of gradients [83]. In some high-dimensional optimization problems, SGD reduces the computational burden by achieving faster iterations with a lower convergence rate [83]. Training neural networks also can be considered as solving the non-convex optimization problem:

arg min L(w), where wRn is the set of parameters and L is the loss function. The iterations of SGD can be given as

wk=wk1ak1L(wk1)

where, wk denotes the kth iteration, ak is the learning rate, and L(wk) denotes the gradient at wk.

We note that the learning rate is user defined parameter which depends on the type of problem and typically it is determined in trial experiments. Hence, extension of the SGD consider adapting the learning rate automatically during the learning process. Adaptive moment estimation (Adam) is an effective stochastic optimization method that only requires first-order gradients with a small amount of memory requirement focused in adapting learning rate [84]. Adam calculates the individual adaptive learning rates of the parameters from the estimates of first and second moments of the gradients. The Adam-based weight update is expressed as follows

wk=wk1ak1·1β2k1β1k·uk1nk1+ϵ

where

uk1=β1uk2+(1β1)f(wk1) (8)
nk1=β2nk2+(1β2)f(wk1)2

where, β1 and β2 are first and second moment estimates, respectively. ϵ is a small scalar used to prevent division by 0.

3.3 Bayesian neural networks

Bayesian neural networks provide a probabilistic implementation of a standard neural network with the key difference where the weights and biases are represented via posterior probability distributions rather than single point estimates [85, 86] as shown in Fig 2. Similar to standard neural networks, Bayesian neural networks also have universal continuous function approximation capabilities.

Fig 2. Bayesian neural network and MCMC sampling.

Fig 2

Note that the posterior distribution is shown that represents weights in Panel (a).

The challenge of Bayesian inference is in sampling to approximate (learn) the posterior distribution of neural network weights and biases. The inference procedure begins by setting prior distributions over the weights and biases. Then the sampling scheme (such as MCMC) employs a likelihood function that takes into account the training data accepting or rejecting a proposed sample. The implementation of Bayesian neural network using MCMC sampling is shown in Fig 2. Due to non-linear activation functions in the Bayesian neural network, the conjugacy of prior and posterior is lost. Moreover, due to large number of parameters given different applications, it is difficult to get informative priors and hence, it is challenging is to sample the posterior distribution.

We can construct the likelihood function using the set of weights and biases (Eq (1))) θ, for M network parameters and S training instances. We note that we use a signal plus noise model where an additional parameter (τ2) is used to cater for the noise. Hence, for model output f(x¯t) with given input features xt¯, we have

p(yS|θ)=1(2πτ2)S/2×exp(12τ2tS(ytf(x¯t))2) (2)

which satisfies the multivariate probability density function.

The prior is based on Gaussian distribution in the case of θ and Gamma distribution in the case of τ2 as shown below.

p(θ)1(2πσ2)L/2×exp{12σ2(i=1Mθ)}×τ2(1+ν1)exp(ν2τ2) (3)

where, σ2 is determined by exploring variance in weights and biases of trained neural networks for similar applications. Moreover, ν1 and ν2 are user defined constants.

3.4 Sampling using parallel tempering MCMC

Parallel tempering MCMC features an ensemble of replica samplers that can run in parallel and has the ability to sample multi-modal posterior distributions [87, 88]. A user defined temperature ladder corresponds to every replica in the ensemble, where the higher temperature values have higher probability to accept weaker proposals. The ensemble is defined by a total of R replicas at temperature level Tm specified by

Θ=(θ[1],,θ[R])

where m denotes the replica. We sample θ from the posterior distribution by proposing θp from a known distribution q(θ). Given the proposed value θp, the chain moves with a probability α or remains at its current position θk. We note that α is chosen to ensure that the chain is reversible and has stationary distribution p(θ|D) given data D.

Algorithm 1 provides further details where parallel tempering MCMC is used for global exploration which then transforms to canonical MCMC via parallel computing for local exploration. The transformation is done by changing the temperature ladder to series of 1’s. The local exploration also features exchange of neighboring replicas. The user needs to set the percentage of samples for global exploration phase in advance along with hyper-parameters like the maximum number of samples (Maxsamples), and the swap interval (Swapint) which is typically set to a few iterations to support efficient inter-process communications in parallel computing environment as shown in Fig 3. After every few iterations (ad defined by Swapint), the algorithm determines if the neighboring replicas require to be swapped which is necessary to improve the efficiency of exploring the posterior distribution. The replica swap is determined by the Metropolis-Hasting acceptance criterion which is similar to within replica transition. After the termination condition is met, the respective replica posterior distributions are combined after discarding the burn-in period which is marked by the parallel tempering MCMC global exploration phase. In this way, we ensure that only the true posterior (replica’s with temperature of 1’s) are part of the posterior distribution.

Fig 3. An overview of the different replicas that are executed on a parallel computing architecture.

Fig 3

We use Langevin-gradient proposal distribution [89] that essentially features a one-step gradient over Gaussian noise. At a given chain position k, our new proposal θp is given as follows

θpN(θ¯k,Σθ),where (4)
θ¯k=θk+r×E(θk), (5)
E=iS(yif(x¯i))2 (6)
E(θ[k])=(Eθ1,EθL) (7)

where r is the learning rate, Σθ=σθ2I, I is the L × L identity matrix, x¯ is the univariate time series input data vector (window) denoted by i for S data instances, yi is the time series data vector for h prediction horizons, and L refers to the total number of model parameters (weights and biases). Based on a user defined probability ϕ, the proposal θp can be either

  • A one-step gradient descent based weight update known as Langevin-gradient (LG) proposal distribution (Eq 4),

  • A random-walk (RW) proposal distribution where Gaussian noise from distribution centered at mean of 0 and variance, N(0,Σθ).

We note that the ensemble of replicas execute in parallel that have stationary distributions which are equal to up to a proportionality constant defined by the temperature ladder, p(θ|D)β; where β ∈ [0, 1] corresponds to the temperature ladder that features geometric spacing. β = 0 corresponds to a uniform stationary distribution and β = 1 refers to the posterior. Hence, the replicas that feature smaller temperature levels from β can provide global exploration, while those with higher values provide local search or exploitation. For each replica m in the ensemble, we propose θmp using Langevin-gradient conditional on current value θmk, that is θmkq(θ|θmk). The within replica transition determines if the proposed value of θmp remains at its original location θmk or gets updated by a probability as given

α=min(1,p(θmp|D)βmq(θmk|θmp)p(θmk|D)βmq(θmp|θmk)). (8)

The proposal becomes part of the posterior distribution once it is accepted.

Algorithm 1: Langevin-gradient Parallel Tempering MCMC

Result: Draw samples from distribution p(θ|D)

i. Define maximum number of samples: Maxsamples, swap interval: Swapint and number of replicas: R

ii. Initialize θm=θm[0] for each replica m

iii. Define the Langevin-gradient probability ϕ

iv. Define percentage of samples for global exploration

v. Define the temperature ladder that uses geometric spacing, β.

while Maxsamples do

for m = 1, …, R do

  for k = 1, …, Swapint do

   Step 1: Within Replica sampling

   1.1 Propose new sample (solution)

   Draw lU[0, 1]

   if ((l < ϕ) then

    Get θmp using Langevin-gradient proposal distribution (Eq 4)

   else

    Get θmp using Random-Walk proposal distribution (N(0,Σθ))

   end

   1.2 Compute acceptance probability α (Eq 8)

   1.3 Acceptance criterion Draw uU[0, 1]

   if u < α then

    θmk=θmp

   else

    θmk=θmk1

   end

  end

end

Step 2: Replica exchange

 2.1.Compute acceptance probability for neighboring replicas

 2.2. Exchange neighboring replica if accepted.

 *Local/Global exploration phase

if global is true then

  βm = βm

else

  βm = 1

end

end

4 Experiments and results

In this section, we provide details about the datasets and present research design with computational results.

4.1 Data

We choose 4 stocks from 4 different countries to evaluate the performance of the respective methods. These include 3M Company(MMM) from United States, China Spacesat Company Limited (600118.SS), Commonwealth Bank of Australia (CBA.AX), and Daimler AG (DAI.DE) from Germany. The respective datasets feature the closing price with the time period from 01/01/2012 to 01/07/2020 from Yahoo Finance. MMM is an American world-renowned multinational company with diversified products, which cover various fields such as household goods and medical supplies. China Spacesat Company Limited is an aerospace high-tech enterprise specializing in the development and application of small satellites in China. Commonwealth Bank is the largest commercial bank in Australia. Daimler AG is a German company which is the largest commercial vehicle manufacturer and the largest luxury car manufacturer in the world. In the case of China Spacesat Company Limited, the data after January 2020 is affected by COVID-19. In the case of the remaining companies, the data after March 2020 is affected by COVID-19.

We apply data normalization to the original data time series data to a boundary of [0, 1] using min-max normalization given as

xi=xixminxmaxxmin, (9)

where, x is the adjusted closing price time series.

4.2 Experiment setup

In parallel tempering MCMC, we use burn-in rate of 0.5 with maximum of 100,000 samples with maximum temperature value of 2. Note that the burn-in rate also defines the first-phase where global search is enforced by parallel tempering in of Algorithm 1. We use 10 replicas with swap interval is 5. The Langevin-gradient proposals use learning rate of 0.1 and applied with a probability of 0.5. We use one hidden layer feedforward neural network with 5 output units where each output unit denotes a prediction horizon for the 5-step-ahead stock blackprice prediction problems. We apply Taken’s embedding theorem to reconstruct data with dimension D = 5 and time-lag T = 2. Hence, 5 input neurons with 5 hidden neurons are used in the respective neural network models.

The experiments are executed as follows.

  • Evaluate multi-step-ahead stock price prediction using novel Bayesian neural networks before COVID-19.

  • Compare the results with feedforward neural network using Adam optimiser (FNN-Adam) and feedforward neural network using stochastic gradient descent (FNN-SGD) training algorithm.

  • Evaluate multi-step-ahead stock price prediction using novel Bayesian neural networks during COVID-19.

The results report the mean and 95% confidence interval from 30 experimental runs with different weight and bias initialisation in the respective models for 5-step ahead prediction. The prediction performance is measured by root mean squared error(RMSE) given as follows.

RMSE=1Ni=1N(yiyi^)2 (10)

where, yi is the actual value and yi^ is the predicted value, and N is the number of data-points for a single prediction horizon (step).} We use RMSE since it is one of the key performance measures for time series forecasting in the literature. Other measures such as Mean Absolute Error (MAE), and Normalised Mean Squared Error (NMSE) can also be used. Note that in our past work [90], we found that NMSE for example, does not change the key conclusions for time series problems, hence we used RMSE only.

4.3 Prediction results pre-COVID-19

The first 80% of the data is used for training and remaining is used for testing. Table 1 gives further details of the time frame considered and indicates the exact dates for the respective stocks.

Table 1. Time span of data considered for each stock-price pre-COVID-19.

MMM 600118.SS CBA.AX DAI.DE
Train (80%) 3.1.2012-24.5.2018 4.1.2012-28.5.2018 3.1.2012-24.7.2018 2.1.2012-22.5.2018
Test (20%) 25.5.2018-31.12.2019 29.5.2018-31.12.2019 25.7.2018-31.12.2019 23.5.2018-30.12.2019

Fig 4 reports the performance of the respective method in prediction of future trends of given stock prices that include MMM, 600118.SS, CBA.AX and DAI.DE, respectively. The results show the prediction horizon (step), the mean RMSE with a 95% confidence interval as error-bars. Table 2 further presents the results numerically.

Fig 4. Prediction performance (mean of normalized RMSE) for the three models (Bayes-FNN, FNN-Adam, and FNN-SGD).

Fig 4

Note that the predictions were normalized in range of [0, 1].

Table 2. Multi-step-ahead prediction (RMSE).

Problem Step Bayes-FNN FNN-Adam FNN-SGD
MMM 1 0.03669±0.00577 0.03800±0.00305 0.03889±0.00073
2 0.04173±0.00413 0.04221±0.00274 0.04305±0.00032
3 0.04434±0.00654 0.04491±0.00249 0.04555±0.00067
4 0.04495±0.00404 0.04826±0.00237 0.04902±0.00050
5 0.04975±0.00652 0.05047±0.00249 0.05107±0.00031
600118.SS 1 0.01789±0.00139 0.01669±0.00331 0.01800±0.00036
2 0.01833±0.00098 0.01860±0.00269 0.01900±0.00032
3 0.01852±0.00128 0.01908±0.00251 0.01904±0.00035
4 0.01954±0.00081 0.02066±0.00202 0.02015±0.00045
5 0.01891±0.00087 0.02088±0.00200 0.01980±0.00044
CBA.AX 1 0.02551±0.00136 0.02886±0.00067 0.02597±0.00028
2 0.02942±0.00058 0.03270±0.00037 0.03021±0.00031
3 0.03475±0.00050 0.03820±0.00031 0.03429±0.00021
4 0.03628±0.00043 0.03996±0.00017 0.03616±0.00020
5 0.04034±0.00055 0.04409±0.00033 0.03929±0.00020
DAI.DE 1 0.02743±0.00131 0.02871±0.00194 0.03212±0.00037
2 0.03228±0.00078 0.03378±0.00163 0.03626±0.00046
3 0.03771±0.00109 0.03918±0.00142 0.04150±0.00050
4 0.04204±0.00136 0.04261±0.00121 0.04410±0.00035
5 0.04562±0.00115 0.04674±0.00104 0.04836±0.00033

In the case of MMM (Fig 4(a)), Bayesian neural network (Bayes-FNN) performs the best with the lowest RMSE compared to FNN-Adam and FNN-SGD where the error increases with the prediction horizon. Fig 4(b) presents results for stock 600118.SS that show that Bayes-FNN gives best performance in all the prediction horizons, expect for 1. The error increases with the prediction horizon. In Fig 4(c), Bayes-FNN shows the best performance in the prediction horizon 1 and 2, which is overtaken by FNN-Adam in the rest, while the RMSE increases with the prediction horizon. In the case for stock DAI.DE, Fig 4(d) shows that Bayes-FNN gives best performance and the RMSE increases as the prediction horizon increases.

Fig 5 (stock MMM), Fig 6 (stock 600118.SS), Fig 7 (stock CBA.AX), and Fig 8 (stock DAI.DE) show the prediction on the test dataset for the prediction horizons 1, 2 and 5 using Bayes-FNN. We notice that the uncertainty (shaded) is relatively small for stock DAI.DE and CBA.AX when compared to stock MMM and 600118.SS. We notice that Bayes-FNN gives very accurate predictions for horizon 1 and 2 in case of stock DAI.DE and CBA.AX and hence, there is lower uncertainty. In case of stock 600118.SS, there is similar level of accuracy in prediction horizons 1 and 2 but higher uncertainty while the stock MMM shows poor prediction and uncertainty quantification in the first half (less than 100 days).

Fig 5. Prediction and uncertainty (shaded region) for test data of stock MMM.

Fig 5

Fig 6. Prediction and uncertainty (shaded region) for test data of stock 600118.SS.

Fig 6

Fig 7. Prediction and uncertainty (shaded region) for test data of stock CBA.AX.

Fig 7

Fig 8. Prediction and uncertainty (shaded region) for test data of stock DAI.DE.

Fig 8

4.4 Results during COVID-19

Next, we apply Taken’s embedding theorem to reconstruct data with dimension D = 5 and time-lag T = 1. The previous section featured data (training and test set) before the COVID-19 pandemic. In certain stocks, we included January and February 2020 data for training since COVID-19 was not widespread in countries such as USA, Germany and Australia during that period. We provide an investigation to check how the stock price changes during COVID-19 and effect of the stock price trend before COVID-19 on the stock price during COVID-19. Hence, we use the data previous of COVID-19 pandemic for different stocks as training data to set the model performance during COVID-19 and refer to it as Setup-1. Table 3 gives details of the dates considered for the training and test dataset for the respective stocks. In Setup-2, we include parts of the data during COVID-19 in the training set with all the training data from Setup-1 where the exact dates are given in Table 3. In general, we appended Setup-1 with data from March and April, 2020 that covers the first phase of the pandemic in the respective countries that affected the stocks. The major reason for doing this is to ensure that the training dataset covers the stock trend during the pandemic. Hence, our test datasets of both setups are different. We have entire COVID-19 dataset used as test dataset for Setup-1, while second half of COVID-19 data is used as test dataset in Setup-2. In this case, we only provide results using the Bayes-FNN method.

Table 3. Timespan considered for respective stocks during COVID-19.

Data MMM 600118.SS CBA.AX DAI.DE
Setup 1 Train (80%) 26.10.2018-28.2.2020 31.1.2018-31.12.2019 31.10.2018-28.2.2020 7.11.2018-28.2.2020
Test (20%) 2.3.2020-30.6.2020 2.1.2020-29.6.2020 2.3.2020-30.6.2020 2.3.2020-29.6.2020
Setup 2 Train (80%) 6.9.2019-30.4.2020 16.4.2019-31.3.2020 2.9.2019-30.4.2020 10.9.2019-30.4.2020
Test (20%) 1.5.2020-30.6.2020 1.4.2020-29.6.2020 1.5.2020-30.6.2020 4.5.2020-29.6.2020

Fig 9 illustrates the performance of the Bayes-FNN method in forecasting the four stocks mentioned before respectively. Setup-1 (left) and Setup-2 (right) bar charts show the prediction horizon and the mean RMSE with a 95% confidence interval. Setup-1 features prediction during the whole COVID-19 period, and Setup-2 features predictions for the second half of the COVID-19. In general, RMSE for all the given stocks improved by reducing for Setup-2. Fig 9(b)) shows that Setup-2 gives a better performance with much lower error (RMSE) when compared to Setup-1 in Fig 9(a). Generally the error get larger with the prediction horizon during Setup 1, but there is not a clear trend when compared to Setup-2. In the case of stock MMM for Setup-1, prediction horizon 3 has the lowest error, while stock CBA.AX and DAI.DE show the best performance in the prediction horizon 1. In the case of stock 600118.SS, the performance in the prediction horizon 2 is the best. Given that we add the first half of time period affected by COVID-19 into the training set, the error during Setup-2 significantly decreases, but become larger with the prediction horizon which is natural for multi-step ahead prediction. Among the respective stocks, the prediction of stock 600118.SS performs the best for Setup-2. Table 4 further reports the results numerically.

Fig 9. Prediction performance (mean of normalized posterior RMSE) of respective stocks during COVID-19 using Bayes-FNN (Setup 1 vs Setup 2).

Fig 9

Note that the predictions were normalized in range of [0, 1].

Table 4. Multi-step-ahead prediction (RMSE) during COVID-19.

Data Step MMM 600118.SS CBA.AX DAI.DE
Setup 1 1 0.10962±0.01153 0.23707±0.00748 0.11478±0.01281 0.12481±0.01401
2 0.10516±0.00673 0.23278±0.01837 0.17165±0.02852 0.16664±0.03704
3 0.08305±0.00319 0.23936±0.00457 0.15478±0.01897 0.14751±0.03054
4 0.09534±0.00385 0.26333±0.00517 0.19222±0.02748 0.21195±0.03008
5 0.09592±0.00186 0.27321±0.00933 0.15961±0.01077 0.16994±0.02151
Setup 2 1 0.07758±0.00533 0.05916±0.00153 0.06984±0.00165 0.06024±0.00458
2 0.09938±0.00567 0.05846±0.00089 0.08072±0.00391 0.07658±0.00394
3 0.11044±0.00154 0.06571±0.00075 0.10260±0.00108 0.08375±0.00398
4 0.11275±0.00149 0.06779±0.00018 0.10231±0.00723 0.09637±0.00783
5 0.11509±0.00090 0.07300±0.00035 0.10410±0.00465 0.09903±0.00927

Fig 10 (stock MMM), Fig 11 (stock 600118.SS), Fig 12 (stock CBA.AX) and Fig 13 (stock DAI.DE) show the prediction on test dataset for the prediction horizons 1, 2 and 5 using Bayes-FNN with Setup-1 and Setup-2. We notice that the uncertainty (shaded) is lower for stocks DAI.DE and MMM for Setup 2 when compared to Setup 1. It is visually clear that the prediction in Setup-2 is generally better when compared to Setup-1.

Fig 10. Prediction and uncertainty over test data of MMM (data setup 1 vs data setup 2).

Fig 10

Fig 11. Prediction and uncertainty over test data of 600118.SS (data setup 1 vs data setup 2).

Fig 11

Fig 12. Prediction and uncertainty over test data of CBA.AX (data setup 1 vs data setup 2).

Fig 12

Fig 13. Prediction and uncertainty over test data of DAI.DE (data setup 1 vs data setup 2).

Fig 13

The monthly volatility for the respective stock is presented with the red shaded area featuring the period during outbreak of COVID-19 (Figs 14 and 15). We measure volatility by the variance between returns from the particular stock index [91] and assume that there is 21 trading days per month. The respective stocks show high volatility in general during COVID-19. In Fig 14 after certain weeks (Fig 14(a) and 14(c)). In the other two stocks (Fig 15(a) and 15(c)), we notice a sharper drop sharply which are bank and luxury goods companies.

Fig 14. Stock price time series and monthly volatility for stocks MMM and 600118.SS.

Fig 14

Fig 15. Stock price time series and monthly volatility for stock CBA.AX and DAI.DE.

Fig 15

5 Discussion

The results show that prior to outbreak of COVID-19, Bayes-FNN provides one of the best performance which would be due to global-local exploration features of parallel tempering MCMC taking into account Langevin-gradient proposals. The accuracy decreased significantly as the prediction horizon increased. This is not surprising when taking into account other time series problems from in the literature by Chandra et al. [92] for one-step ahead prediction. Our work extended Bayes-FNN using parallel tempering MCMC used for one-step ahead prediction into multi-step ahead prediction for a challenging problem of stock price forecasting before and during COVID-19. We learned that Bayes-FNN scales well for multi-ahead prediction and provides better or competing performance when compared to state-of-art method (FNN-Adam). We note that typically stock price prediction does not have such a high level of uncertainty as weather predictions [93], however the situation during COVID-19 has significantly changed as shown by the volatility in Fig 14. The investors need to have confidence in the particular stock and rigorous uncertainty quantification in prediction. We know that good prediction accuracy is needed not just for the day ahead, but the behaviour of the stock days ahead is also of interest. A number of automated forecasting models are part of stock markets [94] and areas of machine learning such as reinforcement learning has made such automation possible [95]. Providing uncertainty quantification in automated stock forecasting models would give information to investors for better decision-making.

At the beginning of the COVID-19 pandemic, the stocks of various countries were greatly affected [96] which is quite clear from monthly volatility visualisation in Fig 14. We note that volatility statistical measure of the dispersion of returns for a stock where the higher the volatility implies riskier security [97]. The volatility also refers to uncertainty or risk related to the changes in the stock [97]. The results during COVID-19 show that the uncertainty in the stock price is much higher due to the high volatility in the stock price. After training the model by adding part of data featuring high volatility (during COVID-19), the prediction accuracy of the model is greatly improved.

We approached the problem purely as a univariate time series forecasting; however, there is scope for taking a multi-variate approach in future study. Future work could consider empirical study of factors that affect the stock price for building model with multivariate forecasting approach during COVID-19. Some of the factors that greatly affect the stock market are the level of infections a country or region [98] has and the level of lock downs [99], which need to be incorporated in a comprehensive forecasting model. Furthermore, other models such as deep learning with LSTM network models [100] and convolutional neural networks [101] could further improve the prediction performance. Furthermore, there is also scope for Bayesian graph convolutional neural networks to capture dependencies amongst related stocks for prediction [102, 103]

6 Conclusions

We applied novel methods in Bayesian neural networks for multi-step-ahead stock price forecasting before and during first phase of COVID-19. The Bayesian neural network used state-of-art sampling strategy that incorporated parallel computing, Langevin gradients and parallel tempering MCMC for improving sampling which provided very promising results when compared to novel neural networks methods. Our investigation revealed that it is important to incorporate data during an extreme event for better model building. In the experiments, the data from initial phase of COVID-19 in the training dataset improved the prediction accuracy is significantly. Hence, high volatility in the stock blackprice makes forecasting very challenging and increases model uncertainty. Although machine learning methods provide accurate prediction, their applicability in stock price prediction remains due to volatility and hence model validity is important. With robust uncertainty quantification via Bayesian inference, investors would find more confidence in predictions using Bayesian neural networks.

The methodology provided better prediction performance prior to the COVID-19 pandemic which is not surprising given the market crash. The results show that Bayesian neural networks provide reasonable predictions with robust uncertainty quantification despite high market volatility during the first phase of the COVID-19 pandemic. This paper provides motivation for multivariate forecasting approaching using Bayesian deep learning methods which could improve the results further.

7 Data and software

We provide open-source Python code and data for further experiments and extensions: https://github.com/sydney-machine-learning/Bayesianneuralnet_stockmarket.

Data Availability

We note that data and code used in the paper is open and processed data has been provided via Github repo with link given in the paper: https://github.com/sydney-machine-learning/Bayesianneuralnet_stockmarket.

Funding Statement

The authors received no specific funding for this work.

References

  • 1.Ding X, Zhang Y, Liu T, Duan J. Deep learning for event-driven stock prediction. In: Twenty-fourth international joint conference on artificial intelligence; 2015.
  • 2. Devadoss AV, Ligori TAA. Stock prediction using artificial neural networks. International Journal of Web Technology. 2013;2(2):42–48. [Google Scholar]
  • 3.Zhang L, Aggarwal C, Qi GJ. Stock price prediction via discovering multi-frequency trading patterns. In: Proceedings of the 23rd ACM SIGKDD international conference on knowledge discovery and data mining; 2017. p. 2141–2149.
  • 4. Simon S, Raoot AD, Lake V. Accuracy driven artificial neural networks in stock market prediction. International Journal of Soft Computing. 2012;3:35–44. doi: 10.5121/ijsc.2012.3203 [DOI] [Google Scholar]
  • 5. Refenes A, Zapranis A, Francis G. Stock Performance Modeling Using Neural Networks: A Comparative Study With Regression Models. Neural Networks. 1994;7:375–388. doi: 10.1016/0893-6080(94)90030-2 [DOI] [Google Scholar]
  • 6. Yoon Y, S G Jr, Margavio TM. A Comparison of Discriminant Analysis versus Artificial Neural Networks. Journal of the Operational Research Society. 1993;44(1):51–60. doi: 10.1057/jors.1993.6 [DOI] [Google Scholar]
  • 7. MacKay DJ. Hyperparameters: Optimize, or integrate out? In: Maximum entropy and Bayesian methods. Springer; 1996. p. 43–59. [Google Scholar]
  • 8. Freedman DA. On the Asymptotic Behavior of Bayes’ Estimates in the Discrete Case. The Annals of Mathematical Statistics. 1963;34(4):1386–1403. doi: 10.1214/aoms/1177703871 [DOI] [Google Scholar]
  • 9. Neal RM. Bayesian Learning for Neural Networks. vol. 118. Springer; New York; 1996. [Google Scholar]
  • 10. MacKay DJC. A Practical Bayesian Framework for Backpropagation Networks. Neural Comput. 1992;4(3):448–472. doi: 10.1162/neco.1992.4.3.448 [DOI] [Google Scholar]
  • 11.Hinton GE, van Camp D. Keeping the Neural Networks Simple by Minimizing the Description Length of the Weights. In: Proceedings of the Sixth Annual Conference on Computational Learning Theory. Association for Computing Machinery; 1993. p. 5–13.
  • 12. Neal RM. Bayesian Learning via Stochastic Dynamics. In: Advances in Neural Information Processing Systems 5. Morgan-Kaufmann; 1993. p. 475–482. [Google Scholar]
  • 13. Neal RM, et al. MCMC using Hamiltonian dynamics. Handbook of Markov Chain Monte Carlo. 2011;2(11). [Google Scholar]
  • 14.Welling M, Teh YW. Bayesian Learning via Stochastic Gradient Langevin Dynamics. In: Proceedings of the 28th International Conference on International Conference on Machine Learning. Omnipress; 2011. p. 681–688.
  • 15. Chandra R, Jain K, Deo RV, Cripps S. Langevin-gradient parallel tempering for Bayesian neural learning. Neurocomputing. 2019;359:315–326. doi: 10.1016/j.neucom.2019.05.082 [DOI] [Google Scholar]
  • 16. Gorbalenya AE, Baker SC, Baric RS, de Groot RJ, Drosten C, Gulyaeva AA, et al. The species Severe acute respiratory syndrome-related coronavirus: classifying 2019-nCoV and naming it SARS-CoV-2. Nature Microbiology. 2020;5(4):536. doi: 10.1038/s41564-020-0695-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17. Monteil V, Kwon H, Prado P, Hagelkrüys A, Wimmer RA, Stahl M, et al. Inhibition of SARS-CoV-2 infections in engineered human tissues using clinical-grade soluble human ACE2. Cell. 2020;. doi: 10.1016/j.cell.2020.04.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Organization WH, et al. Coronavirus disease 2019 (COVID-19): situation report, 72; 2020. World Health Organization.
  • 19. Cucinotta D, Vanelli M. WHO declares COVID-19 a pandemic. Acta bio-medica: Atenei Parmensis. 2020;91(1):157–160. doi: 10.23750/abm.v91i1.9397 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20. Andersen KG, Rambaut A, Lipkin WI, Holmes EC, Garry RF. The proximal origin of SARS-CoV-2. Nature medicine. 2020;26(4):450–452. doi: 10.1038/s41591-020-0820-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Atkeson A. What will be the economic impact of COVID-19 in the us? rough estimates of disease scenarios. National Bureau of Economic Research; 2020. [Google Scholar]
  • 22.Fernandes N. Economic effects of coronavirus outbreak (COVID-19) on the world economy. Available at SSRN 3557504. 2020.
  • 23. Maliszewska M, Mattoo A, Van Der Mensbrugghe D. The potential impact of COVID-19 on GDP and trade: A preliminary assessment; 2020. [Google Scholar]
  • 24.Hart CE, Hayes DJ, Jacobs KL, Schulz LL, Crespi JM. The Impact of COVID-19 on Iowa’s Corn, Soybean, Ethanol, Pork, and Beef Sectors. Center for Agricultural and Rural Development, Iowa State University CARD Policy Brief. 2020.
  • 25. Siche R. What is the impact of COVID-19 disease on agriculture? Scientia Agropecuaria. 2020;11(1):3–6. doi: 10.17268/sci.agropecu.2020.01.00 [DOI] [Google Scholar]
  • 26. Yang Z, Zeng Z, Wang K, Wong SS, Liang W, Zanin M, et al. Modified SEIR and AI prediction of the epidemics trend of COVID-19 in China under public health interventions. Journal of Thoracic Disease. 2020;12(3):165. doi: 10.21037/jtd.2020.02.64 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27. [da Silva] RG, Ribeiro MHDM, Mariani VC, dos Santos Coelho L. Forecasting Brazilian and American COVID-19 cases based on artificial intelligence coupled with climatic exogenous variables. Chaos, Solitons & Fractals. 2020;139:110027. doi: 10.1016/j.chaos.2020.110027 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28. Saba AI, Elsheikh AH. Forecasting the prevalence of COVID-19 outbreak in Egypt using nonlinear autoregressive artificial neural networks. Process Safety and Environmental Protection. 2020;141:1—8. doi: 10.1016/j.psep.2020.05.029 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29. Yousaf M, Zahir S, Riaz M, Hussain SM, Shah K. Statistical analysis of forecasting COVID-19 for upcoming month in Pakistan. Chaos, Solitons & Fractals. 2020;138:109926. doi: 10.1016/j.chaos.2020.109926 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30. Arias Velásquez RM, Mejía Lara JV. Forecast and evaluation of COVID-19 spreading in USA with reduced-space Gaussian process regression. Chaos, Solitons & Fractals. 2020;136:109924. doi: 10.1016/j.chaos.2020.109924 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31. Chimmula VKR, Zhang L. Time series forecasting of COVID-19 transmission in Canada using LSTM networks. Chaos, Solitons & Fractals. 2020;135:109864. doi: 10.1016/j.chaos.2020.109864 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32. Chakraborty T, Ghosh I. Real-time forecasts and risk assessment of novel coronavirus (COVID-19) cases: A data-driven analysis. Chaos, Solitons & Fractals. 2020;135:109850. doi: 10.1016/j.chaos.2020.109850 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33. Ren H, Zhao L, Zhang A, Song L, Liao Y, Lu W, et al. Early forecasting of the potential risk zones of COVID-19 in China’s megacities. Science of The Total Environment. 2020;729:138995. doi: 10.1016/j.scitotenv.2020.138995 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34. Huang Y. Machine Learning for Stock Prediction Based on Fundamental Analysis; 2019. Master of Engineering Science, The University of Western Ontario. [Google Scholar]
  • 35. Larsen JI. Predicting stock prices using technical analysis and machine learning; 2010. Master of Science in Computer Science, Norwegian Institute for Science and Technology, Norway. [Google Scholar]
  • 36. Kourtis G, Kourtis E, Kourtis M, Curtis P. Fundamental analysis, stock returns and high B/M companies. International Journal of Economics and Business Administration. 2017;5:3–18. doi: 10.35808/ijeba/139 [DOI] [Google Scholar]
  • 37. Nazário RTF, e Silva JL, Sobreiro VA, Kimura H. A literature review of technical analysis on stock markets. The Quarterly Review of Economics and Finance. 2017;66:115–126. doi: 10.1016/j.qref.2017.01.014 [DOI] [Google Scholar]
  • 38. Chavan PS, Patil ST. Parameters for stock market prediction. International Journal of Computer Technology and Applications. 2013;4(2):337. [Google Scholar]
  • 39. Petrusheva N, Jordanoski I. Comparative analysis between the fundamental and technical analysis of stocks. Journal of Process Management New Technologies. 2016;4(2):26–31. doi: 10.5937/JPMNT1602026P [DOI] [Google Scholar]
  • 40. Patel J, Shah S, Thakkar P, Kotecha K. Predicting stock market index using fusion of machine learning techniques. Expert Systems with Applications. 2015;42(4):2162–2172. doi: 10.1016/j.eswa.2014.10.031 [DOI] [Google Scholar]
  • 41. Schumaker RP, Chen H. A discrete stock price prediction engine based on financial news. Computer. 2010;43(1):51–56. doi: 10.1109/MC.2010.2 [DOI] [Google Scholar]
  • 42.Lin Y, Guo H, Hu J. An SVM-based approach for stock market trend prediction. In: The 2013 international joint conference on neural networks (IJCNN). IEEE; 2013. p. 1–7.
  • 43.Devi KN, Bhaskaran VM, Kumar GP. Cuckoo optimized SVM for stock market prediction. In: 2015 International Conference on Innovations in Information, Embedded and Communication Systems (ICIIECS). IEEE; 2015. p. 1–5.
  • 44. Dase R, Pawar D. Application of Artificial Neural Network for stock market predictions: A review of literature. International Journal of Machine Intelligence. 2010;2(2):14–17. [Google Scholar]
  • 45. Liao Z, Wang J. Forecasting model of global stock index by stochastic time effective neural network. Expert Systems with Applications. 2010;37(1):834–841. doi: 10.1016/j.eswa.2009.05.086 [DOI] [Google Scholar]
  • 46. Moghaddam AH, Moghaddam MH, Esfandyari M. Stock market index prediction using artificial neural network. Journal of Economics, Finance and Administrative Science. 2016;21(41):89–93. doi: 10.1016/j.jefas.2016.07.002 [DOI] [Google Scholar]
  • 47. Chopra S, Yadav D, Chopra A. Artificial neural networks based Indian stock market price prediction: before and after demonetization. J Swarm Intel Evol Comput. 2019;8(174):2. [Google Scholar]
  • 48. Strader TJ, Rozycki JJ, Root TH, Huang YHJ. Machine Learning Stock Market Prediction Studies: Review and Research Directions. Journal of International Technology and Information Management. 2020;28(4):63–83. [Google Scholar]
  • 49. Khatibi V, Khatibi E, Rasouli A. A new support vector machine-genetic algorithm (SVM-GA) based method for stock market forecasting. International Journal of Physical Sciences. 2011;6(25):6091–6097. [Google Scholar]
  • 50. Qiu M, Song Y. Predicting the direction of stock market index movement using an optimized artificial neural network model. PloS one. 2016;11(5):e0155133. doi: 10.1371/journal.pone.0155133 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51. Qiu M, Song Y, Akagi F. Application of artificial neural network for the prediction of stock market returns: The case of the Japanese stock market. Chaos, Solitons & Fractals. 2016;85:1–7. doi: 10.1016/j.chaos.2016.01.004 [DOI] [Google Scholar]
  • 52. Guresen E, Kayakutlu G, Daim TU. Using artificial neural network models in stock market index prediction. Expert Systems with Applications. 2011;38(8):10389–10397. doi: 10.1016/j.eswa.2011.02.068 [DOI] [Google Scholar]
  • 53.Rathnayaka RKT, Seneviratna D, Jianguo W, Arumawadu HI. A hybrid statistical approach for stock market forecasting based on artificial neural network and ARIMA time series models. In: 2015 International Conference on Behavioral, Economic and Socio-cultural Computing (BESC). IEEE; 2015. p. 54–60.
  • 54. Zhong X, Enke D. Forecasting daily stock market return using dimensionality reduction. Expert Systems with Applications. 2017;67:126–139. doi: 10.1016/j.eswa.2016.09.027 [DOI] [Google Scholar]
  • 55. Deng L, Yu D. Deep learning: methods and applications. Foundations and trends in signal processing. 2014;7(3–4):197–387. doi: 10.1561/2000000039 [DOI] [Google Scholar]
  • 56. LeCun Y, Bengio Y, Hinton G. Deep learning. nature. 2015;521(7553):436. doi: 10.1038/nature14539 [DOI] [PubMed] [Google Scholar]
  • 57. Schmidhuber J. Deep learning in neural networks: An overview. Neural networks. 2015;61:85–117. doi: 10.1016/j.neunet.2014.09.003 [DOI] [PubMed] [Google Scholar]
  • 58. Tang D, Qin B, Liu T. Deep Learning for Sentiment Analysis: Successful Approaches and Future Challenges. Wiley Int Rev Data Min and Knowl Disc. 2015;5(6):292–303. doi: 10.1002/widm.1171 [DOI] [Google Scholar]
  • 59. Shen D, Wu G, Suk HI. Deep Learning in Medical Image Analysis. Annual Review of Biomedical Engineering. 2017;19(1):221–248. doi: 10.1146/annurev-bioeng-071516-044442 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60. Li H. Deep learning for natural language processing: advantages and challenges. National Science Review. 2017;5(1):24–26. doi: 10.1093/nsr/nwx110 [DOI] [Google Scholar]
  • 61. Najafabadi MM, Villanustre F, Khoshgoftaar TM, Seliya N, Wald R, Muharemagic E. Deep learning applications and challenges in big data analytics. Journal of Big Data. 2015;2(1):1–21. doi: 10.1186/s40537-014-0007-7 [DOI] [Google Scholar]
  • 62.Amarasinghe K, Marino DL, Manic M. Deep neural networks for energy load forecasting. In: 2017 IEEE 26th International Symposium on Industrial Electronics (ISIE); 2017. p. 1483–1488.
  • 63. Huang CJ, Kuo PH. A Deep CNN-LSTM Model for Particulate Matter (PM2.5) Forecasting in Smart Cities. Sensors (Basel, Switzerland). 2018;18(7). doi: 10.3390/s18072220 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64.Sudriani Y, Ridwansyah I, Rustini HA. Long short term memory (LSTM) recurrent neural network (RNN) for discharge level prediction and forecast in Cimandiri river, Indonesia. In: IOP Conference Series: Earth and Environmental Science. vol. 299. IOP Publishing; 2019. p. 012037.
  • 65.Ding X, Zhang Y, Liu T, Duan J. Deep Learning for Event-Driven Stock Prediction. In: Proceedings of the 24th International Conference on Artificial Intelligence. AAAI Press; 2015. p. 2327–2333.
  • 66.Nelson DM, Pereira AC, de Oliveira RA. Stock market’s price movement prediction with LSTM neural networks. In: 2017 International joint conference on neural networks (IJCNN). IEEE; 2017. p. 1419–1426.
  • 67. Chandra R, Chand S. Evaluation of co-evolutionary neural network architectures for time series prediction with mobile application in finance. Applied Soft Computing. 2016;49:462–473. doi: 10.1016/j.asoc.2016.08.029 [DOI] [Google Scholar]
  • 68. Mirikitani DT, Nikolaev N. Recursive Bayesian Recurrent Neural Networks for Time-Series Modeling. Neural Networks, IEEE Transactions on. 2010;21(2):262–274. doi: 10.1109/TNN.2009.2036174 [DOI] [PubMed] [Google Scholar]
  • 69. Kocadagli O, Asikgil B. Nonlinear time series forecasting with Bayesian neural networks. Expert Syst Appl. 2014;41:6596–6610. doi: 10.1016/j.eswa.2014.04.035 [DOI] [Google Scholar]
  • 70. F L. Bayesian neural networks for nonlinear time series forecasting. Neurocomputing. 2019;359:315–326. [Google Scholar]
  • 71. Lau KT, Guo W, Kiernan B, Slater C, Diamond D. Non-linear carbon dioxide determination using infrared gas sensors and neural networks with Bayesian regularization. Sensors and Actuators B: Chemical. 2009;136(1):242—247. doi: 10.1016/j.snb.2008.11.030 [DOI] [Google Scholar]
  • 72. Blonbou R. Very short-term wind power forecasting with neural networks and adaptive Bayesian learning. Renewable Energy. 2011;36(3):1118—1124. doi: 10.1016/j.renene.2010.08.026 [DOI] [Google Scholar]
  • 73. Li G, Shi J, Zhou J. Bayesian adaptive combination of short-term wind speed forecasts from neural network models. Renewable Energy. 2011;36(1):352–359. doi: 10.1016/j.renene.2010.06.049 [DOI] [Google Scholar]
  • 74. Ahmar AS, [del Val] EB. SutteARIMA: Short-term forecasting method, a case: COVID-19 and stock market in Spain. Science of The Total Environment. 2020;729:138883. doi: 10.1016/j.scitotenv.2020.138883 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 75. Ali M, Alam N, Rizvi SAR. Coronavirus (COVID-19)–An epidemic or pandemic for financial markets. Journal of Behavioral and Experimental Finance. 2020; p. 100341. doi: 10.1016/j.jbef.2020.100341 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 76. Salisu AA, Vo XV. Predicting stock returns in the presence of COVID-19 pandemic: The role of health news. International Review of Financial Analysis. 2020; p. 101546. doi: 10.1016/j.irfa.2020.101546 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 77.James L, Mark P, Thorpe J. The possible economic consequences of a novel coronavirus (COVID-19) pandemic; 2020. PwC.
  • 78. Nicola M, Alsafi Z, Sohrabi C, Kerwan A, Al-Jabir A, Iosifidis C, et al. The socio-economic implications of the coronavirus and COVID-19 pandemic: a review. International Journal of Surgery. 2020;. doi: 10.1016/j.ijsu.2020.04.018 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 79. McKibbin WJ, Fernando R. The global macroeconomic impacts of COVID-19: Seven scenarios. Asian Economic Papers. 2021;20:1–30. doi: 10.1162/asep_a_00796 [DOI] [Google Scholar]
  • 80. Guan D, Wang D, Hallegatte S, Davis SJ, Huo J, Li S, et al. Global supply-chain effects of COVID-19 control measures. Nature Human Behaviour. 2020; p. 1–11. [DOI] [PubMed] [Google Scholar]
  • 81. Deyle ER, Sugihara G. Generalized theorems for nonlinear state space reconstruction. Plos one. 2011;6(3):e18295. doi: 10.1371/journal.pone.0018295 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 82. Takens F. Detecting strange attractors in turbulence. In: Dynamical Systems and Turbulence, Warwick 1980. Lecture Notes in Mathematics; 1981. p. 366–381. [Google Scholar]
  • 83. Bottou L, Bousquet O. The tradeoffs of large scale learning. In: Advances in neural information processing systems; 2008. p. 161–168. [Google Scholar]
  • 84.Kingma DP, Ba J. Adam: A method for stochastic optimization. arXiv preprint arXiv:14126980. 2014.
  • 85. Carlin BP, Louis TA. Bayes and empirical Bayes methods for data analysis. Statistics and Computing. 1997;7(2):153–154. doi: 10.1023/A:1018577817064 [DOI] [Google Scholar]
  • 86. Liang F. Bayesian neural networks for nonlinear time series forecasting. Statistics and computing. 2005;15(1):13–29. doi: 10.1007/s11222-005-4786-8 [DOI] [Google Scholar]
  • 87. Earl DJ, Deem MW. Parallel tempering: Theory, applications, and new perspectives. Physical Chemistry Chemical Physics. 2005;7(23):3910–3916. doi: 10.1039/b509983h [DOI] [PubMed] [Google Scholar]
  • 88. Neal RM. Sampling from multimodal distributions using tempered transitions. Statistics and computing. 1996;6(4):353–366. doi: 10.1007/BF00143556 [DOI] [Google Scholar]
  • 89. Chandra R, Jain K, Deo RV, Cripps S. Langevin-gradient parallel tempering for Bayesian neural learning. Neurocomputing. 2019. doi: 10.1016/j.neucom.2019.05.082 [DOI] [Google Scholar]
  • 90. Chandra R, Zhang M. Cooperative coevolution of Elman recurrent neural networks for chaotic time series prediction. Neurocomputing. 2012;186:116—123. doi: 10.1016/j.neucom.2012.01.014 [DOI] [PubMed] [Google Scholar]
  • 91. Steiner B. Mastering financial calculations: a step-by-step guide to the mathematics of financial market instruments. Pearson Education; 2007. [Google Scholar]
  • 92. Chandra R, Ong YS, Goh CK. Co-evolutionary multi-task learning with predictive recurrence for multi-step chaotic time series prediction. Neurocomputing. 2017;243:21–34. doi: 10.1016/j.neucom.2017.02.065 [DOI] [Google Scholar]
  • 93. Sikorska A, Scheidegger A, Banasik K, Rieckermann J. Bayesian uncertainty assessment of flood predictions in ungauged urban basins for conceptual rainfall-runoff models. Hydrology and Earth System Sciences. 2012;16(4):1221–1236. doi: 10.5194/hess-16-1221-2012 [DOI] [Google Scholar]
  • 94.Bhat AA, Kamath SS. Automated stock price prediction and trading framework for Nifty intraday trading. In: 2013 Fourth International Conference on Computing, Communications and Networking Technologies (ICCCNT). IEEE; 2013. p. 1–6.
  • 95. Meng TL, Khushi M. Reinforcement learning in financial markets. Data. 2019;4(3):110. doi: 10.3390/data4030110 [DOI] [Google Scholar]
  • 96. Albuquerque R, Koskinen Y, Yang S, Zhang C. Resiliency of environmental and social stocks: An analysis of the exogenous COVID-19 market crash. The Review of Corporate Finance Studies. 2020;9(3):593–621. doi: 10.1093/rcfs/cfaa011 [DOI] [Google Scholar]
  • 97. Bhowmik D. Stock market volatility: An evaluation. International Journal of Scientific and Research Publications. 2013;3(10):1–17. [Google Scholar]
  • 98. Al-Awadhi AM, Al-Saifi K, Al-Awadhi A, Alhamadi S. Death and contagious infectious diseases: Impact of the COVID-19 virus on stock market returns. Journal of Behavioral and Experimental Finance. 2020; p. 100326. doi: 10.1016/j.jbef.2020.100326 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 99. Phan DHB, Narayan PK. Country responses and the reaction of the stock market to COVID-19–A preliminary exposition. Emerging Markets Finance and Trade. 2020;56(10):2138–2150. doi: 10.1080/1540496X.2020.1784719 [DOI] [Google Scholar]
  • 100. Zhao Z, Chen W, Wu X, Chen PC, Liu J. LSTM network: a deep learning approach for short-term traffic forecast. IET Intelligent Transport Systems. 2017;11(2):68–75. doi: 10.1049/iet-its.2016.0208 [DOI] [Google Scholar]
  • 101.Koprinska I, Wu D, Wang Z. Convolutional neural networks for energy time series forecasting. In: 2018 International Joint Conference on Neural Networks (IJCNN). IEEE; 2018. p. 1–8.
  • 102.Chandra R, Bhagat A, Maharana M, Krivitsky PN. Bayesian graph convolutional neural networks via tempered MCMC. arXiv preprint arXiv:210408438. 2021.
  • 103.Matsunaga D, Suzumura T, Takahashi T. Exploring graph neural networks for stock market predictions with rolling window analysis. arXiv preprint arXiv:190910660. 2019.

Decision Letter 0

Junhuan Zhang

30 Apr 2021

PONE-D-21-10570

Bayesian neural networks for stock market forecasting  before and during COVID-19 pandemic

PLOS ONE

Dear Dr. Chandra,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Jun 14 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Junhuan Zhang, PhD

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

  1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please ensure that you refer to Figure 5-8 and 10-13 in your text as, if accepted, production will need this reference to link the reader to the figure.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: No

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: No

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The abstract should be reformulated. This is a very important part of the article. The authors spend too much writing on background. Instead, the authors should better present their research, the main contributions and results, and the conclusions that might be drawn from these results.

In my opinion, the introduction is not well focused. Background should be brief. The originality (novelty) and relevancy of the study should be established with better efforts. My understanding is that this study is an empirical one. Thus, it is expected that the introduction should also include hypothesis and objectives of the study, followed by a justification of the methodology. The current manuscript needs much improvements on all these aspects.

In the introduction, authors claim to study stock market prediction using Bayesian neural networks. However, from the data description, I see the study actually only considers 4 stocks, each from a different country though. Predicting the stock price of a particular company is a very different task from stock market prediction. I believe the authors need more careful wording, and should articulate their research question in the introduction.

More on the data, I don't see any explanation as to why these companies are selected for experiments. Are there any criteria? Also, how to decide the date after which the stock price is affected by COVID-19. Moreover, the authors choose different dates for different stocks/countries. Reasons and justifications are expected here.

The abbreviation FNN-Adam and FNN-SGD are used without mentioning full names. Please check if the use of other abbreviations is of the same problem.

I think one of the main tasks of this study is to compare Bayesian neural network with other neural network methods in terms of forecasting performance. Is it sound to consider only one performance measure (RMSE) here?

From my perspective, the conclusion is quite weak. A more detailed conclusion is needed. The novel method applied here does not seem to outperform state-of-art machine learning methods, at least I can't see it from the conclusion part. The better performance prior COVID-19 is no surprise indeed, and thus does not add any weight to the conclusion part. I expect to see more intelligent conclusions such as the real advantages of this novel Bayesian neural network method. Probably, a comparison with other state-of-art studies would be helpful. This can also be added to the discussion part.

English writing must be carefully revised. Usually, use of WE/OUR in the academic writing should be avoided. I have encountered many grammar mistakes and typos while reading. I list them as below but there are probably more in the manuscript. Thought I am not a native speaker, I feel like the manuscript would benefit from a proof read by a native speaker.

Mistakes I have spotted:

Line 16, “Markov Chain Monte Carlo (MCMC) methods provides a means…”

Line 17, “As the size of model and data continues increases…”

Line 205-206, “The probabilistic neural network model employs the posterior distribution to provides uncertainty quantification on the predictions.”

Line 293, “…we set the burn-in rate is 0.5”

Line 376, “We good prediction accuracy is needed not just for the day ahead, …”

Line 430, “COVID-19 which is not surprising given internal market-crush”

Line 431, “…its is more challenging to provide forecasting during COVID-19…”

Reviewer #2: This paper applies a Bayesian neural networks for multi-step-ahead stock market forecasting before and during COVID-19. But in this version,I don't think the author has made it clear where their novelty ies, is it the novelty of method, or is it the novelty of predicting the stock price changes before and after covid-19? In the paper, the author mentioned that “In the literature, there has not much been done using Bayesian neural networks for stock markets that features robust uncertainty quantification. We can use them to harness power of neural networks that provides good prediction accuracy and also quantify uncertainty. Moreover, there has not been much work that shows how robust machine learning models such as neural networks perform post COVID-19 given major changes in the international stock market with disruptions in international trade and prediction.”,but actually in the listed or not listed references, there are some papers for forecasting in COVID-19. The authors should state exactly the difference of this paper from others, not in a general way. Besides, I list other questions for revision reference.

1. Why Bayes neural network is suitable(even superior)for stock price forecasting. The highlights of this paper should be further addressed.

2. The authors choose 4 stock prices from 4 countries. But I think the selected stocks are not the most representative market stock. Why these datasets are selected should be further clarified.

3. Two other methods are compared with the Bayes neural network, that is, FNN-Adam and FNN-SGD, whose full names should be given when they first appear.

4. The abstract is poorly written. The authors do not clearly show the highlights and significance of this work.

5. The parameter settings in the experiment is very important. I think the authors should give an illustration of parameters in different forecasting models.

6. “”In Setup-2, we include parts of the data during COVID-19 in the training set with all the training data from Setup-1.” What does ”parts” mean here? I suggest the authors to give an exact time period and data length to show them. Besides, what is the objective of Setup 2.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2021 Jul 1;16(7):e0253217. doi: 10.1371/journal.pone.0253217.r002

Author response to Decision Letter 0


6 May 2021

The authors thank reviewers for valuable comments. Pls find the response to review comments attached with the manuscript.

Attachment

Submitted filename: ReviewPLOSoneBayes_stockmarket.pdf

Decision Letter 1

Junhuan Zhang

25 May 2021

PONE-D-21-10570R1

Bayesian neural networks for stock market forecasting  before and during COVID-19 pandemic

PLOS ONE

Dear Dr. Chandra,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Jul 09 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Junhuan Zhang, PhD

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: N/A

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The revised manuscript has addressed all my comments. I only have one additional suggestion. In the study, the authors use the novel method to predict stock prices of 4 individual companies, rather than market indices. I think it is more appropriate to describe it as "stock price forecasting" than "stock market forecasting". Thus, I suggest to revise the title and relevant parts in the main text.

Reviewer #2: The authors have carefully revised paper accorrdingt to my comments and other reviewr's comment. I suggest an

acceptance.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Decision Letter 2

Junhuan Zhang

31 May 2021

Bayesian neural networks for stock price forecasting  before and during COVID-19 pandemic

PONE-D-21-10570R2

Dear Dr. Chandra,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Junhuan Zhang, PhD

Academic Editor

PLOS ONE

Acceptance letter

Junhuan Zhang

17 Jun 2021

PONE-D-21-10570R2

Bayesian neural networks for stock price forecasting before and during COVID-19 pandemic

Dear Dr. Chandra:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Junhuan Zhang

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: ReviewPLOSoneBayes_stockmarket.pdf

    Attachment

    Submitted filename: finalreview.pdf

    Data Availability Statement

    We note that data and code used in the paper is open and processed data has been provided via Github repo with link given in the paper: https://github.com/sydney-machine-learning/Bayesianneuralnet_stockmarket.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES