Skip to main content
PLOS One logoLink to PLOS One
. 2024 Mar 7;19(3):e0298426. doi: 10.1371/journal.pone.0298426

Gold price prediction by a CNN-Bi-LSTM model along with automatic parameter tuning

Amirhossein Amini 1, Robab Kalantari 2,*
Editor: Jayesh Soni3
PMCID: PMC10919698  PMID: 38452043

Abstract

Banking and stock markets consider gold to be an important component of their economic and financial status. There are various factors that influence the gold price trend and its fluctuations. Accurate and reliable prediction of the gold price is an essential part of financial and portfolio management. Moreover, it could provide insights about potential buy and sell points in order to prevent financial damages and reduce the risk of investment. In this paper, different architectures of deep neural network (DNN) have been proposed based on long short-term memory (LSTM) and convolutional-based neural networks (CNN) as a hybrid model, along with automatic parameter tuning to increase the accuracy, coefficient of determination, of the forecasting results. An illustrative dataset from the closing gold prices for 44 years, from 1978 to 2021, is provided to demonstrate the effectiveness and feasibility of this method. The grid search technique finds the optimal set of DNNs’ parameters. Furthermore, to assess the efficiency of DNN models, three statistical indices of RMSE, RMAE, and coefficient of determination (R2), were calculated for the test set. Results indicate that the proposed hybrid model (CNN-Bi-LSTM) outperforms other models in total bias, capturing extreme values and obtaining promising results. In this model, CNN is used to extract features of input dataset. Furthermore, Bi-LSTM uses CNN’s outputs to predict the daily closing gold price.

1. Introduction

Gold is one of the most critical minerals in economics and politics because central banks of countries hold gold reserves as a guarantee to pay for trade on the world market. Moreover, gold is the most popular choice for investments among all commodities [1]. However, there are significant volatility and strong gold price fluctuations due to various reasons such as supply and demand, political issues, and inflation. Developing an accurate closing gold price prediction is not a simple task due to its nonlinear nature, the large number of unpredictable factors involved, and the significant volatilities of the gold price, resulting in complicated temporal time series [2]. Literature reviews show that several traditional methods have been developed for gold price forecasting, such as linear regression, support vector regression (SVR), autoregressive moving averages (ARMA), autoregressive integrated moving averages (ARIMA), and artificial neural networks (ANNs) [36]. However, statistical models are usually based on stationarity and linear correlation assumptions. In addition, machine learning models appear to be unable to detect and capture the nonlinear and complex behavior of the gold price time sequence. Therefore, none of these methods is able to provide a robust and reliable forecasting model [7].

DNNs address ANNs’ shortcomings, such as gradient vanishing. In addition, the automatic learning capabilities of DNNs enable them to find relationships between input and output variables of raw data on a complex and high-level basis [8]. In recent years, the development of DNNs for the prediction of the gold price has gained popularity among the scientific community [9]. There has also been widespread use of CNNs and recurrent neural networks (RNNs) to model time series [10]. Using these models, it is possible to enhance the accuracy of time sequence prediction by addressing the problem of long-term dependencies in the input data using the LSTM architecture design and the CNN models by eliminating the noise from the input data and extracting more useful features for the prediction [11]. Although many studies have been conducted on gold price forecasting using DNNs, the automatic parameter tuning (grid search method) of DNNs considering sensitivity analysis for closing gold price has not been collectively studied. The main objective of this paper is to propose a tuned hybrid deep neural network for the adaptive prediction of closing gold prices. Specifically, the model predicts the gold price for the next time step based on the analysis of historical gold prices. The proposed model was assessed against state-of-the-art DNNs such as CNN, CNN-LSTM, Conv-LSTM, and Stacked LSTM for the adaptive prediction of closing gold price. Using the one parameter at a time method for sensitivity analysis, we also analyze the effect of look back, learning rate, and mini-batch size parameters.

The rest of this paper is organized as follows: “section 2: Related work” summarizes recent studies on the application of various models and parameter tuning for gold, stock, and Bitcoin price prediction. “Section 3: Methods” presents a general description of DNNs, grid search technique, statistical criteria, and the dataset utilized in the study. “Section 4: Proposed CNN-Bi-LSTM model” provides a detailed description of the proposed deep learning model. In “Section 5: Results and discussion”, the modeling results and sensitivity analysis are presented. “Section 6: Conclusion” summarizes the research findings, and some outlines are provided for future research.

2. Related work

The scientific community has grown increasingly interested in using DNNs and other models to predict gold’s price movement over the last decade. The introduced methodologies provided some useful insight into the closing gold price behavior. Pindoriya et al. [12] used an adaptive wavelet neural network (AWNN) to predict energy price in electricity markets. Results showed that AWNN predicted the sequence more accurately than other methods such as wavelet-ARIMA, multi-layer perceptron (MLP), and radial basis function (RBF). Vidya & Hari [13] predicted gold price time series, 10850 data from 1979 to 2020, using LSTM and CNN deep neural networks. They utilized an adaptive moment estimation (Adam) optimizer, which could efficiently deal with nonlinear problems such as gold price prediction. Researchers reported that LSTM forecasts were superior to common models, such as ARIMA, SVR, and CNN. Yadav et al. [14] predicted the real-time stock price using the best features of Fast Recurrent Neural Networks (FastRNN), CNN, and Bi-LSTM algorithms. The model predicted the stock price for daily time steps using four companies’ minute time step data. The proposed model outperformed ARIMA and other real-time hybrid DNNs. Li et al. [15] proposed an LSTM model to forecast the daily stock price in China using limit order books (LOBs) data. Based on experimental results, they highlighted the effectiveness of the proposed model. Madziwa et al. [16] predicted the annual gold price using autoregressive distribution lag (ARDL), stochastic mean reverting, and ARIMA models. Results demonstrated that ARDL model performed better compared to other models.

An important point to note is that many of the datasets used in these studies include recent COVID-19 pandemic values at the beginning of 2020, accompanied by substantial volatility and departures from usual behavior. Mohtasham Khani et al. [17] conducted a gold price prediction on daily time steps during the Coronavirus pandemic using DNNs, namely LSTM, CNN, Bi-LSTM, and CNN-LSTM models. They tried to explore the effect of new daily cases of Coronavirus on financial markets, especially gold price prediction using data collected on the gold price and new daily cases of Coronavirus from 2015 to 2020. Based on the mean squared error (MSE) indicator, the authors’ highlighted the potential of Stacked LSTM compared to CNN-LSTM and Bi-LSTM for adaptive and multi-step gold price forecasting. Hansun & Suryadibrata [18] utilized a simple three-layer LSTM to predict gold price during the COVID-19 crisis using 5000 data. The LSTM parameters, such as the number of hidden units, batch size, and dropout probability are selected using the trial and error approach. They reported that the determination coefficient and RMSE of the LSTM network equal 97.2% and 39.94, respectively.

According to studies, hybrid DNNs are more effective than individual DNNs. For example, He et al. [19] predicted daily gold price based on the integration of the LSTM and CNN models. The results indicated that their model outperforms traditional models such as SVR, ARIMA, CNN, and deep regression. Li & Dai [20] proposed a CNN-LSTM network for predicting closing Bitcoin prices based on the closing Bitcoin price itself, macroeconomic indicators, and investors’ data. The results illustrate that the CNN-LSTM hybrid network can effectively enhance value and direction prediction accuracy compared to backpropagation (BP), CNN, and LSTM. Livieris et al. [9] used a multivariable CNN-LSTM model for the prediction of cryptocurrencies such as Bitcoin (BTC), Ethereum (ETH), and Ripple (XPR) and their movements. The results showed that the proposed hybrid DNN has the ability to efficiently reduce overfitting and computational cost compared to traditional fully connected DNNs. Lu et al. [21] proposed a hybrid DNN (CNN-Bi-LSTM) to predict the daily closing stock price. The optimal parameters and hyperparameter of the CNN-Bi-LSTM model are hand-picked through a trial and error approach. They choose the Adam optimizer with a 0.001 learning rate and loss function equal to MAE. Results demonstrate that the performance of this model is superior to the performance of MLP, LSTM, CNN, Bi-LSTM, and CNN-LSTM with MAE, RMSE, and R2 are 21.95, 31.69, and 0.98 respectively. Kim and Jang [20] used hybrid deep neural networks (CNN-LSTM and CNN-GRU) to predict the price of crude oil. Results show that their proposed models are more accurate (0.985 and 0.988 correlation coefficients) than other DNNs. Zhao and Yang [22] proposed a new hybrid LSTM based model to predict stock price movement prediction. Evaluations demonstrated that the accuracy of proposed model is higher compared to other models. Chen et al. [23] used the combination of the K-means clustering algorithm and LSTM to predict China’s commercial bank stock price. A K-means method was utilized to find other banks with similar price trends, which were then used as input sequences to an LSTM network in order to predict stock prices over the next multiple time intervals. The researchers discovered that their methodology exhibited a superior level of accuracy, as evidenced by lower MSE and MAE values and higher R2, compared to alternative approaches in the prediction of stock prices. Liu et al. [24] utilized a compound of variational mode decomposition and metal-learning algorithms to predict stock price. First, they broke the sequence into smaller sequences using the variational mode decomposition method, then the They first broke down the sequence to subsequences using variational mode decomposition, then used the model-agnostic meta-learning algorithm and LSTM network to forecast the subsequences. Finally, the subsequences predictions are combined to produce the final output sequence. According to the results, this method performed better (higher coefficient of determination and lower RMSE) than other methods for different datasets. Vo et al. [11] predicted the Brent oil price using the Bi-LSTM network. Researchers reported that Bi-LSTM Network outperformed other networks such as LSTM, CNN-LSTM, and CNN-Bi-LSTM. Lu et al. [25] predicted the gold price using a hybrid model called complete ensemble empirical mode decomposition with adaptive noise (CEEMDAN)- grey wolf optimizer (GWO)- SVR and high-frequency intrinsic mode functions subsequence (IMFH), low-frequency intrinsic mode functions subsequence (IMFL), and the residual as input data. The IMFL and the residual have higher correlations compared to IMFH. These factors showed the prominent factors and overall trend of gold price sequence. The final output is gained through combining the aforementioned timeseries. Results showed that the proposed model outperforms other models such as ARIMA, generalized auto-regressive conditional heteroscedasticity (GARCH), ANN, classification and regression tree (CART), and BPNN (back propagation neural network). Zhao and Yang [26] employed emotion-enhanced CNN, denoising autoencoder (DAE), and LSTM models to forecast stock prices. The findings indicate that the combined model exhibited superior accuracy compared to the LSTM and CNN base models. Madhika et al. [27] applied the ARIMA and LSTM models for the purpose of forecasting gold prices. The findings indicate that the LSTM model outperforms the ARIMA model in terms of the RMSE and Mean Absolute Percent Error (MAPE) statistical indicators. Pangestu et al. [28] employed a Long Short-Term Memory (LSTM) model and a Fuzzy system in their study to forecast the price of gold. The results of the study revealed that the RMSE of the LSTM model exhibited superior performance compared to the fuzzy system. Liang et al. [29] used the improved version of CEEMDAN (ICEEMDAN) algorithm at first stage. Next, they forecasted the gold price using the long short-term memory, convolutional neural networks, and convolutional block attention module (LSTM-CNN-CBAM) for different sublayers. They reported the proposed model (i.e., ICEEMDAN-LSTM-CNN-CBAM (ILCC)) exhibited a superior level of accuracy (R-square) compared to other related models. The use of individual and hybrid DNNs for gold price prediction has been extensively studied; however, the comparison between various DNNs, the use of CNN-Bi-LSTM architecture for closing gold price prediction, the automatic parameter tuning using grid search to determine optimal parameters of each DNN, and the sensitivity analysis of three hyperparameters of the best DNN have not been appropriately explored. Our research contribution addresses the aforementioned shortcoming in closing gold price prediction.

3. Methods

In this section, four deep learning networks, grid search hyperparameter tuning technique, statistical evaluation, and data used for closing gold price forecasting are explained in detail.

3.1 Deep learning models

3.1.1 Stacked LSTM

The LSTM network, an artificial RNN, is first introduced by S. Hochreiter and J. Schmidhuber in 1997 [30]. This network is designed to address the problem of long-term dependencies in time sequences and gradient explosion and disappearance in RNNs [31]. This architecture has loops so it can retain information or omit unneeded information. The network adjusts information flow by using gate control mechanisms and systematically determines how much incoming information is retained in each step [32]. The stacked LSTM consists of multiple LSTM layers with specified hidden units stacked on top of each other. Fig 1 illustrates the three parts of the LSTM memory cell: the input gate, the forget gate, and the output gate. Input gates (it) control the flow of new values into the cell and store some information about the current value. The forget gate (gt) determines whether values need to be retained or forgotten. Also, the output gate (ot) select useful information from the current cell state. The outputs of each cell that being transferred to the next cell are hidden state (ht) and cell state (ct) [13].

Fig 1. The schematic figure of the Long-Short Term Memory (LSTM) network.

Fig 1

U and W are weight matrices in the below equations, b is a bias term, (σ) is a sigmoid activation function, and the symbol * represents multiplication. The following equations present the architecture of the LSTM network.

it=σ(Uixt+Wiht-1+bi) (1)
gt=σ(Ugxt+Wght-1+bf) (2)
ot=σ(Uoxt+Woht-1+bo) (3)

Eqs (1)–(3) represent input, forget, and output gate expressions, which take the weighted sum of the hidden state at time t-1 (ht−1), and the input at time t (xt) as inputs of activation functions to determine the output [33]. The Eqs (4)–(6) are as follows for the candidate state (c˜t), cell state (ct), and final output or hidden state at time t (ht).

c˜t=tanh(Ucxt+Wcht-1+bc) (4)
ct=gt*ct-1+it*c˜t (5)
ht=ot*tanh(ct) (6)

3.1.2 CNN

The convolutional neural network (CNN) consists of many layers, each with its distinct architecture, called convolutional layers and pooling layers [34]. It emulates the perception of local information using local information by biological vision cells. The data features are extracted using a local connection and layer-by-layer computation, and then the global information is synthesized with the full connection [20]. In convolutional-based networks such as ConvLSTM and CNN-LSTM, CNN is used to extract features of input time series efficiently and learn the internal representation of sequence [35]. Several parameters are required for CNN architectures, such as the stride type, pooling type, and filter size. The filter size shows the set of inputs to which the kernel will be applied during the convolution process. The stride is the number of steps the filtering process will take. The pooling type indicates which pooling process will be applied to the filtered output [36]. Convolution layers contain a plurality of convolution kernels, and their calculation is shown in Eq (7). The extracted features of the convolutional layer are very high. After the convolution layer, this problem is solved by adding a pooling layer to reduce the feature dimensions and the cost of training the network [37].

lt=tanhxt*kt+bt (7)

In Eq (7), lt is the output of the convolutional layer, xt is the input vector, bt is the bias of the convolution kernel, tangent hyperbolic (tanh) as the activation function, and kt is the weight of the convolution kernel.

3.1.3 CNN-LSTM

In many applications, such as image recognition, signal processing and natural language processing, convolutional neural networks (CNNs) showed good results. In short-term forecasting time sequences, LSTMs performed well. As a result of combined CNN and LSTM layers, the CNN-LSTM method can extract complex features from datasets and identify complex trends. CNN layers consist of an input layer that receives input time series, an output layer that extracts features for LSTMs, and a certain number of hidden layers. There is usually a convolutional layer, ReLU activation function, and a max pooling layer in the hidden layer. In the CNN layer, local features are extracted from high layer inputs and transferred to the lower layers in order to produce more complex features [38].

3.1.4 ConvLSTM

ConvLSTM has been introduced for forecasting temporal sequences, where both input and target are temporal sequences [39]. Additionally, it can capture temporal patterns of large-scale sequential datasets [40]. A convolution operation is applied between raw input data and convolution kernels to produce new features [7]. In addition to its accomplishments in image and speech detection, speech synthesis, and facial recognition, this network has demonstrated outstanding capabilities in exploring time series data [36].

The learning rate is the most important parameter to configure in the tuning process of DNNs. Large learning rates reduce the model’s run time at the cost of converging a sub-optimal solution. However, the small learning rate may converge to the global optimum solution of the model with a significantly longer run time. The grid search approach can help highlight the optimal learning rate and describe the effect of the learning rate on the model’s performance [41]. By combining Adaptive Gradient (AdaGrad) and Root Mean Square Propagation (RMSProp), adaptive moment estimation (Adam) calculates adaptive learning rates for each parameter. Adam’s algorithm maintains an exponentially decaying average of past gradients, similar to momentum, in addition to an exponentially decaying average of past squared gradients, like Adadelta and RMSprop. In practice, Adam works well and compares favorably to other stochastic optimization methods [42]. In all layers except the output layer, where linear activation was used, the Rectified Linear Unit (ReLU) activation function was employed.

Networks combined with dropout layers are less prone to overfitting by randomly removing hidden units from a network with the tuned probability. As a result, useful features for producing correct answers operated [43]. The number of "lookback" is another important hyperparameter of DNNs. Each combination of parameters has a different effect on the performance of the gold price prediction model. Though using conventional hyperparameters and parameters generate good forecasts in DNNs, we apply automatic parameter tuning to find the optimal performance of these models.

3.2 Hyperparameter tuning using the grid search

Many hyperparameters are associated with DNNs, including mini-batch size, learning rate, lookback, etc. Parameters that cannot be updated during the model’s training are known as hyperparameters [44]. This can result in suboptimal results if the hyperparameters and parameters are selected incorrectly. Hyperparameter tuning may also be referred to as the process of choosing the most appropriate values for hyperparameters and parameters. Generally, parameters were selected as rules of thumb in the literature of forecasting papers [45].

For deep learning hyperparameter and parameter tuning, a vast potential space must be explored [46]. For example, the parameters of the ConvLSTM network are the number of hidden units, the filter size, mini-batch size, learning rate, and the dropout probability. In general, applying parameters using the rule of thumb or previous research results in the best results, but we have tried to find the best parameter to achieve optimal performance, using automatic parameter tuning for DNNs [47]. There are different automatic hyperparameter tuning algorithms including search, random search, and Bayesian optimization. In this paper, the grid search algorithm is used to adjust the parameters of the DNN. The grid search explores the best combination of parameters within the specified number of trials.

Grid search is one of the basic methods for parameter optimization. Grid search performs an exhaustive search on the set of parameters specified by the user. The users generate all candidates, so they must have preliminary knowledge about these parameters. Grid search may be applied to several parameters with a limited search space. Grid search is the most straightforward algorithm that will always yield the best results as long as sufficient resources are available [44]. In comparison to other tuning techniques, grid search runs in parallel because each trial runs separately without affecting the time sequence. In other words, results for one trial are not affected by those from other trials.

The grid search approach is usually preferred when users are familiar with these parameters enough to define a narrow search space. Despite the fact that other search algorithms may offer more favorable features, grid search remains the most popular method due to its mathematical simplicity [48].

In this study, the maximum number of trials is set at 50. The following parameters, based on previous studies [18,22], ranges are used:

  • The number of hidden units per layer related to LSTM and Dense layers is assigned to an integer ranging from 10 to 50 with 10 steps

  • The mini-batch size is set to the power of two numbers from 8 to 1028

  • The learning rate of the Adam optimizer is set to choose from one of four choices [0.05, 0.005, 0.0005, 0.00005]

  • The lookback is assigned to a value ranging from 1 to 40 with one step

  • The filter size in convolutional-based networks is set to an integer ranging from 4 to 128 with 12 steps

  • The kernel size in convolutional-based networks is set to choose from one of four choices [1, 2, 3, 4]

  • The dropout probability is assigned to an integer ranging from 0.1 to 0.9 with 0.1 steps

3.3. Statistical evaluation

Three statistical criteria, namely Root Mean Square Error, Root Mean Absolute Error (RMAE), and coefficient of determination (R2), are utilized to evaluate the performance of DNNs (Table 1).

Table 1. Formulation of the selected statistical criteria.

Statistical criteria Formula Description Perfect/no skill
Coefficient of determination (R2) 1-i=1NGfi-Goi2i=1NGfi-Gf¯2 Agreement on forecasted and observed gold prices 1/-Inf
RMAE 1Ni=1N(Gfi-Goi) The magnitude of the difference between forecasted and observed gold prices 0/
RMSE 1Ni=1N(Gfi-Goi)2 The closeness between forecasted and observed gold prices 0/

Note: GfI and Goi are the values of the forecasted and observed dataset for the ith of N closing gold price. Also, Gf¯ is the mean of the observed closed gold price dataset.

Because deep learning networks are stochastic, five realizations were done for each model, and the results were presented as box plots in the result and discussion section.

3.4 Data

Deep neural networks (DNNs) are used to predict closing gold prices based on reference closing gold price data. From 1978 to 2021, the closing gold price from the last 44 years, including the Covid-19 era, is used as an input into models. In standard ANNs, time series are entered into the network in a different way than in RNNs. Standard ANNs are typically trained with random input data. The data in time-dependent networks such as DNNs, however, is presented in occurrence time order to maintain the time-dependency structure. Out-of-sample prediction is the standard method for testing time-dependent networks. Extrapolation beyond the reference dataset constitutes true out-of-sample prediction [49]. For evaluation purposes, the closing gold price data were divided into training, validation, and testing sets (Fig 2). More analytically, the first 70% of the acquired daily data, from 29 December 1978 to 11 September 2008, is used for model training. The following 20% and 10% of the gold price data are selected for model validation (from 12 September 2008 to 8 March 2017) and testing (from 9 March 2017 to 4 June 2021), which ensured a substantial amount of out-of-sample data that had previously not been seen by models.

Fig 2. Daily closing gold price in USD from December 1978 to 4 June 2021 (The black, red, and yellow lines show the training, validation, and test data).

Fig 2

4. Proposed CNN-Bi-LSTM model

By using CNN-Bi-LSTM as a combined model, the CNN and Bi-LSTM models can extract valuable information with the convolutional layer and learn from forward as well as backward time sequences by utilizing LSTM [32]. Using its architecture consisting of forward and backward LSTM layers, Bi-LSTM processes data in both directions—from first to last input and last to the first output and uses both forward and backward information [50,51].

Fig 3 illustrates the flowchart of the proposed methodology for predicting the closing gold price using DNNs and the grid search tuning technique. According to Fig 3, our method is comprised of five steps. The first step is to acquire the daily closing gold price. The reference dataset is then normalized to minimize sequencing depth differences, improve the initialization of parameters, and speed up the convergence of the model. Next, grid search algorithms are used to tune the parameters of adaptive DNN models. Three statistical indices are used to evaluate the performance of the models in the evaluation phase. Finally, the best adaptive model is chosen to be implemented to forecast the closing gold price.

Fig 3. Flowchart of the adaptive forecasting of the closing gold price.

Fig 3

5. Results and discussion

5.1 Model optimization

The first step is to select the DNN architecture that is best suited to the problem at hand [52]. We will apply the grid search technique in this study to optimize the parameters of each DNN, resulting in optimal performance of the models. The parameters’ optimal values are shown in Table 2. According to Fig 4, the CNN-Bi-LSTM network outperforms all other models in the comprehensive comparison between five different state-of-the-art forecasting approaches (Stacked-LSTM, CNN, CNN-LSTM, ConvLSTM, and CNN-Bi-LSTM). The general rule of thumb for manually tuning a model’s structures is to start with a simple model, optimize its parameters and check if it can obtain good metrics on the training data. If the model is unable to obtain satisfactory training metrics, it may be necessary to add more layers (i.e., increase its complexity).

Table 2. The parameters’ optimal value of DNNs (daily).

Stacked-LSTM CNN CNN-LSTM ConvLSTM Proposed CNN-Bi-LSTM
Lag time 10 10 12 12 24
# of sequences - - 2 3 3
# of hidden units 40 20 10 - 20
Filter size - 128 64 128 32
Kernel size - 1 2 (1,2) 1
Mini batch size 512 256 1024 128 30
Learning rate 0.005 0.0005 0.005 0.005 0.0005
Dropout probability - - - 0.2 0.1
RMSE 35.24 35.43 36.85 38.01 34.87
RMAE 4.96 5.19 5.24 6.65 5.15
R2 0.94 0.94 0.94 0.91 0.95

Fig 4. Box plots of DNNs performance.

Fig 4

Each box plot is calculated from five realizations. The average coefficient of determination for the gold price forecasting in the validation set for (a) Stacked-LSTM, (b) CNN, (c) Bi-LSTM, (d) CNN-LSTM, (e) ConvLSTM, and (f) proposed CNN-Bi-LSTM (The green triangle and orange line indicate the mean and median coefficients of determination for each model realization).

In comparison to the trial-and-error approach, tuning time for parameters is significantly reduced. Fig 5 illustrates the reference closing gold price and the predicted price in the test period using five tuned DNNs. As can be perceived from Fig 5, all models could predict the gold price movement well for test data. CNN-Bi-LSTM outperforms other models in total bias and captures extreme values, which are recommended buy and sell points. Furthermore, the proposed CNN-Bi-LSTM and CNN models adapt to trends and peaks faster than other models. In order to make better predictions, the trend of the closing gold price must quickly adapt to new prices. The rest of the models do not have the ability to use new data as much as the CNN-Bi-LSTM model. For instance, the CNN model has a problem in predicting buy points, low extremes, which come after sell points, including high prices in the period.

Fig 5. Observation and forecasted charts for test data of (a) Stacked-LSTM, (b) CNN, (c) Bi-LSTM, (d) CNN-LSTM, (e) ConvLSTM, and (f) proposed CNN-Bi-LSTM.

Fig 5

For the aforementioned learning rate, Fig 6 illustrates the variation of training and cross-validation losses versus epoch for the CNN-Bi-LSTM model. The loss (RMSE) value converges to a global minimum smoothly as it decreases. Furthermore, the difference between the losses of the cross-validation and training datasets is very small, which indicates that this model avoids overfitting. It has been demonstrated that the model converges well and remains stable after eight epochs.

Fig 6. Loss value versus epoch in CNN-Bi-LSTM.

Fig 6

The primary constraint of the aforementioned models is its reliance on a substantial volume of data in order to make accurate predictions.

5.2 sensitivity analysis

In all mathematical modeling, sensitivity analysis should be an integral part of the process. An important benefit of performing a sensitivity analysis is identifying parameters or processes that are sensitive to changes in model output [53]. The sensitivity analysis of DNN modeling, like that of any mathematically based model, helps determine which input parameters have the greatest impact on the model’s performance. The sensitivity analysis of the top three important hyperparameters of the CNN-Bi-LSTM (look back, learning rate, and mini-batch size) is shown in Fig 7.

Fig 7. Box plots of CNN-Bi-LSTM model.

Fig 7

Each box plot is calculated from five realizations. The average Coefficient of determination for the forecasting of the validation data for (a) look back, (b) learning rate, and (c) mini-batch sizes.

5.2.1 Lookback

The lookback for the closing gold price was tested from 1 to 40-time steps. The results (Fig 7a) indicate that the CNN-Bi-LSTM network with a 24 look back has the highest average coefficient of determination, equal to 0.92. The accuracy of models with less look back than the optimal value is lower due to a lack of extreme data in the previous time step. Higher look back yields lower accuracy, which may be caused by the noise created by irrelevant input and the model’s inability to predict extreme values that cause significant errors. In other words, models with more look back adapt more slowly and show a less wavy and smoother curve when extreme prices are involved in training sets. Based on the results, the observation of the last 24-look back is utilized in the rest of the analysis. For the Stacked-LSTM, CNN, CNN-LSTM, ConvLSTM, and proposed CNN-Bi-LSTM models, the gold price reference dataset from the last 1 to 40-time steps were tested, and the 10, 10, 12, 12, and 24-look back were set as optimal look backs in the DNNs.

5.2.2 Learning rate

Four different learning rates, 0.05, 0.005, 0.0005, and 0.00005, were applied for the proposed CNN-Bi-LSTM network. Results (Fig 7b) of five realizations indicate that the model with a 0.0005 learning rate has the highest coefficient of determination, possibly the most converging learning rate compared to other choices. It is noteworthy that the learning rate equal to 0.05 results in negative values in many runs, indicating a non-converging value.

5.2.3 Mini-batch size

In this analysis, the mini-batch size was adjusted from 8 to 128. Results (Fig 7c) illustrated that the average R2 for the 32-mini-batch size fluctuates from 0.91 to 0.94. There is no significant difference in performance when using both 64 and 128 mini-batch sizes. Therefore, there could be various data in each mini-batch of the training dataset, which is necessary to calculate gradient descent in mini-batch sizes. Because there are insufficient extreme data in each mini-batch, the results lead to an unstable output.

TensorFlow was used as the backend to train all DNNs using a Python package called Keras [54]. For running the models, a computer system with GeForce GTX 1060 GPU and Intel(R) Core (TM) i7-7700HQ CPU is used.

6. Conclusion

This paper utilizes parameter tuning prior to the DNNs learning process for closing gold price forecasting on a daily time step. The closing gold price for 44 years, from 1978 to 2021, is considered the input of models. The grid search tuning approach is applied to obtain the optimal performance of each DNN model. The forecasted closing gold price of LSTM, CNN, CNN-LSTM, ConvLSTM, and CNN-Bi-LSTM models are compared. The proposed CNN-Bi-LSTM model outperforms other models in general and can capture extreme values, which can be used as indications of where to buy or sell. Moreover, the CNN-Bi-LSTM model adapts to new closing gold price data more quickly than other models.

The performance of the proposed model depends on a number of characteristics. First, Due to a lack of extreme data in the previous time step, models with less look back than optimal value have lower accuracy. Lookback results in lower accuracy, which may be due to noise created by irrelevant inputs and the model’s inability to predict extreme values. Second, according to the tuning procedure for the proposed CNN-Bi-LSTM model, the 32-mini batch size is considered a reliable value to reduce bias. Due to fluctuation in extreme data within each mini-batch, small mini-batch sizes result in unstable outputs. Compared to other models, the proposed CNN-Bi-LSTM performs better in total bias and capture of extreme values with coefficients of determination, RMSE, and RMAE equal to 0.95, 37.94, and 5.27, respectively, for the test data. Compared to other existing research, the following features distinguish the proposed framework:

  • In order to predict closing gold prices, a novel model is implemented by combining CNN and LSTM features.

  • In order to determine the optimal set of parameters for the CNN-Bi-LSTM model, the grid search tuning technique is used

  • This paper performs a sensitivity analysis of three important parameters so as to evaluate the grid search tuning technique and identify parameters that are sensitive to changes in model output.

The closing gold price dataset is the only input sequence used by the suggested model. The model’s performance may be enhanced by incorporating additional variables, such as the price of silver, crude oil, and the US dollar index, or by switching from grid search to other automatic hyperparameter tuning techniques like random search or Bayesian methods. Furthermore, the accuracy of the models may be improved by applying methods like wavelet transform, which reduce noise and use denoised data as input for deep learning models.

Supporting information

S1 Dataset

(ZIP)

pone.0298426.s001.zip (185.9KB, zip)

Data Availability

The gold price dataset is available at https://github.com/amirhosseinse7/Closing-gold-price.

Funding Statement

The authors received no specific funding for this work.

References

  • 1.Makala D, Li Z (2021) Prediction of gold price with ARIMA and SVM. J Phys Conf Ser 1767:. doi: 10.1088/1742-6596/1767/1/012022 [DOI] [Google Scholar]
  • 2.Yurtesver M (2021) Gold Price Forecasting Using LSTM, Bi-LSTM and GRU. Eur J Sci Technol. doi: 10.31590/ejosat.959405 [DOI] [Google Scholar]
  • 3.Dubey AD (2016) Gold price prediction using support vector regression and ANFIS models. 2016 Int Conf Comput Commun Informatics, ICCCI 2016.
  • 4.Guha Banhi. Bandyopadhyay G (2016) Gold Price Forecasting Using ARIMA Model. J Adv Manag Sci 4:117–121. doi: 10.12720/joams.4.2.117-121 [DOI] [Google Scholar]
  • 5.Sadorsky P (2021) Predicting Gold and Silver Price Direction Using Tree-Based Classifiers. J Risk Financ Manag 14:198. doi: 10.3390/jrfm14050198 [DOI] [Google Scholar]
  • 6.Yang X (2019) The Prediction of Gold Price Using ARIMA Model. 196:273–276. doi: 10.2991/ssphe-18.2019.66 [DOI] [Google Scholar]
  • 7.Livieris IE, Pintelas E, Pintelas P (2020) A CNN–LSTM model for gold price time-series forecasting. Neural Comput Appl 32:17351–17360. doi: 10.1007/s00521-020-04867-x [DOI] [Google Scholar]
  • 8.Barzegar R, Aalami MT, Adamowski J (2021) Coupling a hybrid CNN-LSTM deep learning model with a Boundary Corrected Maximal Overlap Discrete Wavelet Transform for multiscale Lake water level forecasting. J Hydrol 598:126196. doi: 10.1016/j.jhydrol.2021.126196 [DOI] [Google Scholar]
  • 9.Livieris IE, Pintelas E, Pintelas P (2020) A CNN–LSTM model for gold price time-series forecasting. Neural Comput Appl 32:17351–17360. doi: 10.1007/s00521-020-04867-x [DOI] [Google Scholar]
  • 10.Zheng J, Fu X, Zhang G (2019) Research on exchange rate forecasting based on deep belief network. Neural Comput Appl 31:573–582. doi: 10.1007/s00521-017-3039-z [DOI] [Google Scholar]
  • 11.Vo AH, Nguyen T, Le T (2020) Brent oil price prediction using bi-lstm network. Intell Autom Soft Comput 26:1307–1317. doi: 10.32604/iasc.2020.013189 [DOI] [Google Scholar]
  • 12.Pindoriya NM, Singh SN, Singh SK (2008) An adaptive wavelet neural network-based energy price forecasting in electricity markets. IEEE Trans Power Syst 23:1423–1432. doi: 10.1109/TPWRS.2008.922251 [DOI] [Google Scholar]
  • 13.Vidya GS, Hari VS (2020) Gold Price Prediction and Modelling using Deep Learning Techniques. 2020 IEEE Recent Adv Intell Comput Syst RAICS 2020 28–31. doi: 10.1109/RAICS51191.2020.9332471 [DOI] [Google Scholar]
  • 14.Yadav K, Yadav M, Saini S (2021) Stock Market Predictions Using FastRNN, CNN, and Bi-LSTM-Based Hybrid Model. Lect Notes Electr Eng 796:1–10. doi: 10.1007/978-981-16-5078-9_1 [DOI] [Google Scholar]
  • 15.Li Y, Li L, Zhao X, et al. (2020) An Attention-Based LSTM Model for Stock Price Trend Prediction Using Limit Order Books. J Phys Conf Ser 1575:. doi: 10.1088/1742-6596/1575/1/012124 [DOI] [Google Scholar]
  • 16.Madziwa L, Pillalamarry M, Chatterjee S (2022) Gold price forecasting using multivariate stochastic model. Resour Policy 76:102544. doi: 10.1016/j.resourpol.2021.102544 [DOI] [Google Scholar]
  • 17.Mohtasham Khani M, Vahidnia S, Abbasi A (2021) A Deep Learning-Based Method for Forecasting Gold Price with Respect to Pandemics. SN Comput Sci 2:1–12. doi: 10.1007/s42979-021-00724-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Hansun S, Suryadibrata A (2021) Gold price prediction in covid-19 era. Int J Comput Intell Control 13:29–34. [Google Scholar]
  • 19.He Z, Zhou J, Dai HN, Wang H (2019) Gold price forecast based on LSTM-CNN model. IEEE 17th Int Conf Dependable, Auton Secur Comput IEEE 17th Int Conf Pervasive Intell Comput IEEE 5th Int Conf Cloud Big Data Comput 4th Cyber Sci 1046–1053. 10.1109/DASC/PiCom/CBDCom/CyberSciTech.2019.00188. [DOI]
  • 20.Li Y, Dai W (2020) Bitcoin price forecasting method based on CNN‐LSTM hybrid neural network model. J Eng 2020:344–347. doi: 10.1049/joe.2019.1203 [DOI] [Google Scholar]
  • 21.Lecun Y, Bottou L, Bengio Y, Ha P (1998) LeNet. Proc IEEE 1–46.
  • 22.Lu W, Li J, Wang J, Qin L (2021) A CNN-BiLSTM-AM method for stock price prediction. Neural Comput Appl 33:4741–4753. doi: 10.1007/s00521-020-05532-z [DOI] [Google Scholar]
  • 23.Chen Y, Wu J, Wu Z (2022) China’s commercial bank stock price prediction using a novel K-means-LSTM hybrid approach. Expert Syst Appl 202:117370. doi: 10.1016/j.eswa.2022.117370 [DOI] [Google Scholar]
  • 24.Liu T, Ma X, Li S, et al. (2022) A stock price prediction method based on meta-learning and variational mode decomposition. Knowledge-Based Syst 252:109324. doi: 10.1016/j.knosys.2022.109324 [DOI] [Google Scholar]
  • 25.Lu W, Qiu T, Shi W, Sun X (2022) International Gold Price Forecast Based on CEEMDAN and Support Vector Regression with Grey Wolf Algorithm. Complexity 2022:. doi: 10.1155/2022/1511479 [DOI] [Google Scholar]
  • 26.Kong W, Dong ZY, Luo F, et al (2018) Effect of automatic hyperparameter tuning for residential load forecasting via deep learning. 2017 Australas Univ Power Eng Conf AUPEC 2017 2017-Novem:1–6.
  • 27.Ferdinandus YRM, Kusrini K, Hidayat T (2023) Gold Price Prediction Using the ARIMA and LSTM Models. Sinkron 8:1255–1264. doi: 10.33395/sinkron.v8i3.12461 [DOI] [Google Scholar]
  • 28.Pangestu P, Rochman A, Zuhdi A (2023) The Comparison of Gold Price Prediction Techniques Using Long Short Term Memory (LSTM) And Fuzzy Time Series (FTS) Method. Intelmatics 3:57–62. [Google Scholar]
  • 29.Liang Y, Lin Y, Lu Q (2022) Forecasting gold price using a novel hybrid model with ICEEMDAN and LSTM-CNN-CBAM. Expert Syst Appl 206:117847. doi: 10.1016/j.eswa.2022.117847 [DOI] [Google Scholar]
  • 30.Hochreiter S. and Schmidhuber J. (1997) Long Short-term Memory. Neural Comput 9:1735–1780. doi: 10.1162/neco.1997.9.8.1735 [DOI] [PubMed] [Google Scholar]
  • 31.Ta VD, Liu CM, Tadesse DA (2020) Portfolio optimization-based stock prediction using long-short term memory network in quantitative trading. Appl Sci 10:1–20. doi: 10.3390/app10020437 [DOI] [Google Scholar]
  • 32.Chen Y, Fang R, Liang T, et al (2021) Stock Price Forecast Based on CNN-BiLSTM-ECA Model. 2021.
  • 33.Ren S, Yang B, Zhang L, Li Z (2018) Traffic speed prediction with convolutional neural network adapted for non-linear spatio-temporal dynamics. Proc 7th ACM SIGSPATIAL Int Work Anal Big Geospatial Data, BigSpatial 2018 32–41. doi: 10.1145/3282834.3282836 [DOI] [Google Scholar]
  • 34.Wang S, Xiang J, Zhong Y, Zhou Y (2018) Convolutional neural network-based hidden Markov models for rolling element bearing fault identification. Knowledge-Based Syst 144:65–76. doi: 10.1016/j.knosys.2017.12.027 [DOI] [Google Scholar]
  • 35.Livieris IE, Kiriakidou N, Stavroyiannis S, Pintelas P (2021) An advanced CNN-LSTM model for cryptocurrency forecasting. Electron 10:1–16. doi: 10.3390/electronics10030287 [DOI] [Google Scholar]
  • 36.Vidal A, Kristjanpoller W (2020) Gold volatility prediction using a CNN-LSTM approach. Expert Syst Appl 157:. doi: 10.1016/j.eswa.2020.113481 [DOI] [Google Scholar]
  • 37.Kamalov F (2020) Forecasting significant stock price changes using neural networks. Neural Comput Appl 32:17655–17667. doi: 10.1007/s00521-020-04942-3 [DOI] [Google Scholar]
  • 38.Agga A, Abbou A, Labbadi M, El Houm Y (2021) Short-term self consumption PV plant power production forecasts based on hybrid CNN-LSTM, ConvLSTM models. Renew Energy 177:101–112. doi: 10.1016/j.renene.2021.05.095 [DOI] [Google Scholar]
  • 39.Shi X, Chen Z, Wang H, et al. (2015) Convolutional LSTM network: A machine learning approach for precipitation nowcasting. Adv Neural Inf Process Syst 2015-Janua:802–810. [Google Scholar]
  • 40.Lee SW, Kim HY (2020) Stock market forecasting with super-high dimensional time-series data using ConvLSTM, trend sampling, and specialized data augmentation. Expert Syst Appl 161:113704. doi: 10.1016/j.eswa.2020.113704 [DOI] [Google Scholar]
  • 41.Kingma DP, Ba JL (2015) Adam: A method for stochastic optimization. 3rd Int Conf Learn Represent ICLR 2015—Conf Track Proc 1–15.
  • 42.Ruder S (2016) An overview of gradient descent optimization algorithms. http://arxiv.org/abs/160904747 1–14.
  • 43.Hinton GE, Srivastava N, Krizhevsky A, et al (2012) Improving neural networks by preventing co-adaptation of feature detectors. arXiv Prepr arXiv1207580 1–18.
  • 44.Yu T, Zhu H (2020) Hyper-Parameter Optimization: A Review of Algorithms and Applications. 1–56. [Google Scholar]
  • 45.Afshin M, Sadeghiant A, Raahemifar K (2007) On efficient tuning of LS-SVM hyper-parameters in short-term load forecasting: A comparative study. 2007 IEEE Power Eng Soc Gen Meet PES. doi: 10.1109/PES.2007.385613 [DOI] [Google Scholar]
  • 46.Ranjit MP, Ganapathy G, Sridhar K, Arumugham V (2019) Efficient deep learning hyperparameter tuning using cloud infrastructure: Intelligent distributed hyperparameter tuning with Bayesian optimization in the cloud. IEEE Int Conf Cloud Comput CLOUD 2019-July:520–522. doi: 10.1109/CLOUD.2019.00097 [DOI] [Google Scholar]
  • 47.Amini A, Dolatshahi M, Kerachian R (2023) Effects of Automatic Hyperparameter Tuning on the Performance of Multi-Variate Deep Learning-Based Rainfall Nowcasting. Water Resour Res 59:1–27. doi: 10.1029/2022WR032789 [DOI] [Google Scholar]
  • 48.Bergstra J, Bengio Y (2012) Random search for hyper-parameter optimization. J Mach Learn Res 13:281–305. [Google Scholar]
  • 49.Razavi S (2021) Deep learning, explained: Fundamentals, explainability, and bridgeability to process-based modelling. Environ Model Softw 144:. doi: 10.1016/j.envsoft.2021.105159 [DOI] [Google Scholar]
  • 50.Kulshrestha A, Krishnaswamy V, Sharma M (2020) Bayesian BILSTM approach for tourism demand forecasting. Ann Tour Res 83:102925. doi: 10.1016/j.annals.2020.102925 [DOI] [Google Scholar]
  • 51.Mahto AK, Alam MA, Biswas R, et al. (2021) Short-Term Forecasting of Agriculture Commodities in Context of Indian Market for Sustainable Agriculture by Using the Artificial Neural Network. J Food Qual 2021:. doi: 10.1155/2021/9939906 [DOI] [Google Scholar]
  • 52.Passos D, Mishra P (2022) A tutorial on automatic hyperparameter tuning of deep spectral modelling for regression and classification tasks. Chemom Intell Lab Syst 223:104520. doi: 10.1016/j.chemolab.2022.104520 [DOI] [Google Scholar]
  • 53.Skaggs TH, Barry DA (1996) Sensitivity methods for time-continuous, spatially discrete groundwater contaminant transport models. Water Resour Res 32:2409–2420. doi: 10.1029/96WR01138 [DOI] [Google Scholar]
  • 54.Chollet F (2017) Deep Learning with Python, Manning. Manning Publications.

Decision Letter 0

M Shamim Kaiser

6 Feb 2023

PONE-D-22-31551Gold Price Prediction by a CNN-Bi-LSTM model along with Automatic Parameter TuningPLOS ONE

Dear Dr. Kalantari,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Mar 23 2023 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

M. Shamim Kaiser, PhD

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, all author-generated code must be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse. New software must comply with the Open Source Definition.

3. Thank you for stating the following financial disclosure: 

  "The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript."

At this time, please address the following queries:

a) Please clarify the sources of funding (financial or material support) for your study. List the grants or organizations that supported your study, including funding received from your institution. 

b) State what role the funders took in the study. If the funders had no role in your study, please state: “The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.”

c) If any authors received a salary from any of your funders, please state which authors and which funders.

d) If you did not receive any funding for this study, please state: “The authors received no specific funding for this work.”

Please include your amended statements within your cover letter; we will change the online submission form on your behalf.

4. In your Data Availability statement, you have not specified where the minimal data set underlying the results described in your manuscript can be found. PLOS defines a study's minimal data set as the underlying data used to reach the conclusions drawn in the manuscript and any additional data required to replicate the reported study findings in their entirety. All PLOS journals require that the minimal data set be made fully available. For more information about our data policy, please see http://journals.plos.org/plosone/s/data-availability.

"Upon re-submitting your revised manuscript, please upload your study’s minimal underlying data set as either Supporting Information files or to a stable, public repository and include the relevant URLs, DOIs, or accession numbers within your revised cover letter. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories. Any potentially identifying patient information must be fully anonymized.

Important: If there are ethical or legal restrictions to sharing your data publicly, please explain these restrictions in detail. Please see our guidelines for more information on what we consider unacceptable restrictions to publicly sharing data: http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. Note that it is not acceptable for the authors to be the sole named individuals responsible for ensuring data access.

We will update your Data Availability statement to reflect the information you provide in your cover letter.

Additional Editor Comments:

1. The literature review seems poor to me, authors must include recently published articles.

2. Add detail about the features.

3. Distinguish your work with the state of the art

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Partly

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: No

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Although the article has corrected some errors in presentation, I keep my opinion as the previous review.

The article's contribution is not convincing enough for publication in PLOS ONE.

I suggest the author to submit the article in other more suitable journals about artificial intelligence

Reviewer #2: Thanks to authors for their contributions.

1. In the introduction: It is clearly stated the gap of research.

2. However, it will be better if they use different data-set source.

3. In conclusion, authors mentioned certain distinguish features of the proposed model in compare with existing models. ----> I think it is better it should be write more precisely into the result-analysis section using certain statistical information.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2024 Mar 7;19(3):e0298426. doi: 10.1371/journal.pone.0298426.r003

Author response to Decision Letter 0


8 Mar 2023

Manuscript No. PONEDEC3

The authors would like to thank you for your kind reply to our submitted manuscript. We prepared a revised version of our manuscript with the title “Gold Price Prediction by a CNN-Bi-LSTM model along with

Automatic Parameter Tuning.” upon the useful comments of respected reviewers.

Response to reviewers' comments

---------------------------------------------------------------------------------------------------------------------

At this time, please address the following queries: a) Please clarify the sources of funding (financial or material support) for your study. List the grants or organizations that supported your study, including funding received from your institution.

b) State what role the funders took in the study. If the funders had no role in your study, please state: “The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.” c) If any authors received a salary from any of your funders, please state which authors and which funders. d) If you did not receive any funding for this study, please state: “The authors received no specific funding for this work.” The authors would like to thank the reviewer for the careful review of the manuscript, and for the comments, corrections, and suggestions. The last section is revised as recommended (Funding part) (L415).

Please include your amended statements within your cover letter; we will change the online submission form on your behalf. 4. In your Data Availability statement, you have not specified where the minimal data set underlying the results described in your manuscript can be found. PLOS defines a study's minimal data set as the underlying data used to reach the conclusions drawn in the manuscript and any additional data required to replicate the reported study findings in their entirety. All PLOS journals require that the minimal data set be made fully available. For more information about our data policy, please see http://journals.plos.org/plosone/s/data-availability.

Done. The gold price dataset is available at https://github.com/amirhosseinse7/Closing-gold-price. It is mentioned at L 416.

"Upon re-submitting your revised manuscript, please upload your study’s minimal underlying data set as either Supporting Information files or to a stable, public repository and include the relevant URLs, DOIs, or accession numbers within your revised cover letter. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories. Any potentially identifying patient information must be fully anonymized.

Done! We added data as information files and public repository as aforementioned stated. Important: If there are ethical or legal restrictions to sharing your data publicly, please explain these restrictions in detail. Please see our guidelines for more information on what we consider unacceptable restrictions to publicly sharing data: http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. Note that it is not acceptable for the authors to be the sole named individuals responsible for ensuring data access. We will update your Data Availability statement to reflect the information you provide in your cover letter. Additional Editor Comments: 1. The literature review seems poor to me, authors must include recently published articles. Done! the manuscript has revised as recommended (L130-134) (Two recently published articles have added to this article)

2. Add detail about the features. The manuscript has revised as recommended (L176-177). About the features which extracted by CNN using time series there are not feature mapping method as there is in image sequence forecasting, so we just add the overall statement about the feature extraction of CNNs.

3. Distinguish your work with the state of the art The manuscript has revised as recommended (L72-80) and (L135-140).

Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Partly Reviewer #2: Partly

The manuscript (especially the discussion and conclusion) is revised as recommended. Thank you! 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: No Reviewer #2: Yes

Done. The text has been revised as suggested (changes are available through track change) 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: Although the article has corrected some errors in presentation, I keep my opinion as the previous review. The article's contribution is not convincing enough for publication in PLOS ONE. I suggest the author to submit the article in other more suitable journals about artificial intelligence. The article has changed in different sections. I hope this time the article’s contribution is convincing enough for publication in PLOS ONE. Reviewer #2: Thanks to authors for their contributions. 1. In the introduction: It is clearly stated the gap of research. 2. However, it will be better if they use different data-set source. Thank you for your suggestion. We utilize different datasets in our future research.

3. In conclusion, authors mentioned certain distinguish features of the proposed model in compare with existing models. ----> I think it is better it should be write more precisely into the result-analysis section using certain statistical information. Done! The manuscript has been revised as suggested. Thank you!

With the best regards

Attachment

Submitted filename: Response to reviewers.docx

pone.0298426.s003.docx (27.1KB, docx)

Decision Letter 1

M Shamim Kaiser

6 Apr 2023

PONE-D-22-31551R1Gold Price Prediction by a CNN-Bi-LSTM model along with Automatic Parameter TuningPLOS ONE

Dear Dr. Kalantari,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

 It is advised that authors must complete the following: 

  • The literature review section must be updated. 

  • Addressed all the comments of Reviewer 1

Please submit your revised manuscript by May 21 2023 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

M. Shamim Kaiser, PhD

Academic Editor

PLOS ONE

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: (No Response)

Reviewer #3: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #3: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #3: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #3: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: No

Reviewer #3: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: I still feel that the contribution of the paper is not enough for publication in PLOS ONE.

My suggestions are as follows:

1. Conduct survey of commodity price prediction studies such as:

+ Chen, Yufeng, Jinwang Wu, and Zhongrui Wu. "China’s commercial bank stock price prediction using a novel K-means-LSTM hybrid approach." Expert Systems with Applications 202 (2022): 117370.

+ Liu, Tengteng, et al. "A stock price prediction method based on meta-learning and variational mode decomposition." Knowledge-Based Systems 252 (2022): 109324.

+ Liu, Chong, et al. "An optimized nonlinear grey Bernoulli prediction model and its application in natural gas production." Expert Systems with Applications 194 (2022): 116448.

+ Vo, Anh H., Trang Nguyen, and Tuong Le. "Brent Oil Price Prediction Using Bi-LSTM Network." Intelligent Automation & Soft Computing 26.6 (2020).

+ Pindoriya, N. M., S. N. Singh, and S. K. Singh. "An adaptive wavelet neural network-based energy price forecasting in electricity markets." IEEE Transactions On power systems 23.3 (2008): 1423-1432.

2. Conduct the comparison with state-of-the-art methods for Gold Price Prediction such as:

+ Madziwa, Lawrence, Mallikarjun Pillalamarry, and Snehamoy Chatterjee. "Gold price forecasting using multivariate stochastic model." Resources Policy 76 (2022): 102544.

+ Lu, Wanbo, et al. "International Gold Price Forecast Based on CEEMDAN and Support Vector Regression with Grey Wolf Algorithm." Complexity 2022 (2022).

+ Liang, Yanhui, Yu Lin, and Qin Lu. "Forecasting gold price using a novel hybrid model with ICEEMDAN and LSTM-CNN-CBAM." Expert Systems with Applications 206 (2022): 117847.

Reviewer #3: (No Response)

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #3: Yes: Sobhana Jahan

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2024 Mar 7;19(3):e0298426. doi: 10.1371/journal.pone.0298426.r005

Author response to Decision Letter 1


10 May 2023

Response to reviewers' comments (Revision 2)

The authors would like to thank the reviewer for the careful review of the manuscript, and for the comments, corrections, and suggestions. This document provides a reply to the comments of reviewer. In order to respond, we have added some paragraphs and explained in different sections

---------------------------------------------------------------------------------------------------------------------

General comment

It is advised that authors must complete the following:

• The literature review section must be updated. Done! The recent studies section has been changed as suggested. Thank you for your constructive comment!

• Addressed all the comments of Reviewer 1. The manuscript has been revised as suggested.

Reviewer #1 comments:

I still feel that the contribution of the paper is not enough for publication in PLOS ONE.

My suggestions are as follows:

1. Conduct survey of commodity price prediction studies such as:

+ Chen, Yufeng, Jinwang Wu, and Zhongrui Wu. "China’s commercial bank stock price prediction using a novel K-means-LSTM hybrid approach." Expert Systems with Applications 202 (2022): 117370.

+ Liu, Tengteng, et al. "A stock price prediction method based on meta-learning and variational mode decomposition." Knowledge-Based Systems 252 (2022): 109324.

+ Liu, Chong, et al. "An optimized nonlinear grey Bernoulli prediction model and its application in natural gas production." Expert Systems with Applications 194 (2022): 116448.

+ Vo, Anh H., Trang Nguyen, and Tuong Le. "Brent Oil Price Prediction Using Bi-LSTM Network." Intelligent Automation & Soft Computing 26.6 (2020).

+ Pindoriya, N. M., S. N. Singh, and S. K. Singh. "An adaptive wavelet neural network-based energy price forecasting in electricity markets." IEEE Transactions On power systems 23.3 (2008): 1423-1432.

2. Conduct the comparison with state-of-the-art methods for Gold Price Prediction such as:

+ Madziwa, Lawrence, Mallikarjun Pillalamarry, and Snehamoy Chatterjee. "Gold price forecasting using multivariate stochastic model." Resources Policy 76 (2022): 102544.

This research is related to prediction of annual gold price which is less volatile compared to daily gold price. In addition, the researchers used traditional models like AR based model

+ Lu, Wanbo, et al. "International Gold Price Forecast Based on CEEMDAN and Support Vector Regression with Grey Wolf Algorithm." Complexity 2022 (2022).

+ Liang, Yanhui, Yu Lin, and Qin Lu. "Forecasting gold price using a novel hybrid model with ICEEMDAN and LSTM-CNN-CBAM." Expert Systems with Applications 206 (2022): 117847.

Thank you for your constructive comment. Done! The aforementioned studies are added to the manuscript as suggested.

1) Other methods mentioned in part one will be explored in our second article about gold price forecasting using the combination of tuned DNNs, various decomposition algorithms, and ensemble models.

2) Regarding the comparison part, due to the different lengths of test data and overall data, the results of different studies are not quite comparable. For example, in Lu et al. (2022) the researchers just utilized the last 5% of data (in our case last 10%) as test data and the length of the data was different from ours.

Attachment

Submitted filename: Response to reviewers.docx

pone.0298426.s004.docx (25.7KB, docx)

Decision Letter 2

M Shamim Kaiser

15 Aug 2023

PONE-D-22-31551R2Gold Price Prediction by a CNN-Bi-LSTM model along with Automatic Parameter TuningPLOS ONE

Dear Dr. Kalantari,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

==============================

All comments suggested by the reviewers must be addressed. The quality of the images are poor. Please improve it. Include recently published articles

==============================

Please submit your revised manuscript by Sep 29 2023 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

M. Shamim Kaiser, PhD

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: (No Response)

Reviewer #2: (No Response)

Reviewer #4: (No Response)

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Yes

Reviewer #4: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: No

Reviewer #2: Yes

Reviewer #4: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: Yes

Reviewer #4: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: No

Reviewer #2: Yes

Reviewer #4: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: I am not satisfied with the author's answer on the comparison part.

Besides, I found the topic of the article to be trivial, nothing new, the scientific quality of the article was not satisfactory for publication in PLOS ONE.

Therefore I suggest reject.

Reviewer #2: Certain observations:

1. Chen et al. [24] ---> you mentioned "more accurate than other methods".

Its better to mention any metrics that represent the mentioned phrase.

2. Liu et al. [25] ---> similar explanation like 1.

3. Liang et al. [27] ----> its better to mention the metrics that represent the term "more accurate".

The best solution, I think, make a table and mention the result and properties of the mentioned schemes and others.

Reviewer #4: The paper presents a novel CNN-Bi-LSTM model for closing gold price forecasting. The model is well-designed and the experimental results are convincing. However, there are a few minor concerns that need to be addressed before the paper can be accepted for publication.

1. The paper only uses the closing gold price dataset. It would be interesting to see how the model performs when other influential factors, such as silver price, crude oil, and the U.S. dollar index, are included in the dataset.

2. The author could also provide more discussion of the limitations of the model.

3. The author mentions that data augmentation methods, such as adding noise to the data, may improve the performance of deep neural networks. However, they do not discuss how they might implement these methods in their models or how these augmentation methods might affect the accuracy of the predictions.

4. The paper only uses grid search for hyperparameter tuning. It would be beneficial to compare the results of grid search with other hyperparameter tuning methods, such as random search and Bayesian optimization.

Overall, the paper is well-written and the research is sound. However, the minor concerns mentioned above need to be addressed before the paper can be accepted for publication.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Reviewer #4: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2024 Mar 7;19(3):e0298426. doi: 10.1371/journal.pone.0298426.r007

Author response to Decision Letter 2


12 Oct 2023

Dear Editors of PlOS ONE

The authors would like to thank for your kind reply on our submitted manuscript. We prepared revised version of our manuscript (No. PONE-D-22-31551R2) upon the useful comments of respected reviewers. Before listing response to the comment.

Response to reviewers' comments (Revision 3)

--------------------------------------------------------------------------------------------------------------------- All comments suggested by the reviewers must be addressed. The quality of the images is poor. Please improve it.

Done! The quality of all Figures is improved (Manuscript file)

Include recently published articles

Done! the manuscript has revised as (L162-170) (considering Reviewer #2 comments in these studies)

============================== Reviewers' comments: Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict-of-interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: (No Response) Reviewer #2: (No Response) Reviewer #4: (No Response)

2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Partly Reviewer #2: Yes Reviewer #4: Yes

3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: No Reviewer #2: Yes Reviewer #4: Yes

4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: No The gold price dataset is available at https://github.com/amirhosseinse7/Closing-gold-price. It is mentioned in L 416.

Reviewer #2: Yes Reviewer #4: Yes

5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: No Done. The text has been revised as suggested.

Reviewer #2: Yes Reviewer #4: Yes

6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: I am not satisfied with the author's answer on the comparison part. Besides, I found the topic of the article to be trivial, nothing new, the scientific quality of the article was not satisfactory for publication in PLOS ONE. Therefore, I suggest reject. We have evaluated the Light Gradient Boosted Machine (LightGBM) ensemble model but were unable to improve statistical criteria; therefore, we have omitted this model from our article.

Reviewer #2: Certain observations: 1. Chen et al. [24] ---> you mentioned "more accurate than other methods". It’s better to mention any metrics that represent the mentioned phrase. Done. The manuscript has been changed as suggested (L142-144). Thank you!

2. Liu et al. [25] ---> similar explanation like 1.

The text has been modified as suggested (L150)

3. Liang et al. [27] ----> its better to mention the metrics that represent the term "more accurate". Done. The manuscript has been changed as suggested (L174-175).

The best solution, I think, make a table and mention the result and properties of the mentioned schemes and others. In order to ensure the comprehensiveness of this table, we think it is better to include other relevant publications that are referenced in the methodology and results sections. This inclusion will necessitate a revision of the overall structure of our article. However, in the event that it is deemed essential, we shall incorporate the aforementioned chart.

Reviewer #4: The paper presents a novel CNN-Bi-LSTM model for closing gold price forecasting. The model is well-designed and the experimental results are convincing. However, there are a few minor concerns that need to be addressed before the paper can be accepted for publication. 1. The paper only uses the closing gold price dataset. It would be interesting to see how the model performs when other influential factors, such as silver price, crude oil, and the U.S. dollar index, are included in the dataset. As we have previously addressed the multivariate predicting of closing gold price by incorporating several relevant elements in a separate article, we refrain from including these components in the present paper. This topic was addressed in the concluding section. Thank you for this comment!

2. The author could also provide more discussion of the limitations of the model. The primary constraint of the aforementioned model is its reliance on a substantial volume of data in order to make accurate predictions. This text has been added in L388.

3. The author mentions that data augmentation methods, such as adding noise to the data, may improve the performance of deep neural networks. However, they do not discuss how they might implement these methods in their models or how these augmentation methods might affect the accuracy of the predictions. The performance of models is contingent upon the quantity of data available and the ability to discern patterns within it. Consequently, augmenting the dataset through the utilization of deep learning models like Generative Adversarial Networks (GANs) or employing data augmentation techniques is expected to enhance the accuracy of the outcomes. Moreover, the application of techniques such as wavelet transform, which involves noise reduction and utilization of the denoised data as input for deep learning models, has the potential to enhance the precision of the outcomes. This description has added in (L459-465).

4. The paper only uses grid search for hyperparameter tuning. It would be beneficial to compare the results of grid search with other hyperparameter tuning methods, such as random search and Bayesian optimization. The rationale for not comparing alternative hyperparameter tuning approaches in our study was mostly due to the near-perfect performance achieved by the various models under consideration. Additionally, it was anticipated that these alternative methods would converge to the same hyperparameter values as those obtained by grid search.

Overall, the paper is well-written and the research is sound. However, the minor concerns mentioned above need to be addressed before the paper can be accepted for publication. Thank you!

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No Reviewer #4: No

With the best regards

R. Kalantari

Attachment

Submitted filename: Response to reviewers1.docx

pone.0298426.s005.docx (28.8KB, docx)

Decision Letter 3

Jayesh Soni

14 Dec 2023

PONE-D-22-31551R3Gold Price Prediction by a CNN-Bi-LSTM model along with Automatic Parameter TuningPLOS ONE

Dear Dr. Kalantari,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Jan 28 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Jayesh Soni

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

Additional Editor Comments:

Your work presents an intriguing approach to predicting gold prices using deep learning techniques.

However, I have a few queries that I believe, once addressed, could further strengthen your manuscript:

Given the unpredictability of financial markets, particularly during extraordinary events like the 2008 financial crisis or the recent pandemic, how does your model perform under such atypical conditions? Is there evidence of its adaptability and robustness in these scenarios?

While your model shows significant promise, a comparative analysis with non-CNN and non-LSTM based machine learning models could offer a broader perspective on its relative performance. Have such comparisons been considered or conducted?

Looking forward, what are the avenues for enhancement in your model? Are there upcoming methodologies or data sources that could be leveraged to improve prediction accuracy?

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #3: All comments have been addressed

********** 

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #3: Yes

********** 

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #3: Yes

********** 

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #3: Yes

********** 

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #3: Yes

********** 

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #3: (No Response)

********** 

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #3: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2024 Mar 7;19(3):e0298426. doi: 10.1371/journal.pone.0298426.r009

Author response to Decision Letter 3


18 Dec 2023

Dear Jayesh Soni

Academic Editor of PlOS one

The authors would like to thank you for your kind reply to our submitted manuscript. We prepared a revised version of our manuscript No. PONE-D-22-31551R3 upon the useful comments of respected reviewers.

Manuscript No. PONE-D-22-31551R3

Response to reviewers' comments (Revision 3)

Additional Editor Comments: Your work presents an intriguing approach to predicting gold prices using deep learning techniques. However, I have a few queries that I believe, once addressed, could further strengthen your manuscript: Given the unpredictability of financial markets, particularly during extraordinary events like the 2008 financial crisis or the recent pandemic, how does your model perform under such atypical conditions? Is there evidence of its adaptability and robustness in these scenarios? In LSTM-based models, the number of lag times (time steps) determines how and when these models adjust to atypical conditions, such as the ones you described. As Fig. 5 illustrates, different DNN models quickly adjust to the sequence's peaks points (small scale atypical condition) (depending on the lag time, their reaction time changes). However, a detailed analysis of these models' accuracy in atypical scenarios (large scale atypical conditions) will be investigated in future studies.

While your model shows significant promise, a comparative analysis with non-CNN and non-LSTM based machine learning models could offer a broader perspective on its relative performance. Have such comparisons been considered or conducted? Due to the findings of earlier studies in literature review indicating that DNN-based models outperform ANN models across multiple domains, we do not perform comparative analysis with non-CNN and non-LSTM based machine learning models (such as MLP, RBF, GRNN, etc.) (L59-66)

Looking forward, what are the avenues for enhancement in your model? Are there upcoming methodologies or data sources that could be leveraged to improve prediction accuracy? Thank you, the text has been changed due to this productive question. The model's performance may be enhanced by incorporating additional variables, such as the price of silver, crude oil, and the US dollar index, or by switching from grid search to other automatic hyperparameter tuning techniques like random search or Bayesian methods. Furthermore, the accuracy of the models may be improved by applying methods like wavelet transform, which reduce noise and use denoised data as input for deep learning models (L456-461).

With the best regards

Attachment

Submitted filename: Response to reviewers(1).docx

pone.0298426.s006.docx (25KB, docx)

Decision Letter 4

Jayesh Soni

25 Jan 2024

Gold Price Prediction by a CNN-Bi-LSTM model along with Automatic Parameter Tuning

PONE-D-22-31551R4

Dear Dr. Robab Kalantari,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Jayesh Soni

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

All the comments are addressed.

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

6. Review Comments to the Author

Reviewer: satisfied with authors response. They have modified the paper in best mode. The paper is in the desired stage to be proceed to publish

**********

Acceptance letter

Jayesh Soni

27 Feb 2024

PONE-D-22-31551R4

PLOS ONE

Dear Dr. Kalantari,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

If revisions are needed, the production department will contact you directly to resolve them. If no revisions are needed, you will receive an email when the publication date has been set. At this time, we do not offer pre-publication proofs to authors during production of the accepted work. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few weeks to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Jayesh Soni

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Dataset

    (ZIP)

    pone.0298426.s001.zip (185.9KB, zip)
    Attachment

    Submitted filename: Response to the comments.docx

    pone.0298426.s002.docx (27.1KB, docx)
    Attachment

    Submitted filename: Response to reviewers.docx

    pone.0298426.s003.docx (27.1KB, docx)
    Attachment

    Submitted filename: Response to reviewers.docx

    pone.0298426.s004.docx (25.7KB, docx)
    Attachment

    Submitted filename: Response to reviewers1.docx

    pone.0298426.s005.docx (28.8KB, docx)
    Attachment

    Submitted filename: Response to reviewers(1).docx

    pone.0298426.s006.docx (25KB, docx)

    Data Availability Statement

    The gold price dataset is available at https://github.com/amirhosseinse7/Closing-gold-price.


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES