Skip to main content
PLOS One logoLink to PLOS One
. 2024 Jun 21;19(6):e0303883. doi: 10.1371/journal.pone.0303883

Impact of economic indicators on rice production: A machine learning approach in Sri Lanka

Sherin Kularathne 1, Namal Rathnayake 2,*, Madhawa Herath 3, Upaka Rathnayake 4, Yukinobu Hoshino 5
Editor: Sandeep Samantaray6
PMCID: PMC11192421  PMID: 38905194

Abstract

Rice is a crucial crop in Sri Lanka, influencing both its agricultural and economic landscapes. This study delves into the complex interplay between economic indicators and rice production, aiming to uncover correlations and build prediction models using machine learning techniques. The dataset, spanning from 1960 to 2020, includes key economic variables such as GDP, inflation rate, manufacturing output, population, population growth rate, imports, arable land area, military expenditure, and rice production. The study’s findings reveal the significant influence of economic factors on rice production in Sri Lanka. Machine learning models, including Linear Regression, Support Vector Machines, Ensemble methods, and Gaussian Process Regression, demonstrate strong predictive accuracy in forecasting rice production based on economic indicators. These results underscore the importance of economic indicators in shaping rice production outcomes and highlight the potential of machine learning in predicting agricultural trends. The study suggests avenues for future research, such as exploring regional variations and refining models based on ongoing data collection.

Introduction

Rice, which provides nearly half the calories for half the world’s population [1, 2] is a vital crop for Sri Lanka’s agricultural and economic landscape, being the staple food, which determines food security and economic resiliency with more than 181kg per year consumption, marking the eighth greatest per capita consumption in the world [3]. Also, rice cultivation accounts for 34% of the total cultivated area, 1.8 million family farms, and 45% of all calories consumed by the average Sri Lankan [4].

The complex interplay of economic forces and rice production necessitates a thorough analysis and a nuanced knowledge of the multifaceted relationships that shape this dynamic. This study aims to contribute to this discussion by doing a thorough investigation and using machine learning approaches to uncover correlations and build prediction models.

The factors examined in this study are diversified facets of economic indicators, each with a unique impact on the intricate web of rice production. At the forefront is Gross Domestic Product (GDP), a vital indicator of the country’s economic health that reflects and drives agricultural development. Along with GDP, the changing inflation is also examined. As Bhattarai mentioned [5], maintaining pricing stability is crucial for the economy‘s long-term growth and development. According to Atigala’s findings on Sri Lanka [6], there is a negative relationship between inflation and economic growth in the short run as well. Furthermore, because growing inflation negatively impacts the economy, testing inflation’s relationship with various economic factors is crucial for making proactive economic decisions. The strength of the manufacturing sector and the demographic underpinnings of population and growth can influence the setting in which rice production succeeds or fails, as they are some of the main economic indicators that have no proof of being tested out in the Sri Lankan context against rice production.

Moreover, the political instability in recent years has also been a motivating factor for conducting this study in Sri Lanka. Political instability can have significant implications for the economy, including the agricultural sector. By studying the relationship between economic indicators and rice production in Sri Lanka, the authors may aim to provide insights that can help mitigate the impact of political instability on rice cultivation and inform more stable and resilient agricultural policies. Additionally, the study seeks to understand how political factors interact with economic indicators to influence rice production, providing valuable insights for policymakers navigating the challenges of political uncertainty.

Furthermore, the study delves beyond the domestic domain, examining the impact of foreign trade using variables like imports. The quantity and quality of arable land, a finite resource crucial to agricultural sustainability, adds an environmental component to our inquiry. To reflect the broader economic backdrop, the study includes the influence of military expenditures, recognizing the interaction between military defence spending and its possible impact on the agricultural sector. Even though most studies in the literature [79] suggest that there is no significant consensus on the impact that military expenditure can have on the economic factors of a region, there is an absence of few studies that specifically focus on against rice production and in Sri Lankan context.

As this study navigates the complicated interdependence of these variables, machine learning can be used as a valuable instrument, capable of discovering patterns from vast data sets and providing insights into the deep relationships that influence economic factors and rice production. This study intends to find correlations while also providing a view into the future, allowing policymakers, agricultural stakeholders, and economists to make informed decisions.

Food production is a critical component of global agricultural systems, and understanding the factors influencing it has been the focus of extensive research. Over the years, numerous studies have investigated the relationship between food production and various methods, including statistical and machine-learning approaches [10, 11]. The following paragraphs aim to provide an overview of key research efforts in this area.

Climate change is one of the significant factors affecting food production, and several studies have explored its potential impact using statistical and machine learning methods Mulla et al. [12] and Rosenzweig et al. [13] simulated the effects of climate change on agriculture, highlighting the importance of temperature and precipitation. Lobell and Burke [14] examined the uncertainty in agricultural impacts of climate change, emphasizing the role of temperature relative to precipitation.

In addition to climate change, researchers have focused on understanding the patterns of crop yield growth and stagnation. Ray et al. [15] analyzed recent patterns of crop yield growth, while Liakos et al. [16] and Krishna et al. [17] reviewed machine learning approaches for crop yield prediction, highlighting the importance of these methods in agricultural research.

Machine learning has emerged as a powerful tool in agriculture, offering new opportunities for analyzing complex agricultural data. Satpathi et al. [18] conducted a comparative study of machine learning models for forecasting rice production, demonstrating the effectiveness of these models in predicting agricultural outcomes.

Vasilyevich [19] and Liakos et al. [16] explored the classification of agricultural data using machine learning algorithms, showcasing the potential of these methods in agricultural data analysis. Additionally, Furbank and Tester [20] discussed the role of phenomics technologies in relieving the phenotyping bottleneck in agriculture.

Machine learning algorithms have been applied to various aspects of agriculture, including agricultural data classification, flood prediction [2125] and crop yield prediction [26, 27]. These studies demonstrate the versatility and effectiveness of machine learning in addressing agricultural challenges.

The study of economic indicators and their impact on rice production is a complex and dynamic field that requires sophisticated analytical tools. In recent years, fuzzy logic has emerged as a valuable approach for modeling complex systems with uncertain or imprecise data, making it particularly well-suited for economic predictions. Fuzzy logic allows for the representation of vague or ambiguous relationships between variables, which is often the case in economic systems where factors can be influenced by multiple variables with varying degrees of impact. For example, studies by Jagielska et al. [28] and Reghis et al. [29] have demonstrated the effectiveness of fuzzy logic in modeling economic systems with uncertain or incomplete information, providing insights into the decision-making processes that govern economic outcomes.

Furthermore, fuzzy logic has been applied to various aspects of economic forecasting, including stock market predictions, consumer behavior analysis, and macroeconomic modeling. For instance, studies by Ecer et al. [30] and Pant et al. [31] have used fuzzy logic to predict stock prices based on a range of economic indicators and market sentiment analysis. These studies have shown that fuzzy logic can capture the non-linear and uncertain nature of financial markets, leading to more accurate predictions compared to traditional statistical methods. In the context of rice production in Sri Lanka, fuzzy logic could offer a valuable tool for modeling the complex interactions between economic factors and agricultural outcomes, providing insights that can enhance the resilience and sustainability of rice cultivation in the face of economic uncertainties [32, 33].

Hybrid fuzzy systems, such as the Adaptive Neuro-Fuzzy Inference System (ANFIS) and its variations [34, 35], have gained popularity for their ability to combine the strengths of fuzzy logic and neural networks in modeling complex economic systems. ANFIS integrates fuzzy logic principles with neural network learning algorithms, allowing for the automatic learning of fuzzy rules from data. This approach has been successfully applied in various economic forecasting tasks, including time series prediction, financial market analysis, and macroeconomic modeling. For example, studies by Jang et al. [36], and Boyacioglu et al. [37] have demonstrated the effectiveness of ANFIS in predicting stock prices and exchange rates, showcasing its ability to capture the non-linear and uncertain relationships inherent in financial markets. In the context of rice production in Sri Lanka, the use of hybrid fuzzy systems like ANFIS could offer a powerful tool for modeling the complex interactions between economic indicators and agricultural outcomes, providing accurate predictions and valuable insights for decision-making.

In conclusion, the relationship between food production and statistical/machine learning methods has been the subject of extensive research. Studies have highlighted the importance of these methods in understanding the complex dynamics of agricultural systems, predicting crop yields, and addressing challenges such as climate change impacts. As technology continues to advance, the application of statistical and machine learning methods in agriculture is expected to play an increasingly significant role in shaping the future of food production.

This research aims to go beyond a surface-level understanding of economic factors and rice production in Sri Lanka. By incorporating machine learning approaches, the research hopes to contribute to a more sophisticated understanding of the connected threads that connect the country’s economic growth to the health of its agricultural backbone. Through such insights, this research can pave the road for evidence-based policies and initiatives that promote a resilient future for Sri Lankan rice production in the face of a dynamic economic landscape.

The comprehensive investigation presented in the above paragraphs points out the following flaws of past studies.

  1. Limited Scope: Past studies may have focused on a narrow set of economic indicators or environmental factors, overlooking the complexity and interdependence of various factors that influence rice production.

  2. Lack of Local Context: Some studies may have lacked a specific focus on the Sri Lankan context, potentially leading to findings that are not directly applicable to the local agricultural landscape.

  3. Static Analysis: Previous research might have relied on static analysis methods without leveraging advanced machine learning techniques, which could limit the depth of insights into the dynamic relationships between economic factors and rice production.

Hence, this study focuses on achieving the following objectives.

  1. Comprehensive Analysis: This research aims to provide a thorough investigation of the multifaceted relationships between economic indicators and rice production in Sri Lanka, offering a more comprehensive understanding of the factors at play.

  2. Integration of Machine Learning: By incorporating machine learning approaches, the study seeks to uncover correlations and build prediction models that can provide insights into the complex dynamics of economic forces and rice cultivation.

  3. Potential Policy Implications: The findings of this research could inform evidence-based policies and initiatives that promote a resilient future for Sri Lankan rice production, highlighting the practical applications of the study’s insights for policymakers and agricultural stakeholders.

This study presents a novel and comprehensive analysis of the multifaceted relationships between economic indicators and rice production in Sri Lanka, integrating machine learning approaches to uncover correlations and build prediction models, with findings directly applicable to the local agricultural landscape and potential implications for evidence-based policies and initiatives.

The findings of this study contribute significantly to addressing the challenges identified in the introduction by offering a comprehensive analysis of the complex relationships between economic indicators and rice production in Sri Lanka. By integrating machine learning approaches, the study provides a deeper understanding of these relationships, which can inform evidence-based policies and initiatives to promote a resilient future for Sri Lankan rice production. Additionally, the focus on the local context ensures that the findings are directly applicable to the specific challenges faced by the country’s agricultural sector, bridging the gap in the literature and providing actionable insights for policymakers and agricultural stakeholders.

Methodology

Fig 1 represents the overall flow of the study. The study follows a structured flow beginning with data collection, where comprehensive datasets spanning from 1960 to 2020 are gathered from various sources. The next step involves data preparation, where the collected data is cleansed, integrated, and transformed to ensure consistency and compatibility for analysis. This leads to the data exploration phase, wherein the distribution and characteristics of the variables are carefully examined to gain insights into their behavior over time.

RiceProduction=function{GDP($B)Inflation(%)Manufacturing($B)PopulationGrowthRate(%)Imports($B)ArableLandMilitary($B) (1)

Fig 1. Overall flow of the study.

Fig 1

Following this, correlation analysis is conducted to identify the variables that exhibit strong correlations with rice production, shedding light on potential influential factors. Eq 1 summarizes the assumed relationship between economic factors and rice production.

Subsequently, traditional and modern state-of-the-art machine learning regression models are trained using the prepared dataset to predict rice production based on the identified factors.

The results of the trained models are evaluated using standard evaluation parameters such as Root Mean Squared Error (RMSE), R-squared (R2), Mean Squared Error (MSE), and Mean Absolute Error (MAE) to assess their predictive performance. Finally, the study concludes with insights and interpretations drawn from the results, highlighting the significant economic factors that influence rice production in Sri Lanka. This comprehensive approach aims to provide a general understanding of the relationship between economic variables and rice production, contributing to the region’s broader understanding of agricultural dynamics.

Data description and preprocessing

The dataset spans from 1960 to 2020 and includes various economic indicators and rice production data from Sri Lanka. These indicators consist of the country’s GDP in billions of dollars, inflation rate as a percentage, manufacturing output in billions of dollars, total population count, population growth rate as a percentage, imports in billions of dollars, arable land area in square kilometres, military expenditure in billions of dollars, and rice production in metric tons. The dataset’s range of values varies across different years and indicators. For instance, the GDP ranges from a minimum of 1.24 billion dollars in 1963 to a maximum of 94.49 billion dollars in 2018, while the inflation rate fluctuates between -0.02% and 0.26%. Similarly, other variables like manufacturing output, population, imports, arable land, military expenditure, and rice production also exhibit varying ranges over the years. In addition to the dataset containing economic indicators and rice production data from Sri Lanka between 1960 and 2020, further statistical information is available for each variable. This includes the count, mean, standard deviation (std), minimum (min), 25th percentile (25%), 50th percentile (50% or median), 75th percentile (75%), and maximum (max) values for each variable. A descriptive analysis of the dataset is presented in Table 1.

Table 1. Summary of statistical measures for economic and agricultural variables in Sri Lanka (1960-2020).

Variables count mean std min 25% 50% 75% max
Year 61 1990 17.75 1960 1975 1990 2005 2020
GDP ($B) 61 23 29.47 1.24 3.36 8.03 24.41 94.49
Inflation (%) 61 0 0.06 -0.02 0.04 0.08 0.12 0.26
Manufacturing ($B) 61 4 4.95 0.2 0.61 1.08 4.76 14.39
Population 61 1.66 × 107 3.64 × 106 9.78 × 106 1.37 × 107 1.72 × 107 1.97 × 107 2.17 × 107
Growth Rate (%) 61 0 0.01 0 0.01 0.01 0.02 0.02
Imports ($B) 61 7 7.64 0.34 1.13 3.06 10.07 26.8
Arable Land 61 9.58 × 105 1.99 × 105 5.77 × 105 8.05 × 1051.97 × 107 9.00 × 105 1.00 × 106 1.37 × 106
Military($B) 61 1 0.66 0.01 0.03 0.28 0.82 2.06
Rice Production (Mt.) 61 2311 883.53 758 1605 2389 2860 4301

To prepare the dataset for analysis, several preprocessing steps can be undertaken. This includes checking for missing or inconsistent values and handling them appropriately, such as through imputation or removal of outliers. Additionally, normalizing or standardizing numeric variables may be necessary to ensure that they are on a similar scale, especially if using machine learning algorithms that are sensitive to scale differences. Feature engineering could also enhance the dataset’s predictive power by creating or transforming new features. Furthermore, throughout this study, the dataset was divided into training and testing sets using a consistent 7:3 ratio. The same train and test dataset was used consistently for all analyses and experiments.

Correlation analysis

The correlation analysis reveals interesting insights into the relationships between various factors and rice production (‘Rice Production (Mt.)’). The highest positive correlation between rice production and the year indicates a potential trend of increasing rice production over time. Additionally, strong positive correlations are found between rice production and economic indicators such as GDP (‘Gross Domestic Product’), manufacturing output, population, imports, arable land, and military expenditure. These findings suggest that economic growth, population dynamics, and resource availability, as represented by arable land and imports, may significantly influence rice production. Conversely, a notable negative correlation is observed between rice production and the growth rate, indicating a potential inverse relationship between economic growth and rice production (Refer to Fig 2).

Fig 2. The correlation coefficients of selected variables with ‘Rice Production (Mt.)’.

Fig 2

Machine learning models

This study employs a range of machine learning models, spanning traditional algorithms like Linear Regression, Decision Trees, and Support Vector Regression, which are renowned for their interpretability and efficiency. On the cutting edge, modern techniques like Neural Networks and Gaussian Process Regression harness the power of deep learning and ensemble methods to capture intricate data relationships. This section comprehensively explores these algorithms, each contributing its unique strengths to accurate prediction.

Traditional ML models

Linear regression (Linear). Linear regression is a basic and widely used statistical method for modeling the relationship between a dependent variable and one or more independent variables. It assumes a linear relationship between the input variables and the output. The pseudo-code of the algorithm is shown in 1 and tuned hyper parameters are shown in Table 2.

Table 2. Tuned hyperparameters for linear regression.
Hyperparameter Value
fit_intercept True
normalize True

Algorithm 1 Linear Regression

1: procedure LinearRegression(X, y)

 (X represents the feature matrix, y represents the target variable)

2:   Initialize the coefficients vector β

3:   Add a column of ones to the feature matrix X (for the intercept term)

4:   Compute the least squares solution: β = (XTX)−1XTy

5:   return β

6: end procedure

Linear regression (Robust). Robust linear regression is a variation of linear regression that is less sensitive to outliers in the data. It uses robust estimation techniques to minimize the impact of outliers on the model’s parameters. The pseudo-code of the algorithm is shown in 2 and tuned hyper parameters are shown in Table 3.

Table 3. Hyperparameters for robust linear regression.
Hyperparameter Value
delta 1.0
num_iterations 100

Algorithm 2 Robust Linear Regression (Huber Regression)

1: procedure RobustLinearRegression(X, y, δ, num_iterations)

 (X represents the feature matrix, y represents the target variable, and δ represents the threshold for the Huber loss function)

2:   Initialize the coefficients vector β

3:   Add a column of ones to the feature matrix X (for the intercept term)

4:   for i = 1 to num_iterations do

5:    Compute the residuals: r = y

6:    Compute the weights using Huber loss function: W = diag(1, …, 1) if |r| ≤ δ, otherwise W = diag(δ/|r|)

7:    Update the coefficients using weighted least squares: β = (XTWX)−1XTWy

8:   end for

9:   return β

10: end procedure

Tree (Fine, Medium). These refer to decision tree models with different levels of complexity. Decision trees partition the input space into regions and assign a constant value to each region, making them easy to interpret. The pseudo-code of the algorithm is shown in 3 and tuned hyperparameters are shown in Table 4.

Table 4. Hyperparameters for decision trees.
Tree Complexity Hyperparameters
Fine Max Depth: 10, Min Samples Split: 2
Medium Max Depth: 5, Min Samples Split: 5

Algorithm 3 Decision Tree

1: procedure DecisionTree(X, y, max_depth)

2:  if max_depth = Fine then

3:   Call RecursiveTree(X, y, max_depth, Fine)

4:  else if max_depth = Medium then

5:   Call RecursiveTree(X, y, max_depth, Medium)

6:  end if

7: end procedure

8: procedure RecursiveTree(X, y, max_depth, complexity)

9:  if stopping_criterion_met then

10:   Create a leaf node with the majority class in y

11:  else if complexity = Fine then

12:   Select the best split attribute and value

13:   Split the data into two subsets based on the selected split

14:   Call RecursiveTree(Xleft, yleft, max_depth—1, Fine)

15:   Call RecursiveTree(Xright, yright, max_depth—1, Fine)

16:  else if complexity = Medium then

17:   Select the best split attribute and value using a faster heuristic

18:   Split the data into two subsets based on the selected split

19:   Call RecursiveTree(Xleft, yleft, max_depth—1, Medium)

20:   Call RecursiveTree(Xright, yright, max_depth—1, Medium)

21:  end if

22: end procedure

SVM (Linear, Quadratic, Cubic, Fine Gaussian, Medium Gaussian, Coarse Gaussian). Support Vector Machines (SVMs) are a class of supervised learning algorithms used for classification and regression tasks. They work by finding the hyperplane that best separates the classes in the input space. The variations you’ve listed refer to different types of kernel functions used in SVMs, which determine the shape of the decision boundary. The pseudo-code of the algorithm is shown in 4 and tuned hyperparameters are shown in Table 5.

Table 5. Hyperparameters for support vector machines.
Kernel Type Hyperparameters
Linear C: 1.0
Quadratic C: 1.0, Degree: 2
Cubic C: 1.0, Degree: 3
Fine Gaussian C: 1.0, Gamma: 0.001
Medium Gaussian C: 1.0, Gamma: 0.01
Coarse Gaussian C: 1.0, Gamma: 0.1

Algorithm 4 Support Vector Machine

1: procedure SVM(X, y, kernel, complexity)

2:  if kernel = Linear then

3:   Call LinearSVM(X, y, complexity)

4:  else if kernel = Quadratic then

5:   Call PolynomialSVM(X, y, 2, complexity)

6:  else if kernel = Cubic then

7:   Call PolynomialSVM(X, y, 3, complexity)

8:  else if kernel = Fine Gaussian then

9:   Call GaussianSVM(X, y, Fine)

10:  else if kernel = Medium Gaussian then

11:   Call GaussianSVM(X, y, Medium)

12:  else if kernel = Coarse Gaussian then

13:   Call GaussianSVM(X, y, Coarse)

14:  end if

15: end procedure

16: procedure LinearSVM(X, y, complexity)

17:  Train SVM with linear kernel using optimization algorithm with complexity setting

18: end procedure

19: procedure PolynomialSVM(X, y, degree, complexity)

20:  Train SVM with polynomial kernel of degree degree using optimization algorithm with complexity setting

21: end procedure

22: procedure GaussianSVM(X, y, granularity)

23:  if granularity = Fine then

24:   Train SVM with fine Gaussian kernel using optimization algorithm

25:  end if granularity = Medium then

26:   Train SVM with medium Gaussian kernel using optimization algorithm

27:  end if granularity = Coarse then

28:   Train SVM with coarse Gaussian kernel using optimization algorithm

29:  end if

30: end procedure

Ensemble (Boosted, Bagged). Ensemble methods combine multiple models to improve predictive performance. Sequentially boosting train models, with each subsequent model focusing on the errors of the previous ones. Bagging trains multiple models independently and combines their predictions through averaging or voting. The pseudo-codes of the algorithm are shown in 5 and 6, and tuned hyperparameters are shown in Table 6.

Table 6. Hyperparameters for ensemble learning.
Ensemble Method Hyperparameters
Boosted Number of Estimators: 100, Learning Rate: 0.1
Bagged Number of Estimators: 100, Bootstrap Samples: True

Algorithm 5 Boosting

1: procedure Boosting(X, y, base_learner, num_learners)

2:  Initialize the weights wi = 1/n, where n is the number of samples

3:  Initialize an empty list of weak learners: h1, h2, …, hnum_learners

4:  for t = 1 to num_learners do

5:   Train a weak learner ht using weighted samples X and y with weights w

6:   Compute the error ϵt of ht on the weighted samples

7:   Compute the weight αt=12log(1-ϵtϵt)

8:   Update the weights: wiwi ⋅ exp(−αtyiht(xi)), where xi is the i-th sample

9:   Normalize the weights: wwi=1nwi

10:   Add ht to the list of weak learners

11:  end for

12:  return h1, h2, …, hnum_learners

13: end procedure

Algorithm 6 Bagging

1: procedure Bagging(X, y, base_learner, num_learners)

2:  Initialize an empty list of weak learners: h1, h2, …, hnum_learners

3:  for t = 1 to num_learners do

4:   Sample n examples from X with replacement

5:   Train a weak learner ht on the sampled data

6:   Add ht to the list of weak learners

7:  end for

8:  return h1, h2, …, hnum_learners

9: end procedure

Modern ML models

Gaussian Process Regression (Squared Exp, Matern 5/2, Exponential, Rational Quadratic Exp). GPR is a non-parametric Bayesian approach to regression. It models the target variable as a Gaussian process, which is characterized by a mean function and a covariance function (kernel). The variations that were captured refer to different kernel functions used in GPR, which capture different types of smoothness in the data. The pseudo-code of the algorithm is shown in 7 and tuned hyperparameters are shown in Table 7.

Table 7. Hyperparameters for gaussian process regression.
Kernel Function Hyperparameters
Squared Exponential Length Scale: 1.0, Variance: 0.1
Matern 5/2 Length Scale: 0.5, Variance: 0.2
Exponential Length Scale: 0.8, Variance: 0.15
Rational Quadratic Exp Length Scale: 1.2, Scale Mixture: 0.3, Variance: 0.1

Algorithm 7 Gaussian Process Regression

1: procedure GaussianProcessRegression(X, y, kernel)

2:  Compute the kernel matrix K using the selected kernel function and input data X

3:  Add a small noise term σ2I to the kernel matrix K for numerical stability

4:  Compute the Cholesky decomposition of the kernel matrix: L = Cholesky(K + σ2I)

5:  Compute the alpha vector: α = (LT)−1(L−1y)

6:  return mean function and covariance function determined by K and α

 (The kernel matrix K is computed based on the chosen kernel function and input data X. To ensure numerical stability, a small noise term σ2I is added to the kernel matrix before performing the Cholesky decomposition.)

7: end procedure

These models represent a diverse set of machine learning tools, each with its own strengths and weaknesses. When choosing a model for a specific task, it’s important to consider factors such as the data’s nature, the problem’s complexity, and the interpretability of the model’s results. These algorithms represent a diverse toolkit, each offering unique advantages for modeling the complex relationships within the economic factors affecting rice production.

Evaluation parameters

This section explores key performance metrics employed to assess the accuracy and reliability of the developed wind power prediction models. Root Mean Squared Error (RMSE), Mean Absolute Error (MAE), Mean Squared Error (MSE), and Coefficient of Determination (R2) are discussed in detail, providing insights into the models’ predictive capabilities. The analysis of these metrics enhances our understanding of the model’s accuracy, highlighting their strengths and areas for improvement.

Root Mean Squared Error (RMSE)

RMSE is a widely used metric that measures the average magnitude of the errors between predicted (yi) and observed (y¯i) values (Eq 2) [38]. It comprehensively assesses the model’s accuracy, where lower values indicate better performance.

RMSE=1Ni=1N(yi-y¯i)2 (2)

Where N is the total number of observations.

Mean Absolute Error (MAE)

MAE quantifies the average absolute difference between predicted (y^i) and observed (yi) values, providing a measure of the model’s accuracy without considering the direction of the errors (Eq 3) [39].

MAE=1Ni=1N|yi-y^i| (3)

Where N is the total number of observations.

Mean Squared Error (MSE)

MSE measures the average of the squares of the errors between predicted (y^i) and observed (yi) values, providing a measure of the model’s accuracy that emphasizes larger errors than MAE (Eq 4) [40].

MSE=1Ni=1N(yi-y^i)2 (4)

Coefficient of Determination (R2)

The Coefficient of Determination, often denoted as R2, assesses the proportion of the variance in the dependent variable (wind power generation) that is predictable from the independent variables (features). It ranges between 0 and 1, where higher values indicate better explanatory power (Eq 5) [33].

R2=1-i=1N(yi-y^i)2i=1N(yi-y¯i)2 (5)

Where N is the total number of observations. yi is the observed value. y^i is the predicted value. y¯i is the mean of the observed values.

These performance indexes collectively evaluate the model’s predictive accuracy, highlighting different aspects of the prediction errors.

Results and discussion

Figs 3 and 4 depict the top eight test results of prediction and actual values using a scatter plot. These results highlight the comparative performance of various machine learning algorithms in predicting the target variable. The evaluated algorithms include Gaussian Process Regression (GPR) with different kernels (Exponential, Squared Exponential, and Matern), Support Vector Machine (SVM) with a linear kernel, Linear Regression (both linear and robust), Ensemble bagging, and Decision Tree with coarse branching.

Fig 3. Comparison of actual and predicted values, residual plot, and residual distribution.

Fig 3

(a) Gaussian Process Regression (GPR)—Exponential (b) Gaussian Process Regression (GPR)—SquaredExponential (c) Gaussian Process Regression (GPR)—Matern (nu = 5/2) and (d) Support Vector Machine (SVM)—Linear.

Fig 4. Comparison of actual and predicted values, residual plot, and residual distribution.

Fig 4

(a) Linear Regression—Linear (b) Linear Regression—Robust (c) Ensemble bagging and (d) Decision tree—Coarse.

Table 8 presents the comprehensive results of the study. The GPR models consistently exhibit strong performance, with all three variants achieving high R2 values of 0.94. This indicates a high degree of explained variance in the predictions, supported by low RMSE, MSE, and MAE values, indicating accurate prediction of the target variable.

Table 8. Performance metrics (R2, RMSE, MSE, MAE) for eight machine learning algorithms in predicting economic factors related to rice production.

Algorithm R2 RMSE MSE MAE
Gaussian Process Regression (GPR) Exponential 0.9437 230 52886 169
Gaussian Process Regression (GPR) SquaredExponential 0.9422 233 54332 181
Gaussian Process Regression (GPR) Matern (nu = 5/2) 0.9415 234 54989 180
Support Vector Machine (SVM) Linear 0.9406 236 55830 191
Linear Regression Linear 0.9226 270 72698 229
Linear Regression Robust 0.9226 270 72698 229
Ensemble bagging 0.9161 281 78819 235
Decision tree Coarse 0.8392 389 151046 336

Similarly, the SVM with a linear kernel demonstrates promising results, achieving an R2 value of 0.94 and competitive RMSE, MSE, and MAE values. The linear regression models also exhibit strong predictive capabilities, with R2 values of 0.92 and low error metrics.

The Ensemble bagging model performs adequately with an R2 value of 0.92, indicating good predictive performance, although slightly lower than the GPR, SVM, and Linear Regression models. In contrast, the Decision Tree model with coarse branching demonstrates the weakest performance among the algorithms considered, with an R2 value of 0.84 and higher error metrics, indicating a less accurate prediction of the target variable compared to other models.

The study compared the responses of the testing predictions generated by eight different algorithms and presented the results in Figs 5 and 6.

Fig 5. Response plot showing predicted and actual values over time.

Fig 5

(a) Gaussian Process Regression (GPR)—Exponential (b) Gaussian Process Regression (GPR)—SquaredExponential (c) Gaussian Process Regression (GPR)—Matern (nu = 5/2) and (d) Support Vector Machine (SVM)—Linear.

Fig 6. Response plot showing predicted and actual values over time.

Fig 6

(a) Linear Regression—Linear (b) Linear Regression—Robust (c) Ensemble bagging and (d) Decision tree—Coarse.

In addition to the performance comparison, it is essential to consider the broader implications of the study. The findings of this study can be compared with similar studies in the literature to provide a more comprehensive understanding of the performance of different machine learning algorithms in predicting economic factors related to rice production. Furthermore, the study provides valuable insights into the advantages of using GPR, SVM with a linear kernel, and Linear Regression models for similar prediction tasks.

The study’s comprehensive analysis of the relationship between economic indicators and rice production in Sri Lanka offers valuable insights that can inform evidence-based policymaking and agricultural strategies. By uncovering the complex interplay of factors influencing rice production, the study provides a nuanced understanding of the challenges and opportunities facing the agricultural sector in Sri Lanka.

One practical implication of the study is its potential to guide policymakers in developing strategies to enhance rice production and agricultural sustainability. By identifying key economic indicators that impact rice cultivation, policymakers can tailor interventions to address specific challenges and capitalize on opportunities for growth. For example, the study’s findings suggest that focusing on boosting the manufacturing sector and maintaining price stability could positively influence rice production. Similarly, strategies to improve arable land quality and manage inflation could lead to more resilient agricultural systems.

Furthermore, the study’s use of machine learning approaches adds a new dimension to its practical implications. By demonstrating the effectiveness of these techniques in predicting rice production outcomes, the study highlights the potential for integrating advanced technologies into agricultural decision-making. This could lead to more accurate and timely predictions, allowing farmers and policymakers to respond proactively to changing conditions.

Moreover, the study opens up avenues for future research. Further investigations could explore the application of other machine learning algorithms or ensemble methods to improve predictive performance. Additionally, incorporating more features or datasets could enhance the accuracy of the predictions. This study lays the foundation for future research in the field of predicting economic factors related to rice production using machine learning algorithms.

In summary, this study provides a detailed comparison of different machine learning algorithms’ performance in predicting economic factors related to rice production. The results offer valuable insights for researchers, policymakers, and stakeholders interested in selecting the most suitable machine learning algorithm for similar prediction tasks.

Conclusion

In conclusion, this study provides a novel and comprehensive analysis of the multifaceted relationships between economic indicators and rice production in Sri Lanka, integrating machine learning approaches to uncover correlations and build prediction models. The findings of this research offer practical implications for policymakers and agricultural stakeholders, providing insights that can inform evidence-based policies and initiatives to promote a resilient future for Sri Lankan rice production. The study’s focus on the local context ensures that the findings are directly applicable to the specific challenges faced by the country’s agricultural sector, bridging the gap in the literature and providing actionable insights for decision-makers. Furthermore, the transferability of the study’s results to other locations facing similar challenges of political instability and reliance on rice as a staple food highlights the broader impact and usefulness of this work.

The transferability of the study’s results to other locations can be significant, particularly in regions facing similar challenges of political instability and where rice is a staple food. The findings can offer valuable insights into how economic indicators impact rice production, which can be applied to other countries experiencing political turmoil. Additionally, since many countries rely on rice as a staple food, the study’s results can be useful for policymakers and stakeholders globally, providing a framework for understanding the complex interplay between economic factors and rice cultivation. By considering the specific context of Sri Lanka’s political instability and the widespread use of rice as a staple, the study’s findings can serve as a blueprint for enhancing agricultural sustainability and economic resilience in other regions facing similar circumstances.

Furthermore, this study paves the way for future research by suggesting several potential avenues. Subsequent investigations could delve into employing alternative machine learning algorithms or ensemble methods to enhance predictive accuracy. Additionally, the incorporation of additional features or datasets could further refine the predictive models. Overall, this study contributes significantly to addressing the challenges identified in the introduction and lays the foundation for future research on using machine learning algorithms to predict economic factors related to rice production.

Data Availability

The data is stored in a public repository for reproducibility (https://www.kaggle.com/datasets/namalrathnayake1990/economy-and-rice-production-sri-lanka-1960-2020).

Funding Statement

This study was funded by the Japan Society for the Promotion of Science (JSPS) in the form of a Grants-in-Aid for Scientific Research (KAKENHI) grant to YH [22KK0160].

References

  • 1. Khoury C., Bjorkman A., Dempewolf H., Ramirez-Villegas J., Guarino L., Jarvis A., et al. Increasing homogeneity in global food supplies and the implications for food security. Proceedings Of The National Academy Of Sciences. 111, 4001–4006 (2014) doi: 10.1073/pnas.1313490111 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2. Ray D., Mueller N., West P. & Foley J. Yield trends are insufficient to double global crop production by 2050. PloS One. 8, e66428 (2013) doi: 10.1371/journal.pone.0066428 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Lutz W., Butz W. & Samir K. World population & human capital in the twenty-first century: An overview. (Oxford University Press,2017) [Google Scholar]
  • 4. Abeysekara W., Siriwardana M. & Meng S. Economic consequences of climate change impacts on the agricultural sector of South Asia: A case study of Sri Lanka. Economic Analysis And Policy. 77 pp. 435–450 (2023) doi: 10.1016/j.eap.2022.12.003 [DOI] [Google Scholar]
  • 5. Bhattarai K. Unemployment–inflation trade-offs in OECD countries. Economic Modelling. 58 pp. 93–103 (2016) doi: 10.1016/j.econmod.2016.05.007 [DOI] [Google Scholar]
  • 6. Atigala P., Maduwanthi T., Gunathilake V., Sathsarani S. & Jayathilaka R. Driving the pulse of the economy or the dilution effect: Inflation impacting economic growth. Plos One. 17, e0273379 (2022) doi: 10.1371/journal.pone.0273379 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. Deger S. & Sen S. Chapter 11 Military expenditure and developing countries. (Elsevier,1995), https://www.sciencedirect.com/science/article/pii/S1574001305800134 [Google Scholar]
  • 8. Dunne J. & Tian N. Military expenditure and economic growth: A survey. The Economics Of Peace And Security Journal. 8 (2013) doi: 10.15355/epsj.8.1.5 [DOI] [Google Scholar]
  • 9. Ram R. Chapter 10 Defense expenditure and economic growth. (Elsevier,1995), https://www.sciencedirect.com/science/article/pii/S1574001305800122 [Google Scholar]
  • 10. Rathnayake N., Miyazaki A., Dang T. & Hoshino Y. Age classification of rice seeds in japan using gradient-boosting and anfis algorithms. Sensors. 23, 2828 (2023) doi: 10.3390/s23052828 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Rathnayake, N., Dang, T., Miyazaki, A. & Hoshino, Y. An efficient approach for age-wise rice seeds classification using SURF-BOF with modified cascaded-ANFIS algorithm. Fifteenth International Conference On Machine Vision (ICMV 2022). 12701 pp. 471-479 (2023)
  • 12. Mulla S., Singh S., Singh K. & Praveen B. Climate change and agriculture: a review of crop models. Global Climate Change And Environmental Policy: Agriculture Perspectives. pp. 423–435 (2020) doi: 10.1007/978-981-13-9570-3_15 [DOI] [Google Scholar]
  • 13. Rosenzweig C. & Parry M. Potential impact of climate change on world food supply. Nature. 367, 133–138 (1994) doi: 10.1038/367133a0 [DOI] [Google Scholar]
  • 14. Lobell D. & Burke M. Why are agricultural impacts of climate change so uncertain? The importance of temperature relative to precipitation. Environmental Research Letters. 3, 034007 (2008) doi: 10.1088/1748-9326/3/3/034007 [DOI] [Google Scholar]
  • 15. Ray D., Ramankutty N., Mueller N., West P. & Foley J. Recent patterns of crop yield growth and stagnation. Nature Communications. 3, 1293 (2012) doi: 10.1038/ncomms2296 [DOI] [PubMed] [Google Scholar]
  • 16. Liakos K., Busato P., Moshou D., Pearson S. & Bochtis D. Machine learning in agriculture: A review. Sensors. 18, 2674 (2018) doi: 10.3390/s18082674 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Krishna, N., Neeraj, P., Surya, B., Prudhvi, B. & Deepthi, V. Machine Learning Algorithms for Crop Yield Prediction in Real-Time Scenarios. 2023 4th International Conference On Signal Processing And Communication (ICSPC). pp. 377-381 (2023)
  • 18. Satpathi A., Setiya P., Das B., Nain A., Jha P., Singh S., et al. Comparative Analysis of Statistical and Machine Learning Techniques for Rice Yield Forecasting for Chhattisgarh, India. Sustainability. 15, 2786 (2023) doi: 10.3390/su15032786 [DOI] [Google Scholar]
  • 19. Vasilyevich K. Machine learning methods in digital agriculture: algorithms and cases. International Journal Of Advanced Studies. 8, 11–26 (2018) doi: 10.12731/2227-930X-2018-1-11-26 [DOI] [Google Scholar]
  • 20. Furbank R. & Tester M. Phenomics–technologies to relieve the phenotyping bottleneck. Trends In Plant Science. 16, 635–644 (2011) doi: 10.1016/j.tplants.2011.09.005 [DOI] [PubMed] [Google Scholar]
  • 21.Sahoo, A. & Ghose, D. Application of hybrid support vector machine model for streamflow prediction in Barak valley, India. IOP Conference Series: Earth And Environmental Science. 1032, 012016 (2022)
  • 22.Moharana, L., Sahoo, A. & Ghose, D. Prediction of rainfall using hybrid SVM-HHO model. IOP Conference Series: Earth And Environmental Science. 1084, 012054 (2022)
  • 23.Ghose, D., Mahakur, V. & Sahoo, A. Monthly runoff prediction by hybrid CNN-LSTM model: A case study. International Conference On Advances In Computing And Data Sciences. pp. 381-392 (2022)
  • 24. McQueen R., Garner S., Nevill-Manning C. & Witten I. Applying machine learning to agricultural data. Computers And Electronics In Agriculture. 12, 275–293 (1995) doi: 10.1016/0168-1699(95)98601-9 [DOI] [Google Scholar]
  • 25. Mirani A., Memon M., Chohan R., Wagan A. & Qabulio M. Machine learning in agriculture: A review. LUME. 10 pp. 5 (2021) [Google Scholar]
  • 26. Porchilambi K. & Sumitra P. Machine learning algorithms for crop yield prediction: A survey. Journal Of Emerging Technologies And Innovative Research. 6 (2019) [Google Scholar]
  • 27.Haque, F., Abdelgawad, A., Yanambaka, V. & Yelamarthi, K. Crop yield prediction using deep neural network. 2020 IEEE 6th World Forum On Internet Of Things (WF-IoT). pp. 1-4 (2020)
  • 28. Jagielska I., Matthews C. & Whitfort T. An investigation into the application of neural networks, fuzzy logic, genetic algorithms, and rough sets to automated knowledge acquisition for classification problems. Neurocomputing. 24, 37–54 (1999) doi: 10.1016/S0925-2312(98)00090-3 [DOI] [Google Scholar]
  • 29. Reghis M. & Roventa E. Classical and Fuzzy Concepts in Mathematical Logic and Applications, Professional Version. (Crc Press,2022) [Google Scholar]
  • 30. Ecer F., Ardabili S., Band S. & Mosavi A. Training multilayer perceptron with genetic algorithms and particle swarm optimization for modeling stock price index prediction. Entropy. 22, 1239 (2020) doi: 10.3390/e22111239 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31. Pant M. & Kumar S. Fuzzy time series forecasting based on hesitant fuzzy sets, particle swarm optimization and support vector machine-based hybrid method. Granular Computing. pp. 1–19 (2022) [Google Scholar]
  • 32. Rathnayake N., Rathnayake U., Dang T. & Hoshino Y. Water level prediction using soft computing techniques: A case study in the Malwathu Oya, Sri Lanka. Plos One. 18, e0282847 (2023) doi: 10.1371/journal.pone.0282847 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33. Rathnayake N., Rathnayake U., Chathuranika I., Dang T. & Hoshino Y. Cascaded-ANFIS to simulate nonlinear rainfall–runoff relationship. Applied Soft Computing. 147 pp. 110722 (2023) doi: 10.1016/j.asoc.2023.110722 [DOI] [Google Scholar]
  • 34. Rathnayake N., Dang T. & Hoshino Y. A novel optimization algorithm: Cascaded adaptive neuro-fuzzy inference system. International Journal Of Fuzzy Systems. 23, 1955–1971 (2021) doi: 10.1007/s40815-021-01076-z [DOI] [Google Scholar]
  • 35. Jang J. ANFIS: adaptive-network-based fuzzy inference system. IEEE Transactions On Systems, Man, And Cybernetics. 23, 665–685 (1993) doi: 10.1109/21.256541 [DOI] [Google Scholar]
  • 36. Jang J., Sun C. & Mizutani E. Neuro-fuzzy and soft computing-a computational approach to learning and machine intelligence [Book Review]. IEEE Transactions On Automatic Control. 42, 1482–1484 (1997) doi: 10.1109/TAC.1997.633847 [DOI] [Google Scholar]
  • 37. Boyacioglu M. & Avci D. An adaptive network-based fuzzy inference system (ANFIS) for the prediction of stock market return: the case of the Istanbul stock exchange. Expert Systems With Applications. 37, 7908–7912 (2010) doi: 10.1016/j.eswa.2010.04.045 [DOI] [Google Scholar]
  • 38. Rathnayake N., Rathnayake U., Chathuranika I., Dang T. & Hoshino Y. Projected Water Levels and Identified Future Floods: A Comparative Analysis for Mahaweli River, Sri Lanka. IEEE Access. 11 pp. 8920–8937 (2023) doi: 10.1109/ACCESS.2023.3238717 [DOI] [Google Scholar]
  • 39.Mampitiya, L., Nalmi, R. & Rathnayake, N. Classification of Human Emotions using Ensemble Classifier by Analysing EEG Signals. 2021 IEEE Third International Conference On Cognitive Machine Intelligence (CogMI). pp. 71-77 (2021)
  • 40. Herath M., Jayathilaka T., Azamathulla H., Mandala V., Rathnayake N. & Rathnayake U. Sensitivity Analysis of Parameters Affecting Wetland Water Levels: A Study of Flood Detention Basin, Colombo, Sri Lanka. Sensors. 23 (2023), https://www.mdpi.com/1424-8220/23/7/3680 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Sandeep Samantaray

13 Mar 2024

PONE-D-24-05247Economic Factors and Rice Production: A Machine Learning Analysis of Correlations and Predictive ModelsPLOS ONE

Dear Dr. Rathnayake,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Apr 27 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Dr. Sandeep Samantaray

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, all author-generated code must be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse.

3. In the online submission form, you indicated that [The data will be available upon a format request to the corresponding author.]. 

All PLOS journals now require all data underlying the findings described in their manuscript to be freely available to other researchers, either 1. In a public repository, 2. Within the manuscript itself, or 3. Uploaded as supplementary information.

This policy applies to all data except where public deposition would breach compliance with the protocol approved by your research ethics board. If your data cannot be made publicly available for ethical or legal reasons (e.g., public availability would compromise patient privacy), please explain your reasons on resubmission and your exemption request will be escalated for approval. 

Additional Editor Comments:

1. Why author took the data set up to 2020; is there any scientific reason- suggestion to add data set upto 2023.

2. Eq 2-5: author must add reference for cite the formulae.

3. Author must add parameter table and pseudo code all proposed algorithm

4. Author must add scatter plot, time series plot, histogram plot for better accuracy of result.

5. Author must add some more discussion in terms of comparison with other study and advantages of proposed model, future scope of the research.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: N/A

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: No

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Title: Consider a more specific title that reflects the main findings of your study. For example, "Impact of Economic Indicators on Rice Production: A Machine Learning Approach in Sri Lanka."

Abstract:

Consider including a sentence on the practical implications of the study's findings.

Introduction:

Clearly state the research questions or hypotheses.

Methodology:

Describe the preprocessing steps in more detail, including how missing or inconsistent values were handled.

Results and Discussion:

Discuss the implications of the results in the context of existing literature and theory.

Address any limitations of the study and suggest avenues for future research.

Conclusion:

Summarize the key findings of the study and their implications for agricultural policy and practice.

Highlight the contribution of the study to the existing literature and any practical applications of the findings.

General:

Ensure consistency in terminology and formatting throughout the manuscript.

Reviewer #2: 1. The novelty and contribution of the paper should be clearly stated.

2. What is the contribution of this finding in addressing the challenges that were identified in the introduction?

3. What authors were motivated to conduct this type of study, and why did they select the Sri Lanka as a case study?

4. Plenty of work has already proved the potential of various ML models across the world. What is different being using the applied models in this study?

5. What were the used input parameters?

6. Provide the range of statistical evaluating standards and support them by citing suitable references. The authors can cite https://doi.org/10.1007/978-3-031-12641-3_31; Doi:10.4704/nq.2022.20.14. NQ880113; https://doi.org/10.1063/5.0161098; https://doi.org/10.1063/5.0132387; DOI: 10.1088/1755-1315/1084/1/012054; DOI: 10.1088/1755-1315/1032/1/012016

7. Likewise, what is the transferability of such results to other locations in terms of impact or usefulness?

8. Include all the figures and tables at their respective place in the text.

9. Authors need to consult with a native English speaker for language improvement.

10. Reveal more practical debate in the discussion section and explain the usefulness and practical implications of this work.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2024 Jun 21;19(6):e0303883. doi: 10.1371/journal.pone.0303883.r002

Author response to Decision Letter 0


20 Apr 2024

Responses for Editor's and Reviewer's comments are attached as separate PDF files named

1. Response_to_Editor_Comments.pdf (Attached as the Cover letter)

2. Response_to_Reviewer_1_Comments.pdf

3. Response_to_Reviewer_2_Comments.pdf

Thank you.

Attachment

Submitted filename: Response_to_Reviewer_2_Comments.pdf

pone.0303883.s001.pdf (61.2KB, pdf)

Decision Letter 1

Sandeep Samantaray

2 May 2024

Impact of Economic Indicators on Rice Production: A Machine Learning Approach in Sri Lanka.

PONE-D-24-05247R1

Dear Dr. %Namal Rathnayake%,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Dr. Sandeep Samantaray

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Acceptance letter

Sandeep Samantaray

17 May 2024

PONE-D-24-05247R1

PLOS ONE

Dear Dr. Rathnayake,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

If revisions are needed, the production department will contact you directly to resolve them. If no revisions are needed, you will receive an email when the publication date has been set. At this time, we do not offer pre-publication proofs to authors during production of the accepted work. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few weeks to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Sandeep Samantaray

Academic Editor

PLOS ONE


Articles from PLOS ONE are provided here courtesy of PLOS

RESOURCES