Table 9.
DRL in economics
Article | Aim of study | Specific approach | Benchmark methods for comparison | Superiority of the proposed method |
---|---|---|---|---|
(Chakole & Kurhekar, 2020) | Make trading decisions | Deep Q-learning | Decision Tree strategy, Buy-and-Hold strategy | Outperforms in terms of some economic indicators: Accumulated Return, Maximum Drawdown, Average daily return, average annual return, Skewness, Kurtosis, Sharpe ratio, and Standard Deviation |
(Zhou et al., 2020b) | Derive optimal power flow | DRL, PPO with IL | IL, PPO | Perform better in accuracy and running time |
(Qiu et al., 2020) | Pricing electric vehicles | PDDPG | Q-learning, DQN, DDPG | Better performance in standard deviation, learning pace, flexibility and computational time |
(Sattarov et al., 2020) | Recommend cryptocurrency trading points | Deep Neural Model of DRL |
Double cross strategy, swing trading, scalping trading |
Best performance in number of actions and quality of Trading |
(Uddin et al., 2020) | Estimate impact of COVID-19 on the spread of the infection, personal satisfaction or quality of life, resource use and economy | DQN, DDPG | Random, Q-Learning, SARSA | Perform better in terms of best rewards and best policy |
Note: IL (Imitation Learning), PDDPG (Prioritized Deep Deterministic Policy Gradient), DQN (Deep Q Network), DDPG (Deep Deterministic Policy Gradient), SARSA (State-Action-Reward-State-Action).