DADE-DQN: Dual Action and Dual Environment Deep Q-Network for Enhancing Stock Trading Strategy
-
Published:2023-08-22
Issue:17
Volume:11
Page:3626
-
ISSN:2227-7390
-
Container-title:Mathematics
-
language:en
-
Short-container-title:Mathematics
Author:
Huang Yuling1ORCID, Lu Xiaoping1ORCID, Zhou Chujin1, Song Yunlin2
Affiliation:
1. School of Computer Science and Engineering, Macau University of Science and Technology, Taipa, Macao, China 2. Department of Engineering Science, Faculty of Innovation Engineering, Macau University of Science and Technology, Taipa, Macao, China
Abstract
Deep reinforcement learning (DRL) has attracted strong interest since AlphaGo beat human professionals, and its applications in stock trading are widespread. In this paper, an enhanced stock trading strategy called Dual Action and Dual Environment Deep Q-Network (DADE-DQN) for profit and risk reduction is proposed. Our approach incorporates several key highlights. First, to achieve a better balance between exploration and exploitation, a dual-action selection and dual-environment mechanism are incorporated into our DQN framework. Second, our approach optimizes the utilization of storage transitions by utilizing independent replay memories and performing dual mini-batch updates, leading to faster convergence and more efficient learning. Third, a novel deep network structure that incorporates Long Short-Term Memory (LSTM) and attention mechanisms is introduced, thereby improving the network’s ability to capture essential features and patterns. In addition, an innovative feature selection method is presented to efficiently enhance the input data by utilizing mutual information to identify and eliminate irrelevant features. Evaluation on six datasets shows that our DADE-DQN algorithm outperforms multiple DRL-based strategies (TDQN, DQN-Pattern, DQN-Vanilla) and traditional strategies (B&H, S&H, MR, TF). For example, on the KS11 dataset, the DADE-DQN strategy has achieved an impressive cumulative return of 79.43% and a Sharpe ratio of 2.21, outperforming all other methods. These experimental results demonstrate the performance of our approach in enhancing stock trading strategies.
Funder
Faculty Research Grants, Macau University of Science and Technology Science and Technology Development Fund, Macau SAR
Subject
General Mathematics,Engineering (miscellaneous),Computer Science (miscellaneous)
Reference62 articles.
1. Hamilton, J.D. (2020). Time Series Analysis, Princeton University Press. 2. Recent Advances in Reinforcement Learning in Finance;Hambly;Math. Financ.,2021 3. Mastering the game of Go without human knowledge;Silver;Nature,2017 4. Grandmaster level in StarCraft II using multi-agent reinforcement learning;Vinyals;Nature,2019 5. Berner, C., Brockman, G., Chan, B., Cheung, V., Debiak, P., Dennison, C., Farhi, D., Fischer, Q., Hashme, S., and Hesse, C. (2019). Dota 2 with Large Scale Deep Reinforcement Learning. arXiv.
Cited by
3 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|