An Ensemble Method of Deep Reinforcement Learning for Automated Cryptocurrency Trading
ArXiv ID: 2309.00626 “View on arXiv”
Authors: Unknown
Abstract
We propose an ensemble method to improve the generalization performance of trading strategies trained by deep reinforcement learning algorithms in a highly stochastic environment of intraday cryptocurrency portfolio trading. We adopt a model selection method that evaluates on multiple validation periods, and propose a novel mixture distribution policy to effectively ensemble the selected models. We provide a distributional view of the out-of-sample performance on granular test periods to demonstrate the robustness of the strategies in evolving market conditions, and retrain the models periodically to address non-stationarity of financial data. Our proposed ensemble method improves the out-of-sample performance compared with the benchmarks of a deep reinforcement learning strategy and a passive investment strategy.
Keywords: Deep Reinforcement Learning, Ensemble Method, Cryptocurrency Trading, Non-Stationarity, Portfolio Trading, Cryptocurrency
Complexity vs Empirical Score
- Math Complexity: 8.5/10
- Empirical Rigor: 7.0/10
- Quadrant: Holy Grail
- Why: The paper employs advanced mathematical concepts like MDPs, policy ensembling, and distributional views of performance, but also includes significant empirical elements like a 4-year backtest, portfolio trading, and performance metrics.
flowchart TD
A["Research Goal: Improve DRL Strategy Generalization for Crypto Trading"] --> B["Data: Intraday Cryptocurrency Prices & Volumes"]
B --> C["Methodology: Ensemble DRL with Model Selection"]
C --> D{"Key Component: Mixture Distribution Policy"}
D --> E["Retraining: Address Data Non-Stationarity"]
E --> F["Outcome: Out-of-Sample Evaluation"]
F --> G["Result: Superior Performance vs. DRL & Passive Benchmarks"]