Model-based Deep Reinforcement Learning for Dynamic Portfolio Optimization
Yu, Pengqian, Lee, Joon Sern, Kulyatin, Ilya, Shi, Zekun, Dasgupta, Sakyasingha
Dynamic portfolio optimization is the process of sequentially allocating wealth to a collection of assets in some consecutive trading periods, based on investors' return-risk profile. Automating this process with machine learning remains a challenging problem. Here, we design a deep reinforcement learning (RL) architecture with an autonomous trading agent such that, investment decisions and actions are made periodically, based on a global objective, with autonomy. In particular, without relying on a purely model-free RL agent, we train our trading agent using a novel RL architecture consisting of an infused prediction module (IPM), a generative adversarial data augmentation module (DAM) and a behavior cloning module (BCM). Our model-based approach works with both on-policy or off-policy RL algorithms. We further design the back-testing and execution engine which interact with the RL agent in real time. Using historical {\em real} financial market data, we simulate trading with practical constraints, and demonstrate that our proposed model is robust, profitable and risk-sensitive, as compared to baseline trading strategies and model-free RL agents from prior work.
Jan-24-2019
- Country:
- Asia (0.28)
- Europe (0.46)
- North America > United States
- Massachusetts > Middlesex County (0.14)
- Genre:
- Research Report > New Finding (0.92)
- Industry:
- Banking & Finance > Trading (1.00)
- Energy > Oil & Gas
- Upstream (0.85)
- Technology: