Modular reinforcement learning for dynamic portfolio optimization in the KOSPI market

2021 ◽  
Vol 32 (1) ◽  
pp. 213-226
Author(s):  
Taeyoon Kim ◽  
Bonggyun Ko
2015 ◽  
Vol 243 (3) ◽  
pp. 921-931 ◽  
Author(s):  
Jan Palczewski ◽  
Rolf Poulsen ◽  
Klaus Reiner Schenk-Hoppé ◽  
Huamao Wang

2020 ◽  
Vol 32 (23) ◽  
pp. 17229-17244
Author(s):  
Giorgio Lucarelli ◽  
Matteo Borrotti

AbstractDeep reinforcement learning is gaining popularity in many different fields. An interesting sector is related to the definition of dynamic decision-making systems. A possible example is dynamic portfolio optimization, where an agent has to continuously reallocate an amount of fund into a number of different financial assets with the final goal of maximizing return and minimizing risk. In this work, a novel deep Q-learning portfolio management framework is proposed. The framework is composed by two elements: a set of local agents that learn assets behaviours and a global agent that describes the global reward function. The framework is tested on a crypto portfolio composed by four cryptocurrencies. Based on our results, the deep reinforcement portfolio management framework has proven to be a promising approach for dynamic portfolio optimization.


2021 ◽  
Vol 8 (3-4) ◽  
pp. 101-125
Author(s):  
Babak Mahdavi-Damghani ◽  
Konul Mustafayeva ◽  
Cristin Buescu ◽  
Stephen Roberts

With the recent rise of Machine Learning (ML) as a candidate to partially replace classic Financial Mathematics (FM) methodologies, we investigate the performances of both in solving the problem of dynamic portfolio optimization in continuous-time, finite-horizon setting for a portfolio of two assets that are intertwined. In the Financial Mathematics approach we model the asset prices not via the common approaches used in pairs trading such as a high correlation or cointegration, but with the cointelation model in Mahdavi-Damghani (2013) that aims to reconcile both short-term risk and long-term equilibrium. We maximize the overall P&L with Financial Mathematics approach that dynamically switches between a mean-variance optimal strategy and a power utility maximizing strategy. We use a stochastic control formulation of the problem of power utility maximization and solve numerically the resulting HJB equation with the Deep Galerkin method introduced in Sirignano and Spiliopoulos (2018). We turn to Machine Learning for the same P&L maximization problem and use clustering analysis to devise bands, combined with in-band optimization. Although this approach is model agnostic, results obtained with data simulated from the same cointelation model gives a slight competitive advantage to the ML over the FM methodology1.


2018 ◽  
Vol 19 (3) ◽  
pp. 519-532 ◽  
Author(s):  
Rongju Zhang ◽  
Nicolas Langrené ◽  
Yu Tian ◽  
Zili Zhu ◽  
Fima Klebaner ◽  
...  

2005 ◽  
Vol 23 (3) ◽  
pp. 579-594 ◽  
Author(s):  
A. Gabih ◽  
W. Grecksch ◽  
R. Wunderlich

Sign in / Sign up

Export Citation Format

Share Document