scholarly journals Portfolio optimization for cointelated pairs: SDEs vs Machine learning

2021 ◽  
Vol 8 (3-4) ◽  
pp. 101-125
Author(s):  
Babak Mahdavi-Damghani ◽  
Konul Mustafayeva ◽  
Cristin Buescu ◽  
Stephen Roberts

With the recent rise of Machine Learning (ML) as a candidate to partially replace classic Financial Mathematics (FM) methodologies, we investigate the performances of both in solving the problem of dynamic portfolio optimization in continuous-time, finite-horizon setting for a portfolio of two assets that are intertwined. In the Financial Mathematics approach we model the asset prices not via the common approaches used in pairs trading such as a high correlation or cointegration, but with the cointelation model in Mahdavi-Damghani (2013) that aims to reconcile both short-term risk and long-term equilibrium. We maximize the overall P&L with Financial Mathematics approach that dynamically switches between a mean-variance optimal strategy and a power utility maximizing strategy. We use a stochastic control formulation of the problem of power utility maximization and solve numerically the resulting HJB equation with the Deep Galerkin method introduced in Sirignano and Spiliopoulos (2018). We turn to Machine Learning for the same P&L maximization problem and use clustering analysis to devise bands, combined with in-band optimization. Although this approach is model agnostic, results obtained with data simulated from the same cointelation model gives a slight competitive advantage to the ML over the FM methodology1.

2007 ◽  
Vol 10 (02) ◽  
pp. 203-233 ◽  
Author(s):  
FLORIAN HERZOG ◽  
GABRIEL DONDI ◽  
HANS P. GEERING

This paper proposes a solution method for the discrete-time long-term dynamic portfolio optimization problem with state and asset allocation constraints. We use the ideas of Model Predictive Control (MPC) to solve the constrained stochastic control problem. MPC is a solution technique which was developed to solve constrained optimal control problems for deterministic control applications. MPC solves the optimal control problem with a receding horizon where a series of consecutive open-loop optimal control problems is solved. The aim of this paper is to develop an MPC approach to the problem of long-term portfolio optimization when the expected returns of the risky assets are modeled using a factor model based on stochastic Gaussian processes. We prove that MPC is a suboptimal control strategy for stochastic systems which uses the new information advantageously and thus is better than the pure optimal open-loop control. For the open-loop optimal control optimization, we derive the conditional portfolio distribution and the corresponding conditional portfolio mean and variance. The mean and the variance depend on future decision about the asset allocation. For the dynamic portfolio optimization problem, we consider constraints on the asset allocation as well as probabilistic constraints on the attainable values of the portfolio wealth. We discuss two different objectives, a classical mean–variance objective and the objective to maximize the probability of exceeding a predetermined value of the portfolio. The dynamic portfolio optimization problem is stated, and the solution via MPC is explained in detail. The results are then illustrated in a case study.


2015 ◽  
Vol 243 (3) ◽  
pp. 921-931 ◽  
Author(s):  
Jan Palczewski ◽  
Rolf Poulsen ◽  
Klaus Reiner Schenk-Hoppé ◽  
Huamao Wang

2020 ◽  
Vol 32 (23) ◽  
pp. 17229-17244
Author(s):  
Giorgio Lucarelli ◽  
Matteo Borrotti

AbstractDeep reinforcement learning is gaining popularity in many different fields. An interesting sector is related to the definition of dynamic decision-making systems. A possible example is dynamic portfolio optimization, where an agent has to continuously reallocate an amount of fund into a number of different financial assets with the final goal of maximizing return and minimizing risk. In this work, a novel deep Q-learning portfolio management framework is proposed. The framework is composed by two elements: a set of local agents that learn assets behaviours and a global agent that describes the global reward function. The framework is tested on a crypto portfolio composed by four cryptocurrencies. Based on our results, the deep reinforcement portfolio management framework has proven to be a promising approach for dynamic portfolio optimization.


2018 ◽  
Vol 21 (07) ◽  
pp. 1850046 ◽  
Author(s):  
SÜHAN ALTAY ◽  
KATIA COLANERI ◽  
ZEHRA EKSI

In this work, we study a dynamic portfolio optimization problem related to pairs trading, which is an investment strategy that matches a long position in one security with a short position in another security with similar characteristics. The relationship between pairs, called a spread, is modeled by a Gaussian mean-reverting process whose drift rate is modulated by an unobservable continuous-time, finite-state Markov chain. Using the classical stochastic filtering theory, we reduce this problem with partial information to an equivalent one with full information and solve it for the logarithmic utility function, where the terminal wealth is penalized by the riskiness of the portfolio according to the realized volatility of the wealth process. We characterize optimal dollar-neutral strategies as well as optimal value functions under full and partial information and show that the certainty equivalence principle holds for the optimal portfolio strategy. Finally, we provide a numerical analysis for a toy example with a two-state Markov chain.


2018 ◽  
Vol 19 (3) ◽  
pp. 519-532 ◽  
Author(s):  
Rongju Zhang ◽  
Nicolas Langrené ◽  
Yu Tian ◽  
Zili Zhu ◽  
Fima Klebaner ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document