scholarly journals Selected GARCH‑type Models in the Metals Market – Backtesting of Value‑at‑Risk

2018 ◽  
Vol 5 (331) ◽  
pp. 185-203
Author(s):  
Dominik Krężołek

 Risk analysis in the financial market requires the correct evaluation of volatility in terms of both prices and asset returns. Disturbances in quality of information, the economic and political situation and investment speculations cause incredible difficulties in accurate forecasting. From the investor’s point of view, the key issue is to minimise the risk of huge losses. This article presents the results of using some selected GARCH‑type models, ARMA‑GARCH and ARMA‑APARCH, in evaluating volatility of asset returns in the metals market. To assess the level of risk, the Value‑at‑Risk measure is used. The comparison between real and estimated losses (in terms of VaR) is made using the backtesting procedure. 

2018 ◽  
Vol 35 (1-2) ◽  
pp. 73-87 ◽  
Author(s):  
Sebastian Geissel ◽  
Jörn Sass ◽  
Frank Thomas Seifried

AbstractThis paper introduces optimal expected utility (OEU) risk measures, investigates their main properties and puts them in perspective to alternative risk measures and notions of certainty equivalents. By taking the investor’s point of view, OEU maximizes the sum of capital available today and the certainty equivalent of capital in the future. To the best of our knowledge, OEU is the only existing utility-based risk measure that is (non-trivial and) coherent if the utility functionuhas constant relative risk aversion. We present several different risk measures that can be derived with special choices ofuand illustrate that OEU is more sensitive than value at risk and average value at risk with respect to changes of the probability of a financial loss.


2021 ◽  
Vol 14 (11) ◽  
pp. 540
Author(s):  
Eyden Samunderu ◽  
Yvonne T. Murahwa

Developments in the world of finance have led the authors to assess the adequacy of using the normal distribution assumptions alone in measuring risk. Cushioning against risk has always created a plethora of complexities and challenges; hence, this paper attempts to analyse statistical properties of various risk measures in a not normal distribution and provide a financial blueprint on how to manage risk. It is assumed that using old assumptions of normality alone in a distribution is not as accurate, which has led to the use of models that do not give accurate risk measures. Our empirical design of study firstly examined an overview of the use of returns in measuring risk and an assessment of the current financial environment. As an alternative to conventional measures, our paper employs a mosaic of risk techniques in order to ascertain the fact that there is no one universal risk measure. The next step involved looking at the current risk proxy measures adopted, such as the Gaussian-based, value at risk (VaR) measure. Furthermore, the authors analysed multiple alternative approaches that do not take into account the normality assumption, such as other variations of VaR, as well as econometric models that can be used in risk measurement and forecasting. Value at risk (VaR) is a widely used measure of financial risk, which provides a way of quantifying and managing the risk of a portfolio. Arguably, VaR represents the most important tool for evaluating market risk as one of the several threats to the global financial system. Upon carrying out an extensive literature review, a data set was applied which was composed of three main asset classes: bonds, equities and hedge funds. The first part was to determine to what extent returns are not normally distributed. After testing the hypothesis, it was found that the majority of returns are not normally distributed but instead exhibit skewness and kurtosis greater or less than three. The study then applied various VaR methods to measure risk in order to determine the most efficient ones. Different timelines were used to carry out stressed value at risks, and it was seen that during periods of crisis, the volatility of asset returns was higher. The other steps that followed examined the relationship of the variables, correlation tests and time series analysis conducted and led to the forecasting of the returns. It was noted that these methods could not be used in isolation. We adopted the use of a mosaic of all the methods from the VaR measures, which included studying the behaviour and relation of assets with each other. Furthermore, we also examined the environment as a whole, then applied forecasting models to accurately value returns; this gave a much more accurate and relevant risk measure as compared to the initial assumption of normality.


Author(s):  
Tomáš Konderla ◽  
Václav Klepáč

The article points out the possibilities of using Hidden Markov model (abbrev. HMM) for estimation of Value at Risk metrics (abbrev. VaR) in sample. For the illustration we use data of the company listed on Prague Stock Exchange in range from January 2011 to June 2016. HMM approach allows us to classify time series into different states based on their development characteristic. Due to a deeper shortage of existing domestic results or comparison studies with advanced volatility governed VaR forecasts we tested HMM with univariate ARMA‑GARCH model based VaR estimates. The common testing via Kupiec and Christoffersen procedures offer generalization that HMM model performs better that volatility based VaR estimation technique in terms of accuracy, even with the simpler HMM with normal‑mixture distribution against previously used GARCH with many types of non‑normal innovations.


2019 ◽  
pp. 28-55
Author(s):  
Hyun Song Shin

An example of a hedge fund illustrates a long-short strategy that maximises expected returns subject to a Value-at-Risk strategy. Balance sheet capacity depends on the measured volatility of asset returns and the book equity of the long-short hedge fund. The principles are illustrated by the case of Long Term Capital Management (LTCM).


Risks ◽  
2020 ◽  
Vol 8 (3) ◽  
pp. 76
Author(s):  
Saswat Patra ◽  
Malay Bhattacharyya

This paper investigates the risk exposure for options and proposes MaxVaR as an alternative risk measure which captures the risk better than Value-at-Risk especially. While VaR is a measure of end-of-horizon risk, MaxVaR captures the interim risk exposure of a position or a portfolio. MaxVaR is a more stringent risk measure as it assesses the risk during the risk horizon. For a 30-day maturity option, we find that MaxVaR can be 40% higher than VaR at a 5% significance level. It highlights the importance of MaxVaR as a risk measure and shows that the risk is vastly underestimated when VaR is used as the measure for risk. The sensitivity of MaxVaR with respect to option characteristics like moneyness, time to maturity and risk horizons at different significance levels are observed. Further, interestingly enough we find that the MaxVar to VaR ratio is higher for stocks than the options and we can surmise that stock returns are more volatile than options. For robustness, the study is carried out under different distributional assumptions on residuals and for different stock index options.


2015 ◽  
Vol 4 (4) ◽  
pp. 188
Author(s):  
HERLINA HIDAYATI ◽  
KOMANG DHARMAWAN ◽  
I WAYAN SUMARJAYA

Copula is already widely used in financial assets, especially in risk management. It is due to the ability of copula, to capture the nonlinear dependence structure on multivariate assets. In addition, using copula function doesn’t require the assumption of normal distribution. There fore it is suitable to be applied to financial data. To manage a risk the necessary measurement tools can help mitigate the risks. One measure that can be used to measure risk is Value at Risk (VaR). Although VaR is very popular, it has several weaknesses. To overcome the weakness in VaR, an alternative risk measure called CVaR can be used. The porpose of this study is to estimate CVaR using Gaussian copula. The data we used are the closing price of Facebook and Twitter stocks. The results from the calculation using 90%  confidence level showed that the risk that may be experienced is at 4,7%, for 95% confidence level it is at 6,1%, and for 99% confidence level it is at 10,6%.


2019 ◽  
Vol 8 (1) ◽  
pp. 15
Author(s):  
NI WAYAN UCHI YUSHI ARI SUDINA ◽  
KOMANG DHARMAWAN ◽  
I WAYAN SUMARJAYA

Conditional value at risk (CVaR) is widely used in risk measure that takes into account losses exceeding the value at risk level. The aim of this research is to compare the performance of the EVT-GJR-vine copula method and EVT-GARCH-vine copula method in estimating CVaR of the portfolio using backtesting. Based on the backtesting results, it was found that the EVT-GJR-vine copula method have better performance when compared to the EVT-GARCH-vine copula method in estimating the CVaR value of the portfolio. This can be seen from the statistical values ??, and  of EVT-GJR-vine copula method which is generally smaller than the statistical values , and of the EVT-GARCH-vine copula method.


2019 ◽  
Vol 181 (2) ◽  
pp. 473-507 ◽  
Author(s):  
E. Ruben van Beesten ◽  
Ward Romeijnders

Abstract In traditional two-stage mixed-integer recourse models, the expected value of the total costs is minimized. In order to address risk-averse attitudes of decision makers, we consider a weighted mean-risk objective instead. Conditional value-at-risk is used as our risk measure. Integrality conditions on decision variables make the model non-convex and hence, hard to solve. To tackle this problem, we derive convex approximation models and corresponding error bounds, that depend on the total variations of the density functions of the random right-hand side variables in the model. We show that the error bounds converge to zero if these total variations go to zero. In addition, for the special cases of totally unimodular and simple integer recourse models we derive sharper error bounds.


Sign in / Sign up

Export Citation Format

Share Document