TY - JOUR
T1 - Prioritizing of volatility models: a computational analysis using data envelopment analysis
AU - Petridis, Konstantinos
AU - Petridis, Nikolaos E.
AU - Emrouznejad, Ali
AU - Ben Abdelaziz, Fouad
N1 - This is the peer reviewed version of the following article: Petridis, K., Petridis, N.E., Emrouznejad, A. and Ben Abdelaziz, F. (2021), Prioritizing of volatility models: a computational analysis using data envelopment analysis. Intl. Trans. in Op. Res.., which has been published in final form at https://doi.org/10.1111/itor.13028. This article may be used for non-commercial purposes in accordance With Wiley Terms and Conditions for self-archiving.
PY - 2021/7/19
Y1 - 2021/7/19
N2 - Economic crisis and uncertainty in global status quo affect stock markets around the world. This fact imposes improvement in the development of volatility models. However, the comparison among volatility models cannot be made based on a single-error measure as a model can perform better in one-error measure and worst in another. In this paper, we propose a two-stage approach for prioritizing volatility models, where in the first stage we develop a novel slack-based data envelopment analysis to rank volatility models. The robustness of the proposed approach has also been investigated using cluster analysis. In the second-stage analysis, it is investigated whether the efficiency scores depend on model characteristics. These attributes concern the time needed in order to estimate the model, the value of Akaike Information Criterion, the number of models' significant parameters, groups and bias terms, and the error sum of squares (ESS). Further, dummy variables have been introduced to the regression model in order to find whether the employed model includes an in-mean effect, whether the assumed distribution is skewed, and whether the employed model belongs to the generalized autoregressive conditional heteroskedasticity (GARCH) family. The main findings of this research show that the number of models' statistically significant coefficients, ESS, and in-mean effects tend to increase the efficiency scores, while time elapsed, the number of statistically significant bias terms, and skewed error distributions tend to decrease the efficiency score.
AB - Economic crisis and uncertainty in global status quo affect stock markets around the world. This fact imposes improvement in the development of volatility models. However, the comparison among volatility models cannot be made based on a single-error measure as a model can perform better in one-error measure and worst in another. In this paper, we propose a two-stage approach for prioritizing volatility models, where in the first stage we develop a novel slack-based data envelopment analysis to rank volatility models. The robustness of the proposed approach has also been investigated using cluster analysis. In the second-stage analysis, it is investigated whether the efficiency scores depend on model characteristics. These attributes concern the time needed in order to estimate the model, the value of Akaike Information Criterion, the number of models' significant parameters, groups and bias terms, and the error sum of squares (ESS). Further, dummy variables have been introduced to the regression model in order to find whether the employed model includes an in-mean effect, whether the assumed distribution is skewed, and whether the employed model belongs to the generalized autoregressive conditional heteroskedasticity (GARCH) family. The main findings of this research show that the number of models' statistically significant coefficients, ESS, and in-mean effects tend to increase the efficiency scores, while time elapsed, the number of statistically significant bias terms, and skewed error distributions tend to decrease the efficiency score.
UR - https://onlinelibrary.wiley.com/doi/10.1111/itor.13028
U2 - 10.1111/itor.13028
DO - 10.1111/itor.13028
M3 - Article
JO - International Transactions in Operational Research
JF - International Transactions in Operational Research
SN - 0969-6016
ER -