Methods for Computing Numerical Standard Errors: Review and Application to Value-at-Risk Estimation


Numerical standard error (NSE) is an estimate of the standard deviation of a simulation result if the simulation experiment were to be repeated many times. We review standard methods for computing NSE, and perform a Monte Carlo experiments to compare their performance in the case of high/extreme autocorrelation. In particular, we propose an application to risk management where we assess the precision of the Value–at–Risk measure when the underlying risk model is estimated by simulation–based methods. Overall, HAC estimators with prewhitening perform best in the presence of large/extreme autocorrelation.

Journal of Time Series Econometrics, Vol. 10, No. 2, pp.1-9