Least Tail-Trimmed Squares for Infinite Variance Autoregressions

26 Pages Posted: 15 May 2012

See all articles by Jonathan B. Hill

Jonathan B. Hill

University of North Carolina (UNC) at Chapel Hill – Department of Economics

Date Written: May 14, 2012

Abstract

We develop a robust least squares estimator for autoregressions with possibly heavy tailed errors. Robustness to heavy tails is ensured by negligibly trimming the squared error according to extreme values of the error and regressor. Tail-trimming ensures asymptotic normality and super-root(n)-convergence with a rate comparable to the highest achieved amongst M-estimators for stationary data. Moreover, tail-trimming ensures robustness to heavy tails in both small and large samples. By comparison, existing robust estimators are not as robust in small samples and have a slower rate of convergence when the variance is infinite, or are not asymptotically normal. We present a consistent estimator of the covariance matrix and treat classic inference without knowledge of the rate of convergence. A simulation study demonstrates the sharpness and approximate normality of the estimator, and we apply the estimator to financial returns data.

Suggested Citation

Hill, Jonathan B., Least Tail-Trimmed Squares for Infinite Variance Autoregressions (May 14, 2012). Available at SSRN: https://ssrn.com/abstract=2058201 or http://dx.doi.org/10.2139/ssrn.2058201

Jonathan B. Hill (Contact Author)

University of North Carolina (UNC) at Chapel Hill – Department of Economics ( email )

102 Ridge Road
Chapel Hill, NC NC 27514
United States