Colloquium: "Statistical learning for time series"

Likai Chen, University of Chicago

Abstract: In statistical learning theory, researchers primarily deal with independent data and there is a huge literature. In comparison, it has been much less investigated for time-dependent data, which are commonly encountered in economics, engineering, finance, geography, physics and other fields. In this talk, we focus on concentration inequalities for suprema of empirical processes which plays a fundamental role in the statistical learning theory. We derive a Gaussian approximation and an upper bound for the tail probability of the suprema under conditions on the size of the function class, the sample size, temporal dependence and the moment conditions of the underlying time series. Due to the dependence and heavy-tailness, our tail probability bound is substantially different from those classical exponential bounds obtained under the independence assumption in that it involves an extra polynomial decaying term. We allow both short- and long-range dependent processes, where the long-range dependence case has never been previously explored. We showed our tail probability inequality is sharp up to a multiplicative constant.  These bounds work as theoretical guarantees for statistical learning applications under dependence.

Host: Jose Figueroa-Lopez

Tea @3:45pm in room 200