SHARP ORACLE INEQUALITIES FOR LEAST SQUARES ESTIMATORS IN SHAPE RESTRICTED REGRESSION

Citation
Pierre C. Bellec, SHARP ORACLE INEQUALITIES FOR LEAST SQUARES ESTIMATORS IN SHAPE RESTRICTED REGRESSION, Annals of statistics , 46(2), 2018, pp. 745-780
Journal title
ISSN journal
00905364
Volume
46
Issue
2
Year of publication
2018
Pages
745 - 780
Database
ACNP
SICI code
Abstract
The performance of Least Squares (LS) estimators is studied in shape-constrained regression models under Gaussian and sub-Gaussian noise. General bounds on the performance of LS estimators over closed convex sets are provided. These results have the form of sharp oracle inequalities that account for the model misspecification error. In the presence of misspecification, these bounds imply that the LS estimator estimates the projection of the true parameter at the same rate as in the well-specified case. In isotonic and unimodal regression, the LS estimator achieves the nonparametric rate n.2/3 as well as a parametric rate of order k/n up to logarithmic factors, where k is the number of constant pieces of the true parameter. In univariate convex regression, the LS estimator satisfies an adaptive risk bound of order q/n up to logarithmic factors, where q is the number of affine pieces of the true regression function. This adaptive risk bound holds for any collection of design points. While Guntuboyina and Sen [Probab. Theory Related Fields 163 (2015) 379.411] established that the nonparametric rate of convex regression is of order n.4/5 for equispaced design points, we show that the nonparametric rate of convex regression can be as slow as n.2/3 for some worst-case design points. This phenomenon can be explained as follows: Although convexity brings more structure than unimodality, for some worstcase design points this extra structure is uninformative and the nonparametric rates of unimodal regression and convex regression are both n.2/3. Higher order cones, such as the cone of .-monotone sequences, are also studied.