|Title:||Consistency and robustness of kernel based regression|
|Abstract:||We investigate properties of kernel based regression (KBR) methods which are inspired by the convex risk minimization method of support vector machines. We first describe the relation between the used loss function of the KBR method and the tail of the response variable Y . We then establish a consistency result for KBR and give assumptions for the existence of the influence function. In particular, our results allow to choose the loss function and the kernel to obtain computational tractable and consistent KBR methods having bounded influence functions. Furthermore, bounds for the sensitivity curve which is a finite sample version of the influence function are developed, and some numerical experiments are discussed.|
|Appears in Collections:||Sonderforschungsbereich (SFB) 475|
This item is protected by original copyright
All resources in the repository are protected by copyright.