
Tatyana Krivobokova
Kernel partial least squares for stationary data
We consider the kernel partial least squares algorithm for the solution of nonparametric regression problems when the data are stationary time series. Probabilistic convergence rates of the kernel partial least squares estimator to the true regression function are established under a source condition. The impact of long range dependence in the data is studied both theoretically and in simulations. A real data example on protein dynamics illustrates the approach.
This is the joint work with Marco Singer and Axel Munk.
Download the slides

Botond Szabó
An asymptotic analysis of nonparametric divideandconquer methods
In the recent years in certain applications datasets have become so large
that it becomes unfeasible, or computationally undesirable, to carry out the
analysis on a single machine. This gave rise to divideandconquer algorithms
where the data is distributed over several "local" machines and the computations
are done on these machines parallel to each other. Then the outcome of
the local computations are somehow aggregated to a global result in a central
machine.
Over the years various divideandconquer algorithms were proposed, many
of them with limited theoretical underpinning. First we compare the theoretical
properties of a (not complete) list of proposed methods on the benchmark
nonparametric signalinwhitenoise model. Most of the investigated algorithms
use information on aspects of the underlying true signal (for instance
regularity), which is usually not available in practice. A central question is
whether one can tune the algorithms in a datadriven way, without using any
additional knowledge about the signal. We show that (a list of) standard
datadriven techniques (both Bayesian and frequentist) can not recover the
underlying signal with the minimax rate. This, however, does not imply the
nonexistence of an adaptive distributed method.
To address the theoretical limitations of datadriven divideandconquer
algorithms we consider a setting where the amount of information sent between
the local and central machines is expensive and limited. We show that it is
not possible to construct datadriven methods which adapt to the unknown
regularity of the underlying signal and at the same time communicates the
optimal amount of information between the machines.
This is a joint work with Harry van Zanten.
Download the slides
