Gatsby Computational Neuroscience Unit


Ingo Steinwart


Wednesday 28th February 2018


Time: 4.00pm


Ground Floor Seminar Room

25 Howland Street, London, W1T 4JG


Learning with (Flexible) Kernels



Reproducing kernel Hilbert spaces (RKHSs) play an important role in machine learning methods such as kernel mean embeddings, support vector machines (SVMs), and Gaussian processes. Analyzing these learning methods requires solid knowledge about the incorporated RKHSs and their interplay with probability measures. In the first part of my talk, I will discuss several aspects of this interplay. In the second part, I will try to explore the so far mostly unexploited flexibility of kernels: On the one hand, I will show that using a simple sum construction for locally defined kernels makes it possible to quickly train SVMs on millions of samples without sacrificing theoretical guarantees. On the other hand, I will discuss a class of hierarchical kernels whose structure mimics parts of deep neural network architectures.