Header logo is ei


2017


no image
Elements of Causal Inference - Foundations and Learning Algorithms

Peters, J., Janzing, D., Schölkopf, B.

Adaptive Computation and Machine Learning Series, The MIT Press, Cambridge, MA, USA, 2017 (book)

PDF [BibTex]

2017

PDF [BibTex]


no image
New Directions for Learning with Kernels and Gaussian Processes (Dagstuhl Seminar 16481)

Gretton, A., Hennig, P., Rasmussen, C., Schölkopf, B.

Dagstuhl Reports, 6(11):142-167, 2017 (book)

DOI [BibTex]


no image
Development and Evaluation of a Portable BCI System for Remote Data Acquisition

Emde, T.

Graduate School of Neural Information Processing, Eberhard Karls Universität Tübingen, Germany, 2017 (mastersthesis)

[BibTex]

[BibTex]


no image
Brain-Computer Interfaces for patients with Amyotrophic Lateral Sclerosis

Fomina, T.

Eberhard Karls Universität Tübingen, Germany, 2017 (phdthesis)

[BibTex]

[BibTex]


no image
Causal models for decision making via integrative inference

Geiger, P.

University of Stuttgart, Germany, 2017 (phdthesis)

[BibTex]

[BibTex]


no image
Learning Optimal Configurations for Modeling Frowning by Transcranial Electrical Stimulation

Sücker, K.

Graduate School of Neural Information Processing, Eberhard Karls Universität Tübingen, Germany, 2017 (mastersthesis)

[BibTex]

[BibTex]

2008


no image
Reinforcement Learning for Motor Primitives

Kober, J.

Biologische Kybernetik, University of Stuttgart, Stuttgart, Germany, August 2008 (diplomathesis)

PDF [BibTex]

2008

PDF [BibTex]


no image
Asymmetries of Time Series under Inverting their Direction

Peters, J.

Biologische Kybernetik, University of Heidelberg, August 2008 (diplomathesis)

PDF [BibTex]

PDF [BibTex]


no image
Learning an Interest Operator from Human Eye Movements

Kienzle, W.

Biologische Kybernetik, Eberhard-Karls-Universität Tübingen, Tübingen, Germany, July 2008 (phdthesis)

[BibTex]

[BibTex]


no image
Machine Learning for Robotics: Learning Methods for Robot Motor Skills

Peters, J.

pages: 107 , (Editors: J Peters), VDM-Verlag, Saarbrücken, Germany, May 2008 (book)

Abstract
Autonomous robots have been a vision of robotics, artificial intelligence, and cognitive sciences. An important step towards this goal is to create robots that can learn to accomplish amultitude of different tasks triggered by environmental context and higher-level instruction. Early approaches to this goal during the heydays of artificial intelligence research in the late 1980s showed that handcrafted approaches do not suffice and that machine learning is needed. However, off the shelf learning techniques often do not scale into real-time or to the high-dimensional domains of manipulator and humanoid robotics. In this book, we investigate the foundations for a general approach to motor skilllearning that employs domain-specific machine learning methods. A theoretically well-founded general approach to representing the required control structures for task representation and executionis presented along with novel learning algorithms that can be applied in this setting. The resulting framework is shown to work well both in simulation and on real robots.

Web [BibTex]

Web [BibTex]


no image
Causal inference from statistical data

Sun, X.

Biologische Kybernetik, Technische Hochschule Karlsruhe, Karlsruhe, Germany, April 2008 (phdthesis)

Web [BibTex]

Web [BibTex]


no image
Pairwise Correlations and Multineuronal Firing Patterns in Primary Visual Cortex

Berens, P.

Biologische Kybernetik, Eberhard Karls Universität Tübingen, Tübingen, Germany, April 2008 (diplomathesis)

[BibTex]

[BibTex]


no image
Development and Application of a Python Scripting Framework for BCI2000

Schreiner, T.

Biologische Kybernetik, Eberhard-Karls-Universität Tübingen, Tübingen, Germany, January 2008 (diplomathesis)

[BibTex]

[BibTex]


no image
Efficient and Invariant Regularisation with Application to Computer Graphics

Walder, CJ.

Biologische Kybernetik, University of Queensland, Brisbane, Australia, January 2008 (phdthesis)

Abstract
This thesis develops the theory and practise of reproducing kernel methods. Many functional inverse problems which arise in, for example, machine learning and computer graphics, have been treated with practical success using methods based on a reproducing kernel Hilbert space perspective. This perspective is often theoretically convenient, in that many functional analysis problems reduce to linear algebra problems in these spaces. Somewhat more complex is the case of conditionally positive definite kernels, and we provide an introduction to both cases, deriving in a particularly elementary manner some key results for the conditionally positive definite case. A common complaint of the practitioner is the long running time of these kernel based algorithms. We provide novel ways of alleviating these problems by essentially using a non-standard function basis which yields computational advantages. That said, by doing so we must also forego the aforementioned theoretical conveniences, and hence need some additional analysis which we provide in order to make the approach practicable. We demonstrate that the method leads to state of the art performance on the problem of surface reconstruction from points. We also provide some analysis of kernels invariant to transformations such as translation and dilation, and show that this indicates the value of learning algorithms which use conditionally positive definite kernels. Correspondingly, we provide a few approaches for making such algorithms practicable. We do this either by modifying the kernel, or directly solving problems with conditionally positive definite kernels, which had previously only been solved with positive definite kernels. We demonstrate the advantage of this approach, in particular by attaining state of the art classification performance with only one free parameter.

PDF [BibTex]

PDF [BibTex]

2002


no image
Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond

Schölkopf, B., Smola, A.

pages: 644, Adaptive Computation and Machine Learning, MIT Press, Cambridge, MA, USA, December 2002, Parts of this book, including an introduction to kernel methods, can be downloaded here. (book)

Abstract
In the 1990s, a new type of learning algorithm was developed, based on results from statistical learning theory: the Support Vector Machine (SVM). This gave rise to a new class of theoretically elegant learning machines that use a central concept of SVMs-kernels—for a number of learning tasks. Kernel machines provide a modular framework that can be adapted to different tasks and domains by the choice of the kernel function and the base algorithm. They are replacing neural networks in a variety of fields, including engineering, information retrieval, and bioinformatics. Learning with Kernels provides an introduction to SVMs and related kernel methods. Although the book begins with the basics, it also includes the latest research. It provides all of the concepts necessary to enable a reader equipped with some basic mathematical knowledge to enter the world of machine learning using theoretically well-founded yet easy-to-use kernel algorithms and to understand and apply the powerful algorithms that have been developed over the last few years.

Web [BibTex]

2002

Web [BibTex]


no image
Nonlinear Multivariate Analysis with Geodesic Kernels

Kuss, M.

Biologische Kybernetik, Technische Universität Berlin, February 2002 (diplomathesis)

GZIP [BibTex]

GZIP [BibTex]


no image
Concentration Inequalities and Empirical Processes Theory Applied to the Analysis of Learning Algorithms

Bousquet, O.

Biologische Kybernetik, Ecole Polytechnique, 2002 (phdthesis) Accepted

Abstract
New classification algorithms based on the notion of 'margin' (e.g. Support Vector Machines, Boosting) have recently been developed. The goal of this thesis is to better understand how they work, via a study of their theoretical performance. In order to do this, a general framework for real-valued classification is proposed. In this framework, it appears that the natural tools to use are Concentration Inequalities and Empirical Processes Theory. Thanks to an adaptation of these tools, a new measure of the size of a class of functions is introduced, which can be computed from the data. This allows, on the one hand, to better understand the role of eigenvalues of the kernel matrix in Support Vector Machines, and on the other hand, to obtain empirical model selection criteria.

PostScript [BibTex]


no image
Support Vector Machines: Induction Principle, Adaptive Tuning and Prior Knowledge

Chapelle, O.

Biologische Kybernetik, 2002 (phdthesis)

Abstract
This thesis presents a theoretical and practical study of Support Vector Machines (SVM) and related learning algorithms. In a first part, we introduce a new induction principle from which SVMs can be derived, but some new algorithms are also presented in this framework. In a second part, after studying how to estimate the generalization error of an SVM, we suggest to choose the kernel parameters of an SVM by minimizing this estimate. Several applications such as feature selection are presented. Finally the third part deals with the incoporation of prior knowledge in a learning algorithm and more specifically, we studied the case of known invariant transormations and the use of unlabeled data.

GZIP [BibTex]

1997


no image
Support vector learning

Schölkopf, B.

pages: 173, Oldenbourg, München, Germany, 1997, Zugl.: Berlin, Techn. Univ., Diss., 1997 (book)

PDF GZIP [BibTex]

1997

PDF GZIP [BibTex]