Header logo is ei


2015


no image
easyGWAS: An Integrated Computational Framework for Advanced Genome-Wide Association Studies

Grimm, Dominik

Eberhard Karls Universität Tübingen, November 2015 (phdthesis)

[BibTex]

2015

[BibTex]


no image
Causal Discovery Beyond Conditional Independences

Sgouritsa, E.

Eberhard Karls Universität Tübingen, Germany, October 2015 (phdthesis)

link (url) [BibTex]

link (url) [BibTex]


no image
From Points to Probability Measures: A Statistical Learning on Distributions with Kernel Mean Embedding

Muandet, K.

University of Tübingen, Germany, University of Tübingen, Germany, September 2015 (phdthesis)

[BibTex]

[BibTex]


no image
Machine Learning Approaches to Image Deconvolution

Schuler, C.

University of Tübingen, Germany, University of Tübingen, Germany, September 2015 (phdthesis)

[BibTex]

[BibTex]


no image
Blind Retrospective Motion Correction of MR Images

Loktyushin, A.

University of Tübingen, Germany, May 2015 (phdthesis)

[BibTex]

[BibTex]


no image
A Cognitive Brain-Computer Interface for Patients with Amyotrophic Lateral Sclerosis

Hohmann, M.

Graduate Training Centre of Neuroscience, University of Tübingen, Germany, 2015 (mastersthesis)

[BibTex]

[BibTex]


no image
Sequential Image Deconvolution Using Probabilistic Linear Algebra

Gao, M.

Technical University of Munich, Germany, 2015 (mastersthesis)

[BibTex]

[BibTex]


no image
Causal Inference in Neuroimaging

Casarsa de Azevedo, L.

Graduate Training Centre of Neuroscience, University of Tübingen, Germany, 2015 (mastersthesis)

[BibTex]

[BibTex]


no image
The effect of frowning on attention

Ibarra Chaoul, A.

Graduate Training Centre of Neuroscience, University of Tübingen, Germany, 2015 (mastersthesis)

[BibTex]

[BibTex]

2002


no image
Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond

Schölkopf, B., Smola, A.

pages: 644, Adaptive Computation and Machine Learning, MIT Press, Cambridge, MA, USA, December 2002, Parts of this book, including an introduction to kernel methods, can be downloaded here. (book)

Abstract
In the 1990s, a new type of learning algorithm was developed, based on results from statistical learning theory: the Support Vector Machine (SVM). This gave rise to a new class of theoretically elegant learning machines that use a central concept of SVMs-kernels—for a number of learning tasks. Kernel machines provide a modular framework that can be adapted to different tasks and domains by the choice of the kernel function and the base algorithm. They are replacing neural networks in a variety of fields, including engineering, information retrieval, and bioinformatics. Learning with Kernels provides an introduction to SVMs and related kernel methods. Although the book begins with the basics, it also includes the latest research. It provides all of the concepts necessary to enable a reader equipped with some basic mathematical knowledge to enter the world of machine learning using theoretically well-founded yet easy-to-use kernel algorithms and to understand and apply the powerful algorithms that have been developed over the last few years.

Web [BibTex]

2002

Web [BibTex]


no image
Nonlinear Multivariate Analysis with Geodesic Kernels

Kuss, M.

Biologische Kybernetik, Technische Universität Berlin, February 2002 (diplomathesis)

GZIP [BibTex]

GZIP [BibTex]


no image
Concentration Inequalities and Empirical Processes Theory Applied to the Analysis of Learning Algorithms

Bousquet, O.

Biologische Kybernetik, Ecole Polytechnique, 2002 (phdthesis) Accepted

Abstract
New classification algorithms based on the notion of 'margin' (e.g. Support Vector Machines, Boosting) have recently been developed. The goal of this thesis is to better understand how they work, via a study of their theoretical performance. In order to do this, a general framework for real-valued classification is proposed. In this framework, it appears that the natural tools to use are Concentration Inequalities and Empirical Processes Theory. Thanks to an adaptation of these tools, a new measure of the size of a class of functions is introduced, which can be computed from the data. This allows, on the one hand, to better understand the role of eigenvalues of the kernel matrix in Support Vector Machines, and on the other hand, to obtain empirical model selection criteria.

PostScript [BibTex]


no image
Support Vector Machines: Induction Principle, Adaptive Tuning and Prior Knowledge

Chapelle, O.

Biologische Kybernetik, 2002 (phdthesis)

Abstract
This thesis presents a theoretical and practical study of Support Vector Machines (SVM) and related learning algorithms. In a first part, we introduce a new induction principle from which SVMs can be derived, but some new algorithms are also presented in this framework. In a second part, after studying how to estimate the generalization error of an SVM, we suggest to choose the kernel parameters of an SVM by minimizing this estimate. Several applications such as feature selection are presented. Finally the third part deals with the incoporation of prior knowledge in a learning algorithm and more specifically, we studied the case of known invariant transormations and the use of unlabeled data.

GZIP [BibTex]

2000


no image
Advances in Large Margin Classifiers

Smola, A., Bartlett, P., Schölkopf, B., Schuurmans, D.

pages: 422, Neural Information Processing, MIT Press, Cambridge, MA, USA, October 2000 (book)

Abstract
The concept of large margins is a unifying principle for the analysis of many different approaches to the classification of data from examples, including boosting, mathematical programming, neural networks, and support vector machines. The fact that it is the margin, or confidence level, of a classification--that is, a scale parameter--rather than a raw training error that matters has become a key tool for dealing with classifiers. This book shows how this idea applies to both the theoretical analysis and the design of algorithms. The book provides an overview of recent developments in large margin classifiers, examines connections with other methods (e.g., Bayesian inference), and identifies strengths and weaknesses of the method, as well as directions for future research. Among the contributors are Manfred Opper, Vladimir Vapnik, and Grace Wahba.

Web [BibTex]

2000

Web [BibTex]


no image
Three-dimensional reconstruction of planar scenes

Urbanek, M.

Biologische Kybernetik, INP Grenoble, Warsaw University of Technology, September 2000 (diplomathesis)

Abstract
For a planar scene, we propose an algorithm to estimate its 3D structure. Homographies between corresponding planes are employed in order to recover camera motion parameters - between camera positions from which images of the scene were taken. Cases of one- and multiple- corresponding planes present on the scene are distinguished. Solutions are proposed for both cases.

ZIP [BibTex]

ZIP [BibTex]


no image
Intelligence as a Complex System

Zhou, D.

Biologische Kybernetik, 2000 (phdthesis)

[BibTex]

[BibTex]


no image
Neural Networks in Robot Control

Peters, J.

Biologische Kybernetik, Fernuniversität Hagen, Hagen, Germany, 2000 (diplomathesis)

[BibTex]

[BibTex]