Header logo is ei


2015


no image
Semi-Autonomous 3rd-Hand Robot

Lopes, M., Peters, J., Piater, J., Toussaint, M., Baisero, A., Busch, B., Erkent, O., Kroemer, O., Lioutikov, R., Maeda, G., Mollard, Y., Munzer, T., Shukla, D.

In Workshop on Cognitive Robotics in Future Manufacturing Scenarios, European Robotics Forum, 2015 (inproceedings)

link (url) [BibTex]

2015

link (url) [BibTex]


no image
A Cognitive Brain-Computer Interface for Patients with Amyotrophic Lateral Sclerosis

Hohmann, M.

Graduate Training Centre of Neuroscience, University of Tübingen, Germany, 2015 (mastersthesis)

[BibTex]

[BibTex]


no image
Neural Adaptive Sequential Monte Carlo

Gu, S., Ghahramani, Z., Turner, R. E.

Advances in Neural Information Processing Systems 28, pages: 2629-2637, (Editors: Corinna Cortes, Neil D. Lawrence, Daniel D. Lee, Masashi Sugiyama, and Roman Garnett), 29th Annual Conference on Neural Information Processing Systems (NIPS), 2015 (conference)

PDF Supplementary [BibTex]

PDF Supplementary [BibTex]


no image
Discovering Temporal Causal Relations from Subsampled Data

Gong, M., Zhang, K., Schölkopf, B., Tao, D., Geiger, P.

In Proceedings of the 32nd International Conference on Machine Learning, 37, pages: 1898–1906, JMLR Workshop and Conference Proceedings, (Editors: F. Bach and D. Blei), JMLR, ICML, 2015 (inproceedings)

PDF link (url) [BibTex]

PDF link (url) [BibTex]


no image
Active Nearest Neighbors in Changing Environments

Berlind, C., Urner, R.

In Proceedings of the 32nd International Conference on Machine Learning, 37, pages: 1870-1879, JMLR Workshop and Conference Proceedings, (Editors: Bach, F. and Blei, D. ), JMLR, ICML, 2015 (inproceedings)

link (url) [BibTex]

link (url) [BibTex]


no image
Learning Inverse Dynamics Models with Contacts

Calandra, R., Ivaldi, S., Deisenroth, M., Rückert, E., Peters, J.

In IEEE International Conference on Robotics and Automation, pages: 3186-3191, ICRA, 2015 (inproceedings)

link (url) DOI [BibTex]

link (url) DOI [BibTex]


no image
A Probabilistic Framework for Semi-Autonomous Robots Based on Interaction Primitives with Phase Estimation

Maeda, G., Neumann, G., Ewerton, M., Lioutikov, R., Peters, J.

In Proceedings of the International Symposium of Robotics Research, ISRR, 2015 (inproceedings)

link (url) [BibTex]

link (url) [BibTex]


Peer grading in a course on algorithms and data structures
Peer grading in a course on algorithms and data structures

Sajjadi, M. S. M., Alamgir, M., von Luxburg, U.

Workshop on Machine Learning for Education (ML4Ed) at the 32th International Conference on Machine Learning (ICML), 2015 (conference)

Arxiv [BibTex]

Arxiv [BibTex]


no image
Removing systematic errors for exoplanet search via latent causes

Schölkopf, B., Hogg, D., Wang, D., Foreman-Mackey, D., Janzing, D., Simon-Gabriel, C. J., Peters, J.

In Proceedings of The 32nd International Conference on Machine Learning, 37, pages: 2218–2226, JMLR Workshop and Conference Proceedings, (Editors: Bach, F. and Blei, D.), JMLR, ICML, 2015 (inproceedings)

Extended version on arXiv link (url) [BibTex]

Extended version on arXiv link (url) [BibTex]


no image
Causal Inference by Identification of Vector Autoregressive Processes with Hidden Components

Geiger, P., Zhang, K., Schölkopf, B., Gong, M., Janzing, D.

In Proceedings of the 32nd International Conference on Machine Learning, 37, pages: 1917–1925, JMLR Workshop and Conference Proceedings, (Editors: F. Bach and D. Blei), JMLR, ICML, 2015 (inproceedings)

PDF link (url) [BibTex]

PDF link (url) [BibTex]


no image
Brain-Computer Interfacing in Amyotrophic Lateral Sclerosis: Implications of a Resting-State EEG Analysis

Jayaram, V., Widmann, N., Förster, C., Fomina, T., Hohmann, M. R., Müller vom Hagen, J., Synofzik, M., Schölkopf, B., Schöls, L., Grosse-Wentrup, M.

In Proceedings of the 37th IEEE Conference for Engineering in Medicine and Biology, pages: 6979-6982, EMBC, 2015 (inproceedings)

PDF DOI [BibTex]

PDF DOI [BibTex]


no image
Identification of the Default Mode Network with Electroencephalography

Fomina, T., Hohmann, M. R., Schölkopf, B., Grosse-Wentrup, M.

In Proceedings of the 37th IEEE Conference for Engineering in Medicine and Biology, pages: 7566-7569, EMBC, 2015 (inproceedings)

DOI [BibTex]

DOI [BibTex]


no image
Towards Cognitive Brain-Computer Interfaces for Patients with Amyotrophic Lateral Sclerosis

Fomina, T., Schölkopf, B., Grosse-Wentrup, M.

In 7th Computer Science and Electronic Engineering Conference, pages: 77-80, Curran Associates, Inc., CEEC, 2015 (inproceedings)

DOI [BibTex]

DOI [BibTex]


no image
Towards Learning Hierarchical Skills for Multi-Phase Manipulation Tasks

Kroemer, O., Daniel, C., Neumann, G., van Hoof, H., Peters, J.

In IEEE International Conference on Robotics and Automation, pages: 1503 - 1510, ICRA, 2015 (inproceedings)

link (url) DOI [BibTex]

link (url) DOI [BibTex]


Probabilistic Line Searches for Stochastic Optimization
Probabilistic Line Searches for Stochastic Optimization

Mahsereci, M., Hennig, P.

In Advances in Neural Information Processing Systems 28, pages: 181-189, (Editors: C. Cortes, N.D. Lawrence, D.D. Lee, M. Sugiyama and R. Garnett), Curran Associates, Inc., 29th Annual Conference on Neural Information Processing Systems (NIPS), 2015 (inproceedings)

Abstract
In deterministic optimization, line searches are a standard tool ensuring stability and efficiency. Where only stochastic gradients are available, no direct equivalent has so far been formulated, because uncertain gradients do not allow for a strict sequence of decisions collapsing the search space. We construct a probabilistic line search by combining the structure of existing deterministic methods with notions from Bayesian optimization. Our method retains a Gaussian process surrogate of the univariate optimization objective, and uses a probabilistic belief over the Wolfe conditions to monitor the descent. The algorithm has very low computational cost, and no user-controlled parameters. Experiments show that it effectively removes the need to define a learning rate for stochastic gradient descent. [You can find the matlab research code under `attachments' below. The zip-file contains a minimal working example. The docstring in probLineSearch.m contains additional information. A more polished implementation in C++ will be published here at a later point. For comments and questions about the code please write to mmahsereci@tue.mpg.de.]

Matlab research code link (url) [BibTex]

Matlab research code link (url) [BibTex]


no image
Assessment of tumor heterogeneity using unsupervised graph based clustering of multi-modality imaging data

Katiyar, P., Divine, M. R., Pichler, B. J., Disselhorst, J. A.

European Molecular Imaging Meeting, 2015 (poster)

[BibTex]

[BibTex]


no image
BACKSHIFT: Learning causal cyclic graphs from unknown shift interventions

Rothenhäusler, D., Heinze, C., Peters, J., Meinshausen, N.

Advances in Neural Information Processing Systems 28, pages: 1513-1521, (Editors: C. Cortes, N.D. Lawrence, D.D. Lee, M. Sugiyama and R. Garnett), Curran Associates, Inc., 29th Annual Conference on Neural Information Processing Systems (NIPS), 2015 (conference)

link (url) [BibTex]

link (url) [BibTex]


no image
Particle Gibbs for Infinite Hidden Markov Models

Tripuraneni*, N., Gu*, S., Ge, H., Ghahramani, Z.

Advances in Neural Information Processing Systems 28, pages: 2395-2403, (Editors: Corinna Cortes, Neil D. Lawrence, Daniel D. Lee, Masashi Sugiyama, and Roman Garnett), 29th Annual Conference on Neural Information Processing Systems (NIPS), 2015, *equal contribution (conference)

PDF [BibTex]

PDF [BibTex]


Peer grading in a course on algorithms and data structures
Peer grading in a course on algorithms and data structures

Sajjadi, M. S. M., Alamgir, M., von Luxburg, U.

Workshop on Crowdsourcing and Machine Learning (CrowdML) Workshop on Machine Learning for Education (ML4Ed) at at the 32th International Conference on Machine Learning (ICML), 2015 (conference)

Arxiv [BibTex]

Arxiv [BibTex]


no image
A Random Riemannian Metric for Probabilistic Shortest-Path Tractography

Hauberg, S., Schober, M., Liptrot, M., Hennig, P., Feragen, A.

In 18th International Conference on Medical Image Computing and Computer Assisted Intervention, 9349, pages: 597-604, Lecture Notes in Computer Science, MICCAI, 2015 (inproceedings)

PDF DOI [BibTex]

PDF DOI [BibTex]


no image
Recent Methodological Advances in Causal Discovery and Inference

Spirtes, P., Zhang, K.

In 15th Conference on Theoretical Aspects of Rationality and Knowledge, pages: 23-35, (Editors: Ramanujam, R.), TARK, 2015 (inproceedings)

[BibTex]

[BibTex]


no image
Learning Optimal Striking Points for A Ping-Pong Playing Robot

Huang, Y., Schölkopf, B., Peters, J.

In IEEE/RSJ International Conference on Intelligent Robots and Systems, pages: 4587-4592, IROS, 2015 (inproceedings)

PDF DOI [BibTex]

PDF DOI [BibTex]


no image
Model-Based Relative Entropy Stochastic Search

Abdolmaleki, A., Peters, J., Neumann, G.

In Advances in Neural Information Processing Systems 28, pages: 3523-3531, (Editors: C. Cortes, N.D. Lawrence, D.D. Lee, M. Sugiyama and R. Garnett), Curran Associates, Inc., 29th Annual Conference on Neural Information Processing Systems (NIPS), 2015 (inproceedings)

link (url) [BibTex]

link (url) [BibTex]


no image
Modeling Spatio-Temporal Variability in Human-Robot Interaction with Probabilistic Movement Primitives

Ewerton, M., Neumann, G., Lioutikov, R., Ben Amor, H., Peters, J., Maeda, G.

In Workshop on Machine Learning for Social Robotics, ICRA, 2015 (inproceedings)

link (url) [BibTex]

link (url) [BibTex]


no image
Extracting Low-Dimensional Control Variables for Movement Primitives

Rueckert, E., Mundo, J., Paraschos, A., Peters, J., Neumann, G.

In IEEE International Conference on Robotics and Automation, pages: 1511-1518, ICRA, 2015 (inproceedings)

link (url) DOI [BibTex]

link (url) DOI [BibTex]


no image
Self-calibration of optical lenses

Hirsch, M., Schölkopf, B.

In IEEE International Conference on Computer Vision (ICCV 2015), pages: 612-620, IEEE, 2015 (inproceedings)

DOI [BibTex]

DOI [BibTex]


no image
Sequential Image Deconvolution Using Probabilistic Linear Algebra

Gao, M.

Technical University of Munich, Germany, 2015 (mastersthesis)

[BibTex]

[BibTex]


no image
Telling cause from effect in deterministic linear dynamical systems

Shajarisales, N., Janzing, D., Schölkopf, B., Besserve, M.

In Proceedings of the 32nd International Conference on Machine Learning, 37, pages: 285–294, JMLR Workshop and Conference Proceedings, (Editors: F. Bach and D. Blei), JMLR, ICML, 2015 (inproceedings)

PDF [BibTex]

PDF [BibTex]


no image
A Cognitive Brain-Computer Interface for Patients with Amyotrophic Lateral Sclerosis

Hohmann, M. R., Fomina, T., Jayaram, V., Widmann, N., Förster, C., Müller vom Hagen, J., Synofzik, M., Schölkopf, B., Schöls, L., Grosse-Wentrup, M.

In Proceedings of the 2015 IEEE International Conference on Systems, Man, and Cybernetics, pages: 3187-3191, SMC, 2015 (inproceedings)

PDF DOI [BibTex]

PDF DOI [BibTex]


no image
Efficient Learning of Linear Separators under Bounded Noise

Awasthi, P., Balcan, M., Haghtalab, N., Urner, R.

In Proceedings of the 28th Conference on Learning Theory, 40, pages: 167-190, (Editors: Grünwald, P. and Hazan, E. and Kale, S.), JMLR, COLT, 2015 (inproceedings)

link (url) [BibTex]

link (url) [BibTex]


no image
Learning multiple collaborative tasks with a mixture of Interaction Primitives

Ewerton, M., Neumann, G., Lioutikov, R., Ben Amor, H., Peters, J., Maeda, G.

In IEEE International Conference on Robotics and Automation, pages: 1535-1542, ICRA, 2015 (inproceedings)

link (url) DOI [BibTex]

link (url) DOI [BibTex]


no image
Disparity estimation from a generative light field model

Köhler, R., Schölkopf, B., Hirsch, M.

IEEE International Conference on Computer Vision (ICCV 2015), Workshop on Inverse Rendering, 2015, Note: This work has been presented as a poster and is not included in the workshop proceedings. (poster)

[BibTex]

[BibTex]


no image
Causal Inference in Neuroimaging

Casarsa de Azevedo, L.

Graduate Training Centre of Neuroscience, University of Tübingen, Germany, 2015 (mastersthesis)

[BibTex]

[BibTex]


no image
The effect of frowning on attention

Ibarra Chaoul, A.

Graduate Training Centre of Neuroscience, University of Tübingen, Germany, 2015 (mastersthesis)

[BibTex]

[BibTex]


no image
Subspace Alignement based Domain Adaptation for RCNN detector

Raj, A., V., N., Tuytelaars, T.

Proceedings of the 26th British Machine Vision Conference (BMVC 2015), pages: 166.1-166.11, (Editors: Xianghua Xie and Mark W. Jones and Gary K. L. Tam), 2015 (conference)

DOI [BibTex]

DOI [BibTex]


no image
Practical Probabilistic Programming with Monads

Ścibior, A., Ghahramani, Z., Gordon, A. D.

Proceedings of the 2015 ACM SIGPLAN Symposium on Haskell, pages: 165-176, Haskell ’15, ACM, 2015 (conference)

DOI [BibTex]

DOI [BibTex]


no image
Developing neural networks with neurons competing for survival

Peng, Z, Braun, DA

pages: 152-153, IEEE, Piscataway, NJ, USA, 5th Joint IEEE International Conference on Development and Learning and on Epigenetic Robotics (IEEE ICDL-EPIROB), August 2015 (conference)

Abstract
We study developmental growth in a feedforward neural network model inspired by the survival principle in nature. Each neuron has to select its incoming connections in a way that allow it to fire, as neurons that are not able to fire over a period of time degenerate and die. In order to survive, neurons have to find reoccurring patterns in the activity of the neurons in the preceding layer, because each neuron requires more than one active input at any one time to have enough activation for firing. The sensory input at the lowest layer therefore provides the maximum amount of activation that all neurons compete for. The whole network grows dynamically over time depending on how many patterns can be found and how many neurons can maintain themselves accordingly. We show in simulations that this naturally leads to abstractions in higher layers that emerge in a unsupervised fashion. When evaluating the network in a supervised learning paradigm, it is clear that our network is not competitive. What is interesting though is that this performance was achieved by neurons that simply struggle for survival and do not know about performance error. In contrast to most studies on neural evolution that rely on a network-wide fitness function, our goal was to show that learning behaviour can appear in a system without being driven by any specific utility function or reward signal.

DOI [BibTex]

DOI [BibTex]

2006


no image
Global Biclustering of Microarray Data

Wolf, T., Brors, B., Hofmann, T., Georgii, E.

In ICDMW 2006, pages: 125-129, (Editors: Tsumoto, S. , C. W. Clifton, N. Zhong, X. Wu, J. Liu, B. W. Wah, Y.-M. Cheung), IEEE Computer Society, Los Alamitos, CA, USA, Sixth IEEE International Conference on Data Mining, December 2006 (inproceedings)

Abstract
We consider the problem of simultaneously clustering genes and conditions of a gene expression data matrix. A bicluster is defined as a subset of genes that show similar behavior within a subset of conditions. Finding biclusters can be useful for revealing groups of genes involved in the same molecular process as well as groups of conditions where this process takes place. Previous work either deals with local, bicluster-based criteria or assumes a very specific structure of the data matrix (e.g. checkerboard or block-diagonal) [11]. In contrast, our goal is to find a set of flexibly arranged biclusters which is optimal in regard to a global objective function. As this is a NP-hard combinatorial problem, we describe several techniques to obtain approximate solutions. We benchmarked our approach successfully on the Alizadeh B-cell lymphoma data set [1].

Web DOI [BibTex]

2006

Web DOI [BibTex]


no image
Conformal Multi-Instance Kernels

Blaschko, M., Hofmann, T.

In NIPS 2006 Workshop on Learning to Compare Examples, pages: 1-6, NIPS Workshop on Learning to Compare Examples, December 2006 (inproceedings)

Abstract
In the multiple instance learning setting, each observation is a bag of feature vectors of which one or more vectors indicates membership in a class. The primary task is to identify if any vectors in the bag indicate class membership while ignoring vectors that do not. We describe here a kernel-based technique that defines a parametric family of kernels via conformal transformations and jointly learns a discriminant function over bags together with the optimal parameter settings of the kernel. Learning a conformal transformation effectively amounts to weighting regions in the feature space according to their contribution to classification accuracy; regions that are discriminative will be weighted higher than regions that are not. This allows the classifier to focus on regions contributing to classification accuracy while ignoring regions that correspond to vectors found both in positive and in negative bags. We show how parameters of this transformation can be learned for support vector machines by posing the problem as a multiple kernel learning problem. The resulting multiple instance classifier gives competitive accuracy for several multi-instance benchmark datasets from different domains.

PDF Web [BibTex]

PDF Web [BibTex]


no image
Some observations on the pedestal effect or dipper function

Henning, B., Wichmann, F.

Journal of Vision, 6(13):50, 2006 Fall Vision Meeting of the Optical Society of America, December 2006 (poster)

Abstract
The pedestal effect is the large improvement in the detectabilty of a sinusoidal “signal” grating observed when the signal is added to a masking or “pedestal” grating of the same spatial frequency, orientation, and phase. We measured the pedestal effect in both broadband and notched noise - noise from which a 1.5-octave band centred on the signal frequency had been removed. Although the pedestal effect persists in broadband noise, it almost disappears in the notched noise. Furthermore, the pedestal effect is substantial when either high- or low-pass masking noise is used. We conclude that the pedestal effect in the absence of notched noise results principally from the use of information derived from channels with peak sensitivities at spatial frequencies different from that of the signal and pedestal. The spatial-frequency components of the notched noise above and below the spatial frequency of the signal and pedestal prevent the use of information about changes in contrast carried in channels tuned to spatial frequencies that are very much different from that of the signal and pedestal. Thus the pedestal or dipper effect measured without notched noise is not a characteristic of individual spatial-frequency tuned channels.

Web DOI [BibTex]

Web DOI [BibTex]


no image
Information-theoretic Metric Learning

Davis, J., Kulis, B., Sra, S., Dhillon, I.

In NIPS 2006 Workshop on Learning to Compare Examples, pages: 1-5, NIPS Workshop on Learning to Compare Examples, December 2006 (inproceedings)

Abstract
We formulate the metric learning problem as that of minimizing the differential relative entropy between two multivariate Gaussians under constraints on the Mahalanobis distance function. Via a surprising equivalence, we show that this problem can be solved as a low-rank kernel learning problem. Specifically, we minimize the Burg divergence of a low-rank kernel to an input kernel, subject to pairwise distance constraints. Our approach has several advantages over existing methods. First, we present a natural information-theoretic formulation for the problem. Second, the algorithm utilizes the methods developed by Kulis et al. [6], which do not involve any eigenvector computation; in particular, the running time of our method is faster than most existing techniques. Third, the formulation offers insights into connections between metric learning and kernel learning.

PDF Web [BibTex]

PDF Web [BibTex]


no image
Pattern Mining in Frequent Dynamic Subgraphs

Borgwardt, KM., Kriegel, H-P., Wackersreuther, P.

In pages: 818-822, (Editors: Clifton, C.W.), IEEE Computer Society, Los Alamitos, CA, USA, Sixth International Conference on Data Mining (ICDM), December 2006 (inproceedings)

Abstract
Graph-structured data is becoming increasingly abundant in many application domains. Graph mining aims at finding interesting patterns within this data that represent novel knowledge. While current data mining deals with static graphs that do not change over time, coming years will see the advent of an increasing number of time series of graphs. In this article, we investigate how pattern mining on static graphs can be extended to time series of graphs. In particular, we are considering dynamic graphs with edge insertions and edge deletions over time. We define frequency in this setting and provide algorithmic solutions for finding frequent dynamic subgraph patterns. Existing subgraph mining algorithms can be easily integrated into our framework to make them handle dynamic graphs. Experimental results on real-world data confirm the practical feasibility of our approach.

Web DOI [BibTex]

Web DOI [BibTex]


no image
3DString: a feature string kernel for 3D object classification on voxelized data

Assfalg, J., Borgwardt, KM., Kriegel, H-P.

In pages: 198-207, (Editors: Yu, P.S. , V.J. Tsotras, E.A. Fox, B. Liu), ACM Press, New York, NY, USA, 15th ACM International Conference on Information and Knowledge Management (CIKM), November 2006 (inproceedings)

Abstract
Classification of 3D objects remains an important task in many areas of data management such as engineering, medicine or biology. As a common preprocessing step in current approaches to classification of voxelized 3D objects, voxel representations are transformed into a feature vector description.In this article, we introduce an approach of transforming 3D objects into feature strings which represent the distribution of voxels over the voxel grid. Attractively, this feature string extraction can be performed in linear runtime with respect to the number of voxels. We define a similarity measure on these feature strings that counts common k-mers in two input strings, which is referred to as the spectrum kernel in the field of kernel methods. We prove that on our feature strings, this similarity measure can be computed in time linear to the number of different characters in these strings. This linear runtime behavior makes our kernel attractive even for large datasets that occur in many application domains. Furthermore, we explain that our similarity measure induces a metric which allows to combine it with an M-tree for handling of large volumes of data. Classification experiments on two published benchmark datasets show that our novel approach is competitive with the best state-of-the-art methods for 3D object classification.

DOI [BibTex]

DOI [BibTex]


no image
Adapting Spatial Filter Methods for Nonstationary BCIs

Tomioka, R., Hill, J., Blankertz, B., Aihara, K.

In IBIS 2006, pages: 65-70, 2006 Workshop on Information-Based Induction Sciences, November 2006 (inproceedings)

Abstract
A major challenge in applying machine learning methods to Brain-Computer Interfaces (BCIs) is to overcome the possible nonstationarity in the data from the datablock the method is trained on and that the method is applied to. Assuming the joint distributions of the whitened signal and the class label to be identical in two blocks, where the whitening is done in each block independently, we propose a simple adaptation formula that is applicable to a broad class of spatial filtering methods including ICA, CSP, and logistic regression classifiers. We characterize the class of linear transformations for which the above assumption holds. Experimental results on 60 BCI datasets show improved classification accuracy compared to (a) fixed spatial filter approach (no adaptation) and (b) fixed spatial pattern approach (proposed by Hill et al., 2006 [1]).

PDF [BibTex]

PDF [BibTex]


no image
Optimizing Spatial Filters for BCI: Margin- and Evidence-Maximization Approaches

Farquhar, J., Hill, N., Schölkopf, B.

Challenging Brain-Computer Interfaces: MAIA Workshop 2006, pages: 1, November 2006 (poster)

Abstract
We present easy-to-use alternatives to the often-used two-stage Common Spatial Pattern + classifier approach for spatial filtering and classification of Event-Related Desychnronization signals in BCI. We report two algorithms that aim to optimize the spatial filters according to a criterion more directly related to the ability of the algorithms to generalize to unseen data. Both are based upon the idea of treating the spatial filter coefficients as hyperparameters of a kernel or covariance function. We then optimize these hyper-parameters directly along side the normal classifier parameters with respect to our chosen learning objective function. The two objectives considered are margin maximization as used in Support-Vector Machines and the evidence maximization framework used in Gaussian Processes. Our experiments assessed generalization error as a function of the number of training points used, on 9 BCI competition data sets and 5 offline motor imagery data sets measured in Tubingen. Both our approaches sho w consistent improvements relative to the commonly used CSP+linear classifier combination. Strikingly, the improvement is most significant in the higher noise cases, when either few trails are used for training, or with the most poorly performing subjects. This a reversal of the usual "rich get richer" effect in the development of CSP extensions, which tend to perform best when the signal is strong enough to accurately find their additional parameters. This makes our approach particularly suitable for clinical application where high levels of noise are to be expected.

PDF PDF [BibTex]

PDF PDF [BibTex]


no image
Extraction of visual features from natural video data using Slow Feature Analysis

Nickisch, H.

Biologische Kybernetik, Technische Universität Berlin, Berlin, Germany, September 2006 (diplomathesis)

Abstract
Das Forschungsprojekt NeuRoBot hat das un{\"u}berwachte Erlernen einer neuronal inspirierten Steuerungsarchitektur zum Ziel, und zwar unter den Randbedingungen biologischer Plausibilit{\"a}t und der Benutzung einer Kamera als einzigen Sensor. Visuelle Merkmale, die ein angemessenes Abbild der Umgebung liefern, sind unerl{\"a}sslich, um das Ziel kollisionsfreier Navigation zu erreichen. Zeitliche Koh{\"a}renz ist ein neues Lernprinzip, das in der Lage ist, Erkenntnisse aus der Biologie des Sehens zu reproduzieren. Es wird durch die Beobachtung motiviert, dass die “Sensoren” der Retina auf deutlich k{\"u}rzeren Zeitskalen variieren als eine abstrakte Beschreibung. Zeitliche Langsamkeitsanalyse l{\"o}st das Problem, indem sie zeitlich langsam ver{\"a}nderliche Signale aus schnell ver{\"a}nderlichen Eingabesignalen extrahiert. Eine Verallgemeinerung auf Signale, die nichtlinear von den Eingaben abh{\"a}ngen, ist durch die Anwendung des Kernel-Tricks m{\"o}glich. Das einzig benutzte Vorwissen ist die zeitliche Glattheit der gewonnenen Signale. In der vorliegenden Diplomarbeit wird Langsamkeitsanalyse auf Bildausschnitte von Videos einer Roboterkamera und einer Simulationsumgebung angewendet. Zuallererst werden mittels Parameterexploration und Kreuzvalidierung die langsamst m{\"o}glichen Funktionen bestimmt. Anschließend werden die Merkmalsfunktionen analysiert und einige Ansatzpunkte f{\"u}r ihre Interpretation angegeben. Aufgrund der sehr großen Datens{\"a}tze und der umfangreichen Berechnungen behandelt ein Großteil dieser Arbeit auch Aufwandsbetrachtungen und Fragen der effizienten Berechnung. Kantendetektoren in verschiedenen Phasen und mit haupts{\"a}chlich horizontaler Orientierung stellen die wichtigsten aus der Analyse hervorgehenden Funktionen dar. Eine Anwendung auf konkrete Navigationsaufgaben des Roboters konnte bisher nicht erreicht werden. Eine visuelle Interpretation der erlernten Merkmale ist jedoch durchaus gegeben.

PDF [BibTex]

PDF [BibTex]


no image
A Linear Programming Approach for Molecular QSAR analysis

Saigo, H., Kadowaki, T., Tsuda, K.

In MLG 2006, pages: 85-96, (Editors: Gärtner, T. , G. C. Garriga, T. Meinl), International Workshop on Mining and Learning with Graphs, September 2006, Best Paper Award (inproceedings)

Abstract
Small molecules in chemistry can be represented as graphs. In a quantitative structure-activity relationship (QSAR) analysis, the central task is to find a regression function that predicts the activity of the molecule in high accuracy. Setting a QSAR as a primal target, we propose a new linear programming approach to the graph-based regression problem. Our method extends the graph classification algorithm by Kudo et al. (NIPS 2004), which is a combination of boosting and graph mining. Instead of sequential multiplicative updates, we employ the linear programming boosting (LP) for regression. The LP approach allows to include inequality constraints for the parameter vector, which turns out to be particularly useful in QSAR tasks where activity values are sometimes unavailable. Furthermore, the efficiency is improved significantly by employing multiple pricing.

PDF PDF [BibTex]

PDF PDF [BibTex]


no image
Incremental Aspect Models for Mining Document Streams

Surendran, A., Sra, S.

In PKDD 2006, pages: 633-640, (Editors: Fürnkranz, J. , T. Scheffer, M. Spiliopoulou), Springer, Berlin, Germany, 10th European Conference on Principles and Practice of Knowledge Discovery in Databases, September 2006 (inproceedings)

Abstract
In this paper we introduce a novel approach for incrementally building aspect models, and use it to dynamically discover underlying themes from document streams. Using the new approach we present an application which we call “query-line tracking” i.e., we automatically discover and summarize different themes or stories that appear over time, and that relate to a particular query. We present evaluation on news corpora to demonstrate the strength of our method for both query-line tracking, online indexing and clustering.

Web DOI [BibTex]

Web DOI [BibTex]


no image
Learning Eye Movements

Kienzle, W., Wichmann, F., Schölkopf, B., Franz, M.

Sensory Coding And The Natural Environment, 2006, pages: 1, September 2006 (poster)

Abstract
The human visual system samples images through saccadic eye movements which rapidly change the point of fixation. Although the selection of eye movement targets depends on numerous top-down mechanisms, a number of recent studies have shown that low-level image features such as local contrast or edges play an important role. These studies typically used predefined image features which were afterwards experimentally verified. Here, we follow a complementary approach: instead of testing a set of candidate image features, we infer these hypotheses from the data, using methods from statistical learning. To this end, we train a non-linear classifier on fixated vs. randomly selected image patches without making any physiological assumptions. The resulting classifier can be essentially characterized by a nonlinear combination of two center-surround receptive fields. We find that the prediction performance of this simple model on our eye movement data is indistinguishable from the physiologically motivated model of Itti & Koch (2000) which is far more complex. In particular, we obtain a comparable performance without using any multi-scale representations, long-range interactions or oriented image features.

Web [BibTex]

Web [BibTex]