Header logo is ei


2020


no image
Algorithmic recourse under imperfect causal knowledge: a probabilistic approach

Karimi*, A., von Kügelgen*, J., Schölkopf, B., Valera, I.

Advances in Neural Information Processing Systems 33, 34th Annual Conference on Neural Information Processing Systems, December 2020, *equal contribution (conference) Accepted

arXiv [BibTex]

2020

arXiv [BibTex]


Grasping Field: Learning Implicit Representations for Human Grasps
Grasping Field: Learning Implicit Representations for Human Grasps

Karunratanakul, K., Yang, J., Zhang, Y., Black, M., Muandet, K., Tang, S.

In International Conference on 3D Vision (3DV), November 2020 (inproceedings)

Abstract
Robotic grasping of house-hold objects has made remarkable progress in recent years. Yet, human grasps are still difficult to synthesize realistically. There are several key reasons: (1) the human hand has many degrees of freedom (more than robotic manipulators); (2) the synthesized hand should conform to the surface of the object; and (3) it should interact with the object in a semantically and physically plausible manner. To make progress in this direction, we draw inspiration from the recent progress on learning-based implicit representations for 3D object reconstruction. Specifically, we propose an expressive representation for human grasp modelling that is efficient and easy to integrate with deep neural networks. Our insight is that every point in a three-dimensional space can be characterized by the signed distances to the surface of the hand and the object, respectively. Consequently, the hand, the object, and the contact area can be represented by implicit surfaces in a common space, in which the proximity between the hand and the object can be modelled explicitly. We name this 3D to 2D mapping as Grasping Field, parameterize it with a deep neural network, and learn it from data. We demonstrate that the proposed grasping field is an effective and expressive representation for human grasp generation. Specifically, our generative model is able to synthesize high-quality human grasps, given only on a 3D object point cloud. The extensive experiments demonstrate that our generative model compares favorably with a strong baseline and approaches the level of natural human grasps. Furthermore, based on the grasping field representation, we propose a deep network for the challenging task of 3D hand-object interaction reconstruction from a single RGB image. Our method improves the physical plausibility of the hand-object contact reconstruction and achieves comparable performance for 3D hand reconstruction compared to state-of-the-art methods. Our model and code are available for research purpose at https://github.com/korrawe/grasping_field.

pdf arXiv code [BibTex]


no image
MYND: Unsupervised Evaluation of Novel BCI Control Strategies on Consumer Hardware

Hohmann, M. R., Konieczny, L., Hackl, M., Wirth, B., Zaman, T., Enficiaud, R., Grosse-Wentrup, M., Schölkopf, B.

Proceedings of the 33rd Annual ACM Symposium on User Interface Software and Technology (UIST), October 2020 (conference) Accepted

arXiv DOI [BibTex]

arXiv DOI [BibTex]


no image
Model-Agnostic Counterfactual Explanations for Consequential Decisions

Karimi, A., Barthe, G., Balle, B., Valera, I.

Proceedings of the 23rd International Conference on Artificial Intelligence and Statistics (AISTATS), 108, pages: 895-905, Proceedings of Machine Learning Research, (Editors: Silvia Chiappa and Roberto Calandra), PMLR, August 2020 (conference)

arXiv link (url) [BibTex]

arXiv link (url) [BibTex]


no image
More Powerful Selective Kernel Tests for Feature Selection

Lim, J. N., Yamada, M., Jitkrittum, W., Terada, Y., Matsui, S., Shimodaira, H.

Proceedings of the 23rd International Conference on Artificial Intelligence and Statistics (AISTATS), 108, pages: 820-830, Proceedings of Machine Learning Research, (Editors: Silvia Chiappa and Roberto Calandra), PMLR, August 2020 (conference)

arXiv link (url) [BibTex]

arXiv link (url) [BibTex]


no image
Bayesian Online Prediction of Change Points

Agudelo-España, D., Gomez-Gonzalez, S., Bauer, S., Schölkopf, B., Peters, J.

Proceedings of the 36th International Conference on Uncertainty in Artificial Intelligence (UAI), 124, pages: 320-329, Proceedings of Machine Learning Research, (Editors: Jonas Peters and David Sontag), PMLR, August 2020 (conference)

link (url) [BibTex]

link (url) [BibTex]


no image
Semi-supervised learning, causality, and the conditional cluster assumption

von Kügelgen, J., Mey, A., Loog, M., Schölkopf, B.

Proceedings of the 36th International Conference on Uncertainty in Artificial Intelligence (UAI) , 124, pages: 1-10, Proceedings of Machine Learning Research, (Editors: Jonas Peters and David Sontag), PMLR, August 2020 (conference)

link (url) [BibTex]

link (url) [BibTex]


no image
Kernel Conditional Moment Test via Maximum Moment Restriction

Muandet, K., Jitkrittum, W., Kübler, J. M.

Proceedings of the 36th International Conference on Uncertainty in Artificial Intelligence (UAI), 124, pages: 41-50, Proceedings of Machine Learning Research, (Editors: Jonas Peters and David Sontag), PMLR, August 2020 (conference)

link (url) [BibTex]

link (url) [BibTex]


no image
On the design of consequential ranking algorithms

Tabibian, B., Gómez, V., De, A., Schölkopf, B., Gomez Rodriguez, M.

Proceedings of the 36th International Conference on Uncertainty in Artificial Intelligence (UAI), 124, pages: 171-180, Proceedings of Machine Learning Research, (Editors: Jonas Peters and David Sontag), PMLR, August 2020 (conference)

link (url) [BibTex]

link (url) [BibTex]


no image
Importance Sampling via Local Sensitivity

Raj, A., Musco, C., Mackey, L.

Proceedings of the 23rd International Conference on Artificial Intelligence and Statistics (AISTATS), 108, pages: 3099-3109, Proceedings of Machine Learning Research, (Editors: Silvia Chiappa and Roberto Calandra), PMLR, August 2020 (conference)

link (url) [BibTex]

link (url) [BibTex]


no image
A Continuous-time Perspective for Modeling Acceleration in Riemannian Optimization

F Alimisis, F., Orvieto, A., Becigneul, G., Lucchi, A.

Proceedings of the 23rd International Conference on Artificial Intelligence and Statistics (AISTATS), 108, pages: 1297-1307, Proceedings of Machine Learning Research, (Editors: Silvia Chiappa and Roberto Calandra), PMLR, August 2020 (conference)

link (url) [BibTex]

link (url) [BibTex]


no image
Fair Decisions Despite Imperfect Predictions

Kilbertus, N., Gomez Rodriguez, M., Schölkopf, B., Muandet, K., Valera, I.

Proceedings of the 23rd International Conference on Artificial Intelligence and Statistics (AISTATS), 108, pages: 277-287, Proceedings of Machine Learning Research, (Editors: Silvia Chiappa and Roberto Calandra), PMLR, August 2020 (conference)

link (url) [BibTex]

link (url) [BibTex]


no image
Integrals over Gaussians under Linear Domain Constraints

Gessner, A., Kanjilal, O., Hennig, P.

Proceedings of the 23rd International Conference on Artificial Intelligence and Statistics (AISTATS), 108, pages: 2764-2774, Proceedings of Machine Learning Research, (Editors: Silvia Chiappa and Roberto Calandra), PMLR, August 2020 (conference)

link (url) [BibTex]

link (url) [BibTex]


no image
Modular Block-diagonal Curvature Approximations for Feedforward Architectures

Dangel, F., Harmeling, S., Hennig, P.

Proceedings of the 23rd International Conference on Artificial Intelligence and Statistics (AISTATS), 108, pages: 799-808, Proceedings of Machine Learning Research, (Editors: Silvia Chiappa and Roberto Calandra), PMLR, August 2020 (conference)

link (url) [BibTex]

link (url) [BibTex]


no image
Testing Goodness of Fit of Conditional Density Models with Kernels

Jitkrittum, W., Kanagawa, H., Schölkopf, B.

Proceedings of the 36th International Conference on Uncertainty in Artificial Intelligence (UAI), 124, pages: 221-230, Proceedings of Machine Learning Research, (Editors: Jonas Peters and David Sontag), PMLR, August 2020 (conference)

link (url) [BibTex]

link (url) [BibTex]


no image
Stochastic Frank-Wolfe for Constrained Finite-Sum Minimization

Negiar, G., Dresdner, G., Tsai, A. Y., El Ghaoui, L., Locatello, F., Freund, R. M., Pedregosa, F.

37th International Conference on Machine Learning (ICML), pages: 296-305, July 2020 (conference)

link (url) [BibTex]

link (url) [BibTex]


no image
Variational Autoencoders with Riemannian Brownian Motion Priors

Kalatzis, D., Eklund, D., Arvanitidis, G., Hauberg, S.

37th International Conference on Machine Learning (ICML), pages: 6789-6799, July 2020 (conference)

link (url) [BibTex]

link (url) [BibTex]


no image
Variational Bayes in Private Settings (VIPS) (Extended Abstract)

Foulds, J. R., Park, M., Chaudhuri, K., Welling, M.

Proceedings of the 29th International Joint Conference on Artificial Intelligence, (IJCAI-PRICAI), pages: 5050-5054, (Editors: Christian Bessiere), International Joint Conferences on Artificial Intelligence Organization, July 2020, Journal track (conference)

link (url) DOI [BibTex]

link (url) DOI [BibTex]


no image
Weakly-Supervised Disentanglement Without Compromises

Locatello, F., Poole, B., Rätsch, G., Schölkopf, B., Bachem, O., Tschannen, M.

37th International Conference on Machine Learning (ICML), pages: 7753-7764, July 2020 (conference)

link (url) [BibTex]

link (url) [BibTex]


no image
Constant Curvature Graph Convolutional Networks

Bachmann*, G., Becigneul*, G., Ganea, O.

37th International Conference on Machine Learning (ICML), pages: 9118-9128, July 2020, *equal contribution (conference)

link (url) [BibTex]

link (url) [BibTex]


no image
Being Bayesian, Even Just a Bit, Fixes Overconfidence in ReLU Networks

Kristiadi, A., Hein, M., Hennig, P.

37th International Conference on Machine Learning (ICML), pages: 1226-1236, July 2020 (conference)

link (url) [BibTex]

link (url) [BibTex]


no image
Differentiable Likelihoods for Fast Inversion of ‘Likelihood-Free’ Dynamical Systems

Kersting, H., Krämer, N., Schiegg, M., Daniel, C., Tiemann, M., Hennig, P.

37th International Conference on Machine Learning (ICML), pages: 2655-2665, July 2020 (conference)

link (url) [BibTex]

link (url) [BibTex]


no image
Kernel Conditional Density Operators

Schuster, I., Mollenhauer, M., Klus, S., Muandet, K.

Proceedings of the 23rd International Conference on Artificial Intelligence and Statistics (AISTATS), 108, pages: 993-1004, Proceedings of Machine Learning Research, (Editors: Silvia Chiappa and Roberto Calandra), PMLR, June 2020 (conference)

link (url) [BibTex]

link (url) [BibTex]


no image
A Kernel Mean Embedding Approach to Reducing Conservativeness in Stochastic Programming and Control

Zhu, J., Diehl, M., Schölkopf, B.

2nd Annual Conference on Learning for Dynamics and Control (L4DC), 120, pages: 915-923, Proceedings of Machine Learning Research, (Editors: Alexandre M. Bayen and Ali Jadbabaie and George Pappas and Pablo A. Parrilo and Benjamin Recht and Claire Tomlin and Melanie Zeilinger), PMLR, June 2020 (conference)

arXiv link (url) [BibTex]

arXiv link (url) [BibTex]


no image
Disentangling Factors of Variations Using Few Labels

Locatello, F., Tschannen, M., Bauer, S., Rätsch, G., Schölkopf, B., Bachem, O.

8th International Conference on Learning Representations (ICLR), April 2020 (conference)

arXiv link (url) [BibTex]

arXiv link (url) [BibTex]


no image
Mixed-curvature Variational Autoencoders

Skopek, O., Ganea, O., Becigneul, G.

8th International Conference on Learning Representations (ICLR), April 2020 (conference)

link (url) [BibTex]

link (url) [BibTex]


Non-linear interlinkages and key objectives amongst the Paris Agreement and the Sustainable Development Goals
Non-linear interlinkages and key objectives amongst the Paris Agreement and the Sustainable Development Goals

Laumann, F., von Kügelgen, J., Barahona, M.

ICLR 2020 Workshop "Tackling Climate Change with Machine Learning", April 2020 (conference)

arXiv PDF [BibTex]

arXiv PDF [BibTex]


no image
Counterfactuals uncover the modular structure of deep generative models

Besserve, M., Mehrjou, A., Sun, R., Schölkopf, B.

8th International Conference on Learning Representations (ICLR), April 2020 (conference)

link (url) [BibTex]

link (url) [BibTex]


Towards causal generative scene models via competition of experts
Towards causal generative scene models via competition of experts

von Kügelgen*, J., Ustyuzhaninov*, I., Gehler, P., Bethge, M., Schölkopf, B.

ICLR 2020 Workshop "Causal Learning for Decision Making", April 2020, *equal contribution (conference)

arXiv PDF [BibTex]

arXiv PDF [BibTex]


no image
On Mutual Information Maximization for Representation Learning

Tschannen, M., Djolonga, J., Rubenstein, P. K., Gelly, S., Lucic, M.

8th International Conference on Learning Representations (ICLR), April 2020 (conference)

arXiv link (url) [BibTex]

arXiv link (url) [BibTex]


From Variational to Deterministic Autoencoders
From Variational to Deterministic Autoencoders

Ghosh*, P., Sajjadi*, M. S. M., Vergari, A., Black, M. J., Schölkopf, B.

8th International Conference on Learning Representations (ICLR) , April 2020, *equal contribution (conference)

Abstract
Variational Autoencoders (VAEs) provide a theoretically-backed framework for deep generative models. However, they often produce “blurry” images, which is linked to their training objective. Sampling in the most popular implementation, the Gaussian VAE, can be interpreted as simply injecting noise to the input of a deterministic decoder. In practice, this simply enforces a smooth latent space structure. We challenge the adoption of the full VAE framework on this specific point in favor of a simpler, deterministic one. Specifically, we investigate how substituting stochasticity with other explicit and implicit regularization schemes can lead to a meaningful latent space without having to force it to conform to an arbitrarily chosen prior. To retrieve a generative mechanism for sampling new data points, we propose to employ an efficient ex-post density estimation step that can be readily adopted both for the proposed deterministic autoencoders as well as to improve sample quality of existing VAEs. We show in a rigorous empirical study that regularized deterministic autoencoding achieves state-of-the-art sample quality on the common MNIST, CIFAR-10 and CelebA datasets.

arXiv link (url) [BibTex]

arXiv link (url) [BibTex]


no image
Radial and Directional Posteriors for Bayesian Deep Learning

Oh, C., Adamczewski, K., Park, M.

Proceedings of the 34th Conference on Artificial Intelligence (AAAI), 34(4):5298-5305, AAAI Press, Febuary 2020, AAAI Technical Track: Machine Learning (conference)

link (url) DOI [BibTex]

link (url) DOI [BibTex]


no image
ODIN: ODE-Informed Regression for Parameter and State Inference in Time-Continuous Dynamical Systems

Wenk, P., Abbati, G., Osborne, M. A., Schölkopf, B., Krause, A., Bauer, S.

Proceedings of the 34th Conference on Artificial Intelligence (AAAI), 34(4):6364-6371, AAAI Press, Febuary 2020, AAAI Technical Track: Machine Learning (conference)

link (url) DOI [BibTex]

link (url) DOI [BibTex]


no image
Interpretable and Differentially Private Predictions

Harder, F., Bauer, M., Park, M.

Proceedings of the 34th Conference on Artificial Intelligence (AAAI), 34(4):4083-4090, AAAI Press, Febuary 2020, AAAI Technical Track: Machine Learning (conference)

link (url) DOI [BibTex]

link (url) DOI [BibTex]


no image
A Commentary on the Unsupervised Learning of Disentangled Representations

Locatello, F., Bauer, S., Lucic, M., Rätsch, G., Gelly, S., Schölkopf, B., Bachem, O.

Proceedings of the 34th Conference on Artificial Intelligence (AAAI), 34(9):13681-13684, AAAI Press, Febuary 2020, Sister Conference Track (conference)

link (url) DOI [BibTex]

link (url) DOI [BibTex]


no image
A Real-Robot Dataset for Assessing Transferability of Learned Dynamics Models

Agudelo-España, D., Zadaianchuk, A., Wenk, P., Garg, A., Akpo, J., Grimminger, F., Viereck, J., Naveau, M., Righetti, L., Martius, G., Krause, A., Schölkopf, B., Bauer, S., Wüthrich, M.

IEEE International Conference on Robotics and Automation (ICRA), 2020 (conference) Accepted

Project Page PDF [BibTex]

Project Page PDF [BibTex]


Worst-Case Risk Quantification under Distributional Ambiguity using Kernel Mean Embedding in Moment Problem
Worst-Case Risk Quantification under Distributional Ambiguity using Kernel Mean Embedding in Moment Problem

Zhu, J., Jitkrittum, W., Diehl, M., Schölkopf, B.

In 59th IEEE Conference on Decision and Control (CDC), 2020 (inproceedings) Accepted

[BibTex]

[BibTex]


no image
Divide-and-Conquer Monte Carlo Tree Search for goal directed planning

Parascandolo*, G., Buesing*, L., Merel, J., Hasenclever, L., Aslanides, J., Hamrick, J. B., Heess, N., Neitz, A., Weber, T.

2020, *equal contribution (conference) Submitted

arXiv [BibTex]

arXiv [BibTex]

2011


no image
Statistical estimation for optimization problems on graphs

Langovoy, M., Sra, S.

In pages: 1-6, NIPS Workshop on Discrete Optimization in Machine Learning (DISCML): Uncertainty, Generalization and Feedback , December 2011 (inproceedings)

Abstract
Large graphs abound in machine learning, data mining, and several related areas. A useful step towards analyzing such graphs is that of obtaining certain summary statistics — e.g., or the expected length of a shortest path between two nodes, or the expected weight of a minimum spanning tree of the graph, etc. These statistics provide insight into the structure of a graph, and they can help predict global properties of a graph. Motivated thus, we propose to study statistical properties of structured subgraphs (of a given graph), in particular, to estimate the expected objective function value of a combinatorial optimization problem over these subgraphs. The general task is very difficult, if not unsolvable; so for concreteness we describe a more specific statistical estimation problem based on spanning trees. We hope that our position paper encourages others to also study other types of graphical structures for which one can prove nontrivial statistical estimates.

PDF Web [BibTex]

2011

PDF Web [BibTex]


no image
On the discardability of data in Support Vector Classification problems

Del Favero, S., Varagnolo, D., Dinuzzo, F., Schenato, L., Pillonetto, G.

In pages: 3210-3215, IEEE, Piscataway, NJ, USA, 50th IEEE Conference on Decision and Control and European Control Conference (CDC - ECC), December 2011 (inproceedings)

Abstract
We analyze the problem of data sets reduction for support vector classification. The work is also motivated by distributed problems, where sensors collect binary measurements at different locations moving inside an environment that needs to be divided into a collection of regions labeled in two different ways. The scope is to let each agent retain and exchange only those measurements that are mostly informative for the collective reconstruction of the decision boundary. For the case of separable classes, we provide the exact conditions and an efficient algorithm to determine if an element in the training set can become a support vector when new data arrive. The analysis is then extended to the non-separable case deriving a sufficient discardability condition and a general data selection scheme for classification. Numerical experiments relative to the distributed problem show that the proposed procedure allows the agents to exchange a small amount of the collected data to obtain a highly predictive decision boundary.

PDF Web DOI [BibTex]

PDF Web DOI [BibTex]


no image
Information, learning and falsification

Balduzzi, D.

In pages: 1-4, NIPS Philosophy and Machine Learning Workshop, December 2011 (inproceedings)

Abstract
There are (at least) three approaches to quantifying information. The first, algorithmic information or Kolmogorov complexity, takes events as strings and, given a universal Turing machine, quantifies the information content of a string as the length of the shortest program producing it [1]. The second, Shannon information, takes events as belonging to ensembles and quantifies the information resulting from observing the given event in terms of the number of alternate events that have been ruled out [2]. The third, statistical learning theory, has introduced measures of capacity that control (in part) the expected risk of classifiers [3]. These capacities quantify the expectations regarding future data that learning algorithms embed into classifiers. Solomonoff and Hutter have applied algorithmic information to prove remarkable results on universal induction. Shannon information provides the mathematical foundation for communication and coding theory. However, both approaches have shortcomings. Algorithmic information is not computable, severely limiting its practical usefulness. Shannon information refers to ensembles rather than actual events: it makes no sense to compute the Shannon information of a single string – or rather, there are many answers to this question depending on how a related ensemble is constructed. Although there are asymptotic results linking algorithmic and Shannon information, it is unsatisfying that there is such a large gap – a difference in kind – between the two measures. This note describes a new method of quantifying information, effective information, that links algorithmic information to Shannon information, and also links both to capacities arising in statistical learning theory [4, 5]. After introducing the measure, we show that it provides a non-universal analog of Kolmogorov complexity. We then apply it to derive basic capacities in statistical learning theory: empirical VC-entropy and empirical Rademacher complexity. A nice byproduct of our approach is an interpretation of the explanatory power of a learning algorithm in terms of the number of hypotheses it falsifies [6], counted in two different ways for the two capacities. We also discuss how effective information relates to information gain, Shannon and mutual information.

PDF Web [BibTex]

PDF Web [BibTex]


no image
A general linear non-Gaussian state-space model: Identifiability, identification, and applications

Zhang, K., Hyvärinen, A.

In JMLR Workshop and Conference Proceedings Volume 20, pages: 113-128, (Editors: Hsu, C.-N. , W.S. Lee ), MIT Press, Cambridge, MA, USA, 3rd Asian Conference on Machine Learning (ACML), November 2011 (inproceedings)

Abstract
State-space modeling provides a powerful tool for system identification and prediction. In linear state-space models the data are usually assumed to be Gaussian and the models have certain structural constraints such that they are identifiable. In this paper we propose a non-Gaussian state-space model which does not have such constraints. We prove that this model is fully identifiable. We then propose an efficient two-step method for parameter estimation: one first extracts the subspace of the latent processes based on the temporal information of the data, and then performs multichannel blind deconvolution, making use of both the temporal information and non-Gaussianity. We conduct a series of simulations to illustrate the performance of the proposed method. Finally, we apply the proposed model and parameter estimation method on real data, including major world stock indices and magnetoencephalography (MEG) recordings. Experimental results are encouraging and show the practical usefulness of the proposed model and method.

PDF Web [BibTex]

PDF Web [BibTex]


no image
Non-stationary correction of optical aberrations

Schuler, C., Hirsch, M., Harmeling, S., Schölkopf, B.

In pages: 659-666 , (Editors: DN Metaxas and L Quan and A Sanfeliu and LJ Van Gool), IEEE, Piscataway, NJ, USA, 13th IEEE International Conference on Computer Vision (ICCV), November 2011 (inproceedings)

Abstract
Taking a sharp photo at several megapixel resolution traditionally relies on high grade lenses. In this paper, we present an approach to alleviate image degradations caused by imperfect optics. We rely on a calibration step to encode the optical aberrations in a space-variant point spread function and obtain a corrected image by non-stationary deconvolution. By including the Bayer array in our image formation model, we can perform demosaicing as part of the deconvolution.

PDF Web DOI [BibTex]

PDF Web DOI [BibTex]


no image
Learning low-rank output kernels

Dinuzzo, F., Fukumizu, K.

In JMLR Workshop and Conference Proceedings Volume 20, pages: 181-196, (Editors: Hsu, C.-N. , W.S. Lee), JMLR, Cambridge, MA, USA, 3rd Asian Conference on Machine Learning (ACML) , November 2011 (inproceedings)

Abstract
Output kernel learning techniques allow to simultaneously learn a vector-valued function and a positive semidefinite matrix which describes the relationships between the outputs. In this paper, we introduce a new formulation that imposes a low-rank constraint on the output kernel and operates directly on a factor of the kernel matrix. First, we investigate the connection between output kernel learning and a regularization problem for an architecture with two layers. Then, we show that a variety of methods such as nuclear norm regularized regression, reduced-rank regression, principal component analysis, and low rank matrix approximation can be seen as special cases of the output kernel learning framework. Finally, we introduce a block coordinate descent strategy for learning low-rank output kernels.

PDF Web [BibTex]

PDF Web [BibTex]


no image
Fast removal of non-uniform camera shake

Hirsch, M., Schuler, C., Harmeling, S., Schölkopf, B.

In pages: 463-470 , (Editors: DN Metaxas and L Quan and A Sanfeliu and LJ Van Gool), IEEE, Piscataway, NJ, USA, 13th IEEE International Conference on Computer Vision (ICCV), November 2011 (inproceedings)

Abstract
Camera shake leads to non-uniform image blurs. State-of-the-art methods for removing camera shake model the blur as a linear combination of homographically transformed versions of the true image. While this is conceptually interesting, the resulting algorithms are computationally demanding. In this paper we develop a forward model based on the efficient filter flow framework, incorporating the particularities of camera shake, and show how an efficient algorithm for blur removal can be obtained. Comprehensive comparisons on a number of real-world blurry images show that our approach is not only substantially faster, but it also leads to better deblurring results.

PDF Web DOI [BibTex]

PDF Web DOI [BibTex]


no image
Stability Condition for Teleoperation System with Packet Loss

Hong, A., Cho, JH., Lee, DY.

In pages: 760-761, 2011 KSME Annual Fall Conference, November 2011 (inproceedings)

Abstract
This paper focuses on the stability condition of teleoperation system where there is a packet loss in communication channel. Communication channel between master and slave cause packet loss and it obviously leads to a performance degradation and instability of teleoperation system. We consider two-channel control architecture for teleoperation system, and control inputs to remote site are produced by position of master and slave. In this paper, teleoperation system is modeled in discrete domain to include packet loss process. Also, the stability condition for teleoperation system with packet loss is discussed with input-to-state stability. Finally, the stability condition is presented in LMI approach.

[BibTex]

[BibTex]


no image
Attenuation correction in MR-BrainPET with segmented T1-weighted MR images of the patient’s head: A comparative study with CT

Wagenknecht, G., Rota Kops, E., Mantlik, F., Fried, E., Pilz, T., Hautzel, H., Tellmann, L., Pichler, B., Herzog, H.

In pages: 2261-2266 , IEEE, Piscataway, NJ, USA, IEEE Nuclear Science Symposium and Medical Imaging Conference (NSS/MIC), October 2011 (inproceedings)

Abstract
Our method for attenuation correction (AC) in MR-BrainPET with segmented T1-weighted MR images of the pa-tient's head was applied to data from different MR-BrainPET scanners (Jülich, Tübingen) and compared to CT-based results. The study objectives presented in this paper are twofold. The first objective is to examine if the segmentation method developed for and successfully applied to 3D MP-RAGE data can also be used to segment other T1-weighted MR data such as 3D FLASH data. The second aim is to show if the similarity of segmented MR-based (SBA) and CT-based AC (CBA) obtained at HR+ PET can also be confirmed for BrainPET for which the new AC method is intended for. In order to reach the first objective, 14 segmented MR data sets (three 3D MP-RAGE data sets from Jülich and eleven 3D FLASH data sets from Tubingen) were compared to the resp. CT data based on the Dice coefficient and scatter plots. For bone, a CT threshold HU>;500 was applied. Dice coefficients (mean±std) for the upper cranial part of the skull, the skull above cavities, and in the caudal part including the cerebellum are 0.73±0.1, 0.79±0.04, and 0.49±0.02 for the Jülich data and 0.7U0.1, 0.72±0.1, and 0.60±0.05 for the Tubingen data. To reach the second aim, SBA and CBA were compared for six subjects based on VOI (AAL atlas) analysis. Mean absolute relative difference (maRD) values are maRD(JUFVBWl-FDG): 0.99%±0.83%, maRD(JüFVBW2-FDG): 0.90%±0.89%, and maRD(JUEP-Fluma- zenil): 1.85%±1.25% for the Jülich data and maRD(TuTP02- FDG): 2.99%±1.65%, maRD(TuNP01-FDG): 5.37%±2.29%, and maRD(TuNP02-FDG): 6.52%±1.69% for the three best-segmented Tübingen data sets. The results show similar segmentation quality for both Tl- weighted MR sequence types. The application to AC in BrainPET - hows a high similarity to CT-based AC if the standardized ACF value for bone used in SBA is in good accordance to the bone density of the patient in question.

Web DOI [BibTex]

Web DOI [BibTex]