Header logo is


2016


Thumb xl untitled
Probabilistic Approximate Least-Squares

Bartels, S., Hennig, P.

Proceedings of the 19th International Conference on Artificial Intelligence and Statistics (AISTATS), 51, pages: 676-684, JMLR Workshop and Conference Proceedings, (Editors: Gretton, A. and Robert, C. C. ), May 2016 (conference)

Abstract
Least-squares and kernel-ridge / Gaussian process regression are among the foundational algorithms of statistics and machine learning. Famously, the worst-case cost of exact nonparametric regression grows cubically with the data-set size; but a growing number of approximations have been developed that estimate good solutions at lower cost. These algorithms typically return point estimators, without measures of uncertainty. Leveraging recent results casting elementary linear algebra operations as probabilistic inference, we propose a new approximate method for nonparametric least-squares that affords a probabilistic uncertainty estimate over the error between the approximate and exact least-squares solution (this is not the same as the posterior variance of the associated Gaussian process regressor). This allows estimating the error of the least-squares solution on a subset of the data relative to the full-data solution. The uncertainty can be used to control the computational effort invested in the approximation. Our algorithm has linear cost in the data-set size, and a simple formal form, so that it can be implemented with a few lines of code in programming languages with linear algebra functionality.

ei pn

link (url) Project Page Project Page [BibTex]

2016


link (url) Project Page Project Page [BibTex]


no image
Learning soft task priorities for control of redundant robots

Modugno, V., Neumann, G., Rueckert, E., Oriolo, G., Peters, J., Ivaldi, S.

IEEE International Conference on Robotics and Automation (ICRA), pages: 221-226, IEEE, May 2016 (conference)

ei

DOI [BibTex]

DOI [BibTex]


no image
On the Reliability of Information and Trustworthiness of Web Sources in Wikipedia

Tabibian, B., Farajtabar, M., Valera, I., Song, L., Schölkopf, B., Gomez Rodriguez, M.

Wikipedia workshop at the 10th International AAAI Conference on Web and Social Media (ICWSM), May 2016 (conference)

ei

[BibTex]

[BibTex]


Thumb xl 2016 peer grading
Peer Grading in a Course on Algorithms and Data Structures: Machine Learning Algorithms do not Improve over Simple Baselines

Sajjadi, M. S. M., Alamgir, M., von Luxburg, U.

Proceedings of the 3rd ACM conference on Learning @ Scale, pages: 369-378, (Editors: Haywood, J. and Aleven, V. and Kay, J. and Roll, I.), ACM, L@S, April 2016, (An earlier version of this paper had been presented at the ICML 2015 workshop for Machine Learning for Education.) (conference)

ei

Arxiv Peer-Grading dataset request [BibTex]

Arxiv Peer-Grading dataset request [BibTex]


no image
On estimation of functional causal models: General results and application to post-nonlinear causal model

Zhang, K., Wang, Z., Zhang, J., Schölkopf, B.

ACM Transactions on Intelligent Systems and Technologies, 7(2):article no. 13, January 2016 (article)

ei

PDF DOI [BibTex]

PDF DOI [BibTex]


no image
Fabular: Regression Formulas As Probabilistic Programming

Borgström, J., Gordon, A. D., Ouyang, L., Russo, C., Ścibior, A., Szymczak, M.

Proceedings of the 43rd Annual ACM SIGPLAN-SIGACT Symposium on Principles of Programming Languages (POPL), pages: 271-283, POPL ’16, ACM, January 2016 (conference)

ei

DOI Project Page [BibTex]

DOI Project Page [BibTex]


Thumb xl cloud tracking
Gaussian Process-Based Predictive Control for Periodic Error Correction

Klenske, E. D., Zeilinger, M., Schölkopf, B., Hennig, P.

IEEE Transactions on Control Systems Technology , 24(1):110-121, 2016 (article)

ei pn

PDF DOI [BibTex]

PDF DOI [BibTex]


no image
Pymanopt: A Python Toolbox for Optimization on Manifolds using Automatic Differentiation

Townsend, J., Koep, N., Weichwald, S.

Journal of Machine Learning Research, 17(137):1-5, 2016 (article)

ei

PDF Arxiv Code Project page link (url) [BibTex]


no image
A Causal, Data-driven Approach to Modeling the Kepler Data

Wang, D., Hogg, D. W., Foreman-Mackey, D., Schölkopf, B.

Publications of the Astronomical Society of the Pacific, 128(967):094503, 2016 (article)

ei

link (url) DOI Project Page [BibTex]

link (url) DOI Project Page [BibTex]


no image
Probabilistic Inference for Determining Options in Reinforcement Learning

Daniel, C., van Hoof, H., Peters, J., Neumann, G.

Machine Learning, Special Issue, 104(2):337-357, (Editors: Gärtner, T., Nanni, M., Passerini, A. and Robardet, C.), European Conference on Machine Learning im Machine Learning, Journal Track, 2016, Best Student Paper Award of ECML-PKDD 2016 (article)

am ei

DOI Project Page [BibTex]

DOI Project Page [BibTex]


no image
Influence of initial fixation position in scene viewing

Rothkegel, L. O. M., Trukenbrod, H. A., Schütt, H. H., Wichmann, F. A., Engbert, R.

Vision Research, 129, pages: 33-49, 2016 (article)

ei

link (url) DOI Project Page [BibTex]

link (url) DOI Project Page [BibTex]


no image
Testing models of peripheral encoding using metamerism in an oddity paradigm

Wallis, T. S. A., Bethge, M., Wichmann, F. A.

Journal of Vision, 16(2), 2016 (article)

ei

DOI Project Page [BibTex]

DOI Project Page [BibTex]


no image
Modeling Confounding by Half-Sibling Regression

Schölkopf, B., Hogg, D., Wang, D., Foreman-Mackey, D., Janzing, D., Simon-Gabriel, C. J., Peters, J.

Proceedings of the National Academy of Science, 113(27):7391-7398, 2016 (article)

ei

Code link (url) DOI Project Page [BibTex]

Code link (url) DOI Project Page [BibTex]


Thumb xl dual control sampled b
Dual Control for Approximate Bayesian Reinforcement Learning

Klenske, E. D., Hennig, P.

Journal of Machine Learning Research, 17(127):1-30, 2016 (article)

ei pn

PDF link (url) [BibTex]

PDF link (url) [BibTex]


no image
A Population Based Gaussian Mixture Model Incorporating 18F-FDG-PET and DW-MRI Quantifies Tumor Tissue Classes

Divine, M. R., Katiyar, P., Kohlhofer, U., Quintanilla-Martinez, L., Disselhorst, J. A., Pichler, B. J.

Journal of Nuclear Medicine, 57(3):473-479, 2016 (article)

ei

DOI [BibTex]

DOI [BibTex]


no image
Painfree and accurate Bayesian estimation of psychometric functions for (potentially) overdispersed data

Schütt, H. H., Harmeling, S., Macke, J. H., Wichmann, F. A.

Vision Research, 122, pages: 105-123, 2016 (article)

ei

link (url) DOI Project Page [BibTex]

link (url) DOI Project Page [BibTex]


no image
Hierarchical Relative Entropy Policy Search

Daniel, C., Neumann, G., Kroemer, O., Peters, J.

Journal of Machine Learning Research, 17(93):1-50, 2016 (article)

ei

link (url) Project Page [BibTex]

link (url) Project Page [BibTex]


no image
Kernel Mean Shrinkage Estimators

Muandet, K., Sriperumbudur, B., Fukumizu, K., Gretton, A., Schölkopf, B.

Journal of Machine Learning Research, 17(48):1-41, 2016 (article)

ei

link (url) [BibTex]

link (url) [BibTex]


no image
Learning to Deblur

Schuler, C. J., Hirsch, M., Harmeling, S., Schölkopf, B.

IEEE Transactions on Pattern Analysis and Machine Intelligence, 38(7):1439-1451, IEEE, 2016 (article)

ei

DOI [BibTex]

DOI [BibTex]


no image
Transfer Learning in Brain-Computer Interfaces

Jayaram, V., Alamgir, M., Altun, Y., Schölkopf, B., Grosse-Wentrup, M.

IEEE Computational Intelligence Magazine, 11(1):20-31, 2016 (article)

ei

PDF DOI [BibTex]

PDF DOI [BibTex]


no image
MERLiN: Mixture Effect Recovery in Linear Networks

Weichwald, S., Grosse-Wentrup, M., Gretton, A.

IEEE Journal of Selected Topics in Signal Processing, 10(7):1254-1266, 2016 (article)

ei

Arxiv Code PDF DOI Project Page [BibTex]

Arxiv Code PDF DOI Project Page [BibTex]


no image
Causal inference using invariant prediction: identification and confidence intervals

Peters, J., Bühlmann, P., Meinshausen, N.

Journal of the Royal Statistical Society, Series B (Statistical Methodology), 78(5):947-1012, 2016, (with discussion) (article)

ei

link (url) DOI [BibTex]

link (url) DOI [BibTex]


no image
Causal discovery and inference: concepts and recent methodological advances

Spirtes, P., Zhang, K.

Applied Informatics, 3(3):1-28, 2016 (article)

ei

DOI [BibTex]

DOI [BibTex]


no image
Self-regulation of brain rhythms in the precuneus: a novel BCI paradigm for patients with ALS

Fomina, T., Lohmann, G., Erb, M., Ethofer, T., Schölkopf, B., Grosse-Wentrup, M.

Journal of Neural Engineering, 13(6):066021, 2016 (article)

ei

link (url) Project Page [BibTex]


no image
Influence Estimation and Maximization in Continuous-Time Diffusion Networks

Gomez-Rodriguez, M., Song, L., Du, N., Zha, H., Schölkopf, B.

ACM Transactions on Information Systems, 34(2):9:1-9:33, 2016 (article)

ei

DOI Project Page Project Page [BibTex]

DOI Project Page Project Page [BibTex]


Thumb xl heteroscedasticgp
Modeling Variability of Musculoskeletal Systems with Heteroscedastic Gaussian Processes

Büchler, D., Calandra, R., Peters, J.

Workshop on Neurorobotics, Neural Information Processing Systems (NIPS), 2016 (conference)

am ei

NIPS16Neurorobotics [BibTex]

NIPS16Neurorobotics [BibTex]


no image
The population of long-period transiting exoplanets

Foreman-Mackey, D., Morton, T. D., Hogg, D. W., Agol, E., Schölkopf, B.

The Astronomical Journal, 152(6):206, 2016 (article)

ei

link (url) Project Page [BibTex]

link (url) Project Page [BibTex]


no image
An overview of quantitative approaches in Gestalt perception

Jäkel, F., Singh, M., Wichmann, F. A., Herzog, M. H.

Vision Research, 126, pages: 3-8, 2016 (article)

ei

link (url) DOI Project Page [BibTex]

link (url) DOI Project Page [BibTex]


no image
Causal and statistical learning

Schölkopf, B., Janzing, D., Lopez-Paz, D.

Oberwolfach Reports, 13(3):1896-1899, (Editors: A. Christmann and K. Jetter and S. Smale and D.-X. Zhou), 2016 (conference)

ei

DOI [BibTex]

DOI [BibTex]


no image
Bootstrat: Population Informed Bootstrapping for Rare Variant Tests

Huang, H., Peloso, G. M., Howrigan, D., Rakitsch, B., Simon-Gabriel, C. J., Goldstein, J. I., Daly, M. J., Borgwardt, K., Neale, B. M.

bioRxiv, 2016, preprint (article)

ei

link (url) DOI [BibTex]

link (url) DOI [BibTex]


no image
Probabilistic Movement Models Show that Postural Control Precedes and Predicts Volitional Motor Control

Rueckert, E., Camernik, J., Peters, J., Babic, J.

Nature PG: Scientific Reports, 6(Article number: 28455), 2016 (article)

ei

DOI Project Page [BibTex]

DOI Project Page [BibTex]


no image
Learning Taxonomy Adaptation in Large-scale Classification

Babbar, R., Partalas, I., Gaussier, E., Amini, M., Amblard, C.

Journal of Machine Learning Research, 17(98):1-37, 2016 (article)

ei

link (url) Project Page [BibTex]

link (url) Project Page [BibTex]


no image
BOiS—Berlin Object in Scene Database: Controlled Photographic Images for Visual Search Experiments with Quantified Contextual Priors

Mohr, J., Seyfarth, J., Lueschow, A., Weber, J. E., Wichmann, F. A., Obermayer, K.

Frontiers in Psychology, 2016 (article)

ei

DOI [BibTex]

DOI [BibTex]


no image
Preface to the ACM TIST Special Issue on Causal Discovery and Inference

Zhang, K., Li, J., Bareinboim, E., Schölkopf, B., Pearl, J.

ACM Transactions on Intelligent Systems and Technologies, 7(2):article no. 17, 2016 (article)

ei

DOI [BibTex]

DOI [BibTex]


no image
Recurrent Spiking Networks Solve Planning Tasks

Rueckert, E., Kappel, D., Tanneberg, D., Pecevski, D., Peters, J.

Nature PG: Scientific Reports, 6(Article number: 21142), 2016 (article)

ei

DOI Project Page [BibTex]

DOI Project Page [BibTex]


no image
Bio-inspired feedback-circuit implementation of discrete, free energy optimizing, winner-take-all computations

Genewein, T, Braun, DA

Biological Cybernetics, 110(2):135–150, June 2016 (article)

Abstract
Bayesian inference and bounded rational decision-making require the accumulation of evidence or utility, respectively, to transform a prior belief or strategy into a posterior probability distribution over hypotheses or actions. Crucially, this process cannot be simply realized by independent integrators, since the different hypotheses and actions also compete with each other. In continuous time, this competitive integration process can be described by a special case of the replicator equation. Here we investigate simple analog electric circuits that implement the underlying differential equation under the constraint that we only permit a limited set of building blocks that we regard as biologically interpretable, such as capacitors, resistors, voltage-dependent conductances and voltage- or current-controlled current and voltage sources. The appeal of these circuits is that they intrinsically perform normalization without requiring an explicit divisive normalization. However, even in idealized simulations, we find that these circuits are very sensitive to internal noise as they accumulate error over time. We discuss in how far neural circuits could implement these operations that might provide a generic competitive principle underlying both perception and action.

ei

DOI [BibTex]

DOI [BibTex]


no image
Decision-Making under Ambiguity Is Modulated by Visual Framing, but Not by Motor vs. Non-Motor Context: Experiments and an Information-Theoretic Ambiguity Model

Grau-Moya, J, Ortega, PA, Braun, DA

PLoS ONE, 11(4):1-21, April 2016 (article)

Abstract
A number of recent studies have investigated differences in human choice behavior depending on task framing, especially comparing economic decision-making to choice behavior in equivalent sensorimotor tasks. Here we test whether decision-making under ambiguity exhibits effects of task framing in motor vs. non-motor context. In a first experiment, we designed an experience-based urn task with varying degrees of ambiguity and an equivalent motor task where subjects chose between hitting partially occluded targets. In a second experiment, we controlled for the different stimulus design in the two tasks by introducing an urn task with bar stimuli matching those in the motor task. We found ambiguity attitudes to be mainly influenced by stimulus design. In particular, we found that the same subjects tended to be ambiguity-preferring when choosing between ambiguous bar stimuli, but ambiguity-avoiding when choosing between ambiguous urn sample stimuli. In contrast, subjects’ choice pattern was not affected by changing from a target hitting task to a non-motor context when keeping the stimulus design unchanged. In both tasks subjects’ choice behavior was continuously modulated by the degree of ambiguity. We show that this modulation of behavior can be explained by an information-theoretic model of ambiguity that generalizes Bayes-optimal decision-making by combining Bayesian inference with robust decision-making under model uncertainty. Our results demonstrate the benefits of information-theoretic models of decision-making under varying degrees of ambiguity for a given context, but also demonstrate the sensitivity of ambiguity attitudes across contexts that theoretical models struggle to explain.

ei

DOI [BibTex]

2005


no image
Kernel Methods for Measuring Independence

Gretton, A., Herbrich, R., Smola, A., Bousquet, O., Schölkopf, B.

Journal of Machine Learning Research, 6, pages: 2075-2129, December 2005 (article)

Abstract
We introduce two new functionals, the constrained covariance and the kernel mutual information, to measure the degree of independence of random variables. These quantities are both based on the covariance between functions of the random variables in reproducing kernel Hilbert spaces (RKHSs). We prove that when the RKHSs are universal, both functionals are zero if and only if the random variables are pairwise independent. We also show that the kernel mutual information is an upper bound near independence on the Parzen window estimate of the mutual information. Analogous results apply for two correlation-based dependence functionals introduced earlier: we show the kernel canonical correlation and the kernel generalised variance to be independence measures for universal kernels, and prove the latter to be an upper bound on the mutual information near independence. The performance of the kernel dependence functionals in measuring independence is verified in the context of independent component analysis.

ei

PDF PostScript PDF [BibTex]

2005


PDF PostScript PDF [BibTex]


no image
Kernel ICA for Large Scale Problems

Jegelka, S., Gretton, A., Achlioptas, D.

In pages: -, NIPS Workshop on Large Scale Kernel Machines, December 2005 (inproceedings)

ei

Web [BibTex]

Web [BibTex]


no image
A Unifying View of Sparse Approximate Gaussian Process Regression

Quinonero Candela, J., Rasmussen, C.

Journal of Machine Learning Research, 6, pages: 1935-1959, December 2005 (article)

Abstract
We provide a new unifying view, including all existing proper probabilistic sparse approximations for Gaussian process regression. Our approach relies on expressing the effective prior which the methods are using. This allows new insights to be gained, and highlights the relationship between existing methods. It also allows for a clear theoretically justified ranking of the closeness of the known approximations to the corresponding full GPs. Finally we point directly to designs of new better sparse approximations, combining the best of the existing strategies, within attractive computational constraints.

ei

PDF [BibTex]

PDF [BibTex]


no image
Training Support Vector Machines with Multiple Equality Constraints

Kienzle, W., Schölkopf, B.

In Proceedings of the 16th European Conference on Machine Learning, Lecture Notes in Computer Science, Vol. 3720, pages: 182-193, (Editors: JG Carbonell and J Siekmann), Springer, Berlin, Germany, ECML, November 2005 (inproceedings)

Abstract
In this paper we present a primal-dual decomposition algorithm for support vector machine training. As with existing methods that use very small working sets (such as Sequential Minimal Optimization (SMO), Successive Over-Relaxation (SOR) or the Kernel Adatron (KA)), our method scales well, is straightforward to implement, and does not require an external QP solver. Unlike SMO, SOR and KA, the method is applicable to a large number of SVM formulations regardless of the number of equality constraints involved. The effectiveness of our algorithm is demonstrated on a more difficult SVM variant in this respect, namely semi-parametric support vector regression.

ei

PDF DOI [BibTex]

PDF DOI [BibTex]


no image
Measuring Statistical Dependence with Hilbert-Schmidt Norms

Gretton, A., Bousquet, O., Smola, A., Schoelkopf, B.

In Algorithmic Learning Theory, Lecture Notes in Computer Science, Vol. 3734, pages: 63-78, (Editors: S Jain and H-U Simon and E Tomita), Springer, Berlin, Germany, 16th International Conference ALT, October 2005 (inproceedings)

Abstract
We propose an independence criterion based on the eigenspectrum of covariance operators in reproducing kernel Hilbert spaces (RKHSs), consisting of an empirical estimate of the Hilbert-Schmidt norm of the cross-covariance operator (we term this a Hilbert-Schmidt Independence Criterion, or HSIC). This approach has several advantages, compared with previous kernel-based independence criteria. First, the empirical estimate is simpler than any other kernel dependence test, and requires no user-defined regularisation. Second, there is a clearly defined population quantity which the empirical estimate approaches in the large sample limit, with exponential convergence guaranteed between the two: this ensures that independence tests based on {methodname} do not suffer from slow learning rates. Finally, we show in the context of independent component analysis (ICA) that the performance of HSIC is competitive with that of previously published kernel-based criteria, and of other recently published ICA methods.

ei

PDF DOI [BibTex]

PDF DOI [BibTex]


no image
Maximal Margin Classification for Metric Spaces

Hein, M., Bousquet, O., Schölkopf, B.

Journal of Computer and System Sciences, 71(3):333-359, October 2005 (article)

Abstract
In order to apply the maximum margin method in arbitrary metric spaces, we suggest to embed the metric space into a Banach or Hilbert space and to perform linear classification in this space. We propose several embeddings and recall that an isometric embedding in a Banach space is always possible while an isometric embedding in a Hilbert space is only possible for certain metric spaces. As a result, we obtain a general maximum margin classification algorithm for arbitrary metric spaces (whose solution is approximated by an algorithm of Graepel. Interestingly enough, the embedding approach, when applied to a metric which can be embedded into a Hilbert space, yields the SVM algorithm, which emphasizes the fact that its solution depends on the metric and not on the kernel. Furthermore we give upper bounds of the capacity of the function classes corresponding to both embeddings in terms of Rademacher averages. Finally we compare the capacities of these function classes directly.

ei

PDF PDF DOI [BibTex]

PDF PDF DOI [BibTex]


no image
An Analysis of the Anti-Learning Phenomenon for the Class Symmetric Polyhedron

Kowalczyk, A., Chapelle, O.

In Algorithmic Learning Theory: 16th International Conference, pages: 78-92, Algorithmic Learning Theory, October 2005 (inproceedings)

Abstract
This paper deals with an unusual phenomenon where most machine learning algorithms yield good performance on the training set but systematically worse than random performance on the test set. This has been observed so far for some natural data sets and demonstrated for some synthetic data sets when the classification rule is learned from a small set of training samples drawn from some high dimensional space. The initial analysis presented in this paper shows that anti-learning is a property of data sets and is quite distinct from overfitting of a training data. Moreover, the analysis leads to a specification of some machine learning procedures which can overcome anti-learning and generate ma- chines able to classify training and test data consistently.

ei

PDF [BibTex]

PDF [BibTex]