You can also find my articles on my Google Scholar profile.

Working papers

  1. [WP2] A. Lindholm and F. Lindsten, Learning dynamical systems with particle stochastic approximation EM. arXiv.org, arXiv:1806.09548, 2019.

    We present the particle stochastic approximation EM (PSAEM) algorithm for learning of dynamical systems. The method builds on the EM algorithm, an iterative procedure for maximum likelihood inference in latent variable models. By combining stochastic approximation EM and particle Gibbs with ancestor sampling (PGAS), PSAEM obtains superior computational performance and convergence properties compared to plain particle-smoothing-based approximations of the EM algorithm. PSAEM can be used for plain maximum likelihood inference as well as for empirical Bayes learning of hyperparameters. Specifically, the latter point means that existing PGAS implementations easily can be extended with PSAEM to estimate hyperparameters at almost no extra computational cost. We discuss the convergence properties of the algorithm, and demonstrate it on several signal processing applications.

    @misc{LindholmL:2019,
      author = {Lindholm, Andreas and Lindsten, Fredrik},
      title = {Learning dynamical systems with particle stochastic approximation {EM}},
      howpublished = {arXiv.org, arXiv:1806.09548},
      url = {https://arxiv.org/abs/1806.09548},
      year = {2019},
      pid = {WP2}
    }
    

Monographs and books

  1. [B1] A. Lindholm, N. Wahlström, F. Lindsten, and T. B. Schön, Machine Learning: A First Course for Engineers and Scientists. Cambridge University Press, 2022.
    DOI  — URL  — BibTeX
    @book{LindholmWLS:2020,
      author = {Lindholm, Andreas and Wahlström, Niklas and Lindsten, Fredrik and Schön, Thomas B.},
      title = {Machine Learning: A First Course for Engineers and Scientists},
      doi = {10.1017/9781108919371},
      publisher = {Cambridge University Press},
      year = {2022},
      place = {Cambridge},
      url = {http://smlbook.org/},
      pid = {B1}
    }
    
  2. [M2] C. A. Naesseth, F. Lindsten, and T. B. Schön, Elements of Sequential Monte Carlo, Foundations and Trends in Machine Learning, vol. 12, no. 3, pp. 307–392, 2019.
    DOI  — arXiv  — Abstract — BibTeX

    A core problem in statistics and probabilistic machine learning is to compute probability distributions and expectations. This is the fundamental problem of Bayesian statistics and machine learning, which frames all inference as expectations with respect to the posterior distribution. The key challenge is to approximate these intractable expectations. In this tutorial, we review sequential Monte Carlo (SMC), a random-sampling-based class of methods for approximate inference. First, we explain the basics of SMC, discuss practical issues, and review theoretical results. We then examine two of the main user design choices: the proposal distributions and the so called intermediate target distributions. We review recent results on how variational inference and amortization can be used to learn efficient proposals and target distributions. Next, we discuss the SMC estimate of the normalizing constant, how this can be used for pseudo-marginal inference and inference evaluation. Throughout the tutorial we illustrate the use of SMC on various models commonly used in machine learning, such as stochastic recurrent neural networks, probabilistic graphical models, and probabilistic programs.

    @article{NaessethLS:2019a,
      author = {Naesseth, Christian A. and Lindsten, Fredrik and Sch\"on, Thomas B.},
      title = {Elements of Sequential {M}onte {C}arlo},
      journal = {Foundations and Trends in Machine Learning},
      year = {2019},
      volume = {12},
      number = {3},
      pages = {307--392},
      pid = {M2},
      doi = {10.1561/2200000074},
      url = {https://arxiv.org/abs/1903.04797}
    }
    
  3. [M1] F. Lindsten and T. B. Schön, Backward simulation methods for Monte Carlo statistical inference, Foundations and Trends in Machine Learning, vol. 6, no. 1, pp. 1–143, 2013.
    DOI  — Abstract — BibTeX

    Monte Carlo methods, in particular those based on Markov chains and on interacting particle systems, are by now tools that are routinely used in machine learning. These methods have had a profound impact on statistical inference in a wide range of application areas where probabilistic models are used. Moreover, there are many algorithms in machine learning which are based on the idea of processing the data sequentially, first in the forward direction and then in the backward direction. In this tutorial, we will review a branch of Monte Carlo methods based on the forward–backward idea, referred to as backward simulators. These methods are useful for learning and inference in probabilistic models containing latent stochastic processes. The theory and practice of backward simulation algorithms have undergone a significant development in recent years and the algorithms keep finding new applications. The foundation for these methods is sequential Monte Carlo (SMC). SMC-based backward simulators are capable of addressing smoothing problems in sequential latent variable models, such as general, nonlinear/non-Gaussian state-space models (SSMs). However, we will also clearly show that the underlying backward simulation idea is by no means restricted to SSMs. Furthermore, backward simulation plays an important role in recent developments of Markov chain Monte Carlo (MCMC) methods. Particle MCMC is a systematic way of using SMC within MCMC. In this framework, backward simulation gives us a way to significantly improve the performance of the samplers. We review and discuss several related backward-simulation-based methods for state inference as well as learning of static parameters, both using a frequentistic and a Bayesian approach.

    @article{LindstenS:2013,
      author = {Lindsten, Fredrik and Sch\"on, Thomas B.},
      title = {Backward simulation methods for {M}onte {C}arlo statistical inference},
      journal = {Foundations and Trends in Machine Learning},
      year = {2013},
      volume = {6},
      number = {1},
      pages = {1--143},
      doi = {10.1561/2200000045},
      pid = {M1}
    }
    

Journal papers

  1. [J19] P. Konold, L. Monrroy, A. Bellisario, D. Filipe, P. Adams, R. Alvarez, R. Bean, J. Bielecki, S. Bódizs, et al., Microsecond time-resolved X-ray scattering by utilizing MHz repetition rate at second-generation XFELs, Nature Methods, 2024.
    @article{KonoldEtAl:2024,
      author = {Konold, Patrick and Monrroy, Leonardo and Bellisario, Alfredo and Filipe, Diogo and Adams, Patrick and Alvarez, Roberto and Bean, Richard and Bielecki, Johan and Bódizs, Szabolcs and Ducroq, Gabriel and Grubmueller, Helmut and Kirian, Richard and Kloos, Marco and Koliyadu, Jayanath and Koua, Faisal and Larkiala, Taru and Letrun, Romain and Lindsten, Fredrik and Maihöfer, Michael and Martin, Andrew and Mészáros, Petra and Mutisya, Jennifer and Nimmrich, Amke and Okamoto, Kenta and Round, Adam and Sato, Tokushi and Valerio, Joana and Westphal, Daniel and Wolter, August and Yenupuri, Tej and You, Tong and Maia, Filipe and Westenhoff, Sebastian},
      title = {Microsecond time-resolved {X}-ray scattering by utilizing {MHz} repetition rate at second-generation {XFELs}},
      journal = {Nature Methods},
      year = {2024},
      note = {Forthcoming},
      pid = {J19}
    }
    
  2. [J18] H. Zimmermann, F. Lindsten, J.-W. van de Meent, and C. A. Naesseth, A Variational Perspective on Generative Flow Networks, Transactions on Machine Learning Research (TMLR), Apr. 2023.
    @article{ZimmermannLMN:2023,
      author = {Zimmermann, Heiko and Lindsten, Fredrik and van de Meent, Jan-Willem and Naesseth, Christian A},
      title = {A Variational Perspective on Generative Flow Networks},
      journal = {Transactions on Machine Learning Research (TMLR)},
      year = {2023},
      month = apr,
      url = {https://openreview.net/forum?id=AZ4GobeSLq},
      pid = {J18}
    }
    
  3. [J17] F. E. Kelvinius, F. Lindsten, and R. Armiento, Graph-based machine learning beyond stable materials and relaxed crystal structures, Physical Review Materials, vol. 6, no. 3, p. 033801, Mar. 2022.
    DOI  — arXiv  — BibTeX
    @article{EkstromAL:2022,
      author = {Kelvinius, Filip Ekstr\"om and Lindsten, Fredrik and Armiento, Rickard},
      title = {Graph-based machine learning beyond stable materials and relaxed crystal structures},
      journal = {Physical Review Materials},
      volume = {6},
      issue = {3},
      pages = {033801},
      numpages = {10},
      year = {2022},
      month = mar,
      publisher = {American Physical Society},
      doi = {10.1103/PhysRevMaterials.6.033801},
      url = {https://arxiv.org/abs/2107.00493},
      pid = {J17}
    }
    
  4. [J16] A. Wigren, J. Wågberg, F. Lindsten, A. Wills, and T. B. Schön, Nonlinear System Identification - Learning while respecting physical models using Sequential Monte Carlo, IEEE Control Systems Magazine, vol. 42, no. 1, pp. 75–102, 2022.
    DOI  — BibTeX
    @article{WigrenWLWS:2022,
      author = {Wigren, Anna and Wågberg, Johan and Lindsten, Fredrik and Wills, Adrian and Schön, Thomas B.},
      title = {Nonlinear System Identification - Learning while respecting physical models using Sequential {M}onte {C}arlo},
      journal = {{IEEE} Control Systems Magazine},
      year = {2022},
      volume = {42},
      number = {1},
      pages = {75--102},
      doi = {https://doi.org/10.1109/MCS.2021.3122269},
      pid = {J16}
    }
    
  5. [J15] J. Alenlöv, A. Doucet, and F. Lindsten, Pseudo-Marginal Hamiltonian Monte Carlo, Journal of Machine Learning Research, vol. 22, no. 141. pp. 1–45, 2021.
    URL  — Abstract — BibTeX

    Bayesian inference in the presence of an intractable likelihood function is computationally challenging. When following a Markov chain Monte Carlo (MCMC) approach to approximate the posterior distribution in this context, one typically either uses MCMC schemes which target the joint posterior of the parameters and some auxiliary latent variables, or pseudo-marginal Metropolis–Hastings (MH) schemes. The latter mimic a MH algorithm targeting the marginal posterior of the parameters by approximating unbiasedly the intractable likelihood. However, in scenarios where the parameters and auxiliary variables are strongly correlated under the posterior and/or this posterior is multimodal, Gibbs sampling or Hamiltonian Monte Carlo (HMC) will perform poorly and the pseudo-marginal MH algorithm, as any other MH scheme, will be inefficient for high-dimensional parameters. We propose here an original MCMC algorithm, termed pseudo-marginal HMC, which combines the advantages of both HMC and pseudo-marginal schemes. Specifically, the PM-HMC method is controlled by a precision parameter N, controlling the approximation of the likelihood and, for any N, it samples the marginal posterior of the parameters. Additionally, as N tends to infinity, its sample trajectories and acceptance probability converge to those of an ideal, but intractable, HMC algorithm which would have access to the intractable likelihood and its gradient. We demonstrate through experiments that PM-HMC can outperform significantly both standard HMC and pseudo-marginal MH schemes.

    @misc{AlenlovDL:2021,
      author = {Alenlöv, Johan and Doucet, Arnaud and Lindsten, Fredrik},
      title = {Pseudo-Marginal {H}amiltonian {M}onte {C}arlo},
      journal = {Journal of Machine Learning Research},
      url = {https://www.jmlr.org/papers/v22/19-486.html},
      year = {2021},
      volume = {22},
      number = {141},
      pages = {1--45},
      pid = {J15}
    }
    
  6. [J14] C. A. Naesseth, F. Lindsten, and T. B. Schön, High-dimensional Filtering using Nested Sequential Monte Carlo, IEEE Transactions on Signal Processing, vol. 67, no. 16, pp. 4177–4188, 2019.
    DOI  — arXiv  — Abstract — BibTeX

    Sequential Monte Carlo (SMC) methods comprise one of the most successful approaches to approximate Bayesian filtering. However, SMC without good proposal distributions struggle in high dimensions. We propose nested sequential Monte Carlo (NSMC), a methodology that generalises the SMC framework by requiring only approximate, properly weighted, samples from the SMC proposal distribution, while still resulting in a correct SMC algorithm. This way we can exactly approximate the locally optimal proposal, and extend the class of models for which we can perform efficient inference using SMC. We show improved accuracy over other state-of-the-art methods on several spatio-temporal state space models.

    @article{NaessethLS:2019,
      author = {Naesseth, Christian A. and Lindsten, Fredrik and Schön, Thomas B.},
      title = {High-dimensional Filtering using Nested Sequential {M}onte {C}arlo},
      journal = {{IEEE} Transactions on Signal Processing},
      year = {2019},
      volume = {67},
      number = {16},
      pages = {4177--4188},
      doi = {10.1109/TSP.2019.2926035},
      url = {https://arxiv.org/abs/1612.09162},
      pid = {J14}
    }
    
  7. [J13] R. S. Risuleo, F. Lindsten, and H. Hjalmarsson, Bayesian nonparametric identification of Wiener systems, Automatica, vol. 108, 2019.
    DOI  — Abstract — BibTeX

    We propose a nonparametric approach for the identification of Wiener systems. We model the impulse response of the linear block and the static nonlinearity using Gaussian processes. The hyperparameters of the Gaussian processes are estimated using an iterative algorithm based on stochastic approximation expectation–maximization. In the iterations, we use elliptical slice sampling to approximate the posterior distribution of the impulse response and update the hyperparameter estimates. The same sampling is finally used to sample the posterior distribution and to compute point estimates. We compare the proposed approach with a parametric approach and a semi-parametric approach. In particular, we show that the proposed method has an advantage when a parametric model for the system is not readily available.

    @article{RisuleoLH:2019,
      author = {Risuleo, Riccardo S. and Lindsten, Fredrik and Hjalmarsson, Håkan},
      title = {Bayesian nonparametric identification of {W}iener systems},
      journal = {Automatica},
      year = {2019},
      volume = {108},
      note = {Brief paper},
      doi = {10.1016/j.automatica.2019.06.032},
      pid = {J13}
    }
    
  8. [J12] P. E. Jacob, F. Lindsten, and T. B. Schön, Smoothing with Couplings of Conditional Particle Filters, Journal of the American Statistical Association, vol. 115, no. 530, pp. 721–729, 2019.
    DOI  — arXiv  — Abstract — BibTeX

    In state–space models, smoothing refers to the task of estimating a latent stochastic process given noisy measurements related to the process. We propose an unbiased estimator of smoothing expectations. The lack-of-bias property has methodological benefits: independent estimators can be generated in parallel, and CI can be constructed from the central limit theorem to quantify the approximation error. To design unbiased estimators, we combine a generic debiasing technique for Markov chains, with a Markov chain Monte Carlo algorithm for smoothing. The resulting procedure is widely applicable and we show in numerical experiments that the removal of the bias comes at a manageable increase in variance. We establish the validity of the proposed estimators under mild assumptions. Numerical experiments are provided on toy models, including a setting of highly informative observations, and for a realistic Lotka–Volterra model with an intractable transition density.

    @article{JacobLS:2019,
      author = {Jacob, Pierre E. and Lindsten, Fredrik and Schön, Thomas B.},
      title = {Smoothing with Couplings of Conditional Particle Filters},
      journal = {Journal of the American Statistical Association},
      year = {2019},
      pages = {721--729},
      volume = {115},
      number = {530},
      pid = {J12},
      doi = {https://doi.org/10.1080/01621459.2018.1548856},
      url = {https://arxiv.org/abs/1701.02002}
    }
    
  9. [J11] F. M. Calafat, T. Wahl, F. Lindsten, J. Williams, and E. Frajka-Williams, Coherent modulation of the sea-level annual cycle in the United States by Atlantic Rossby waves, Nature Communications, vol. 9, no. 2571, 2018.
    DOI  — Abstract — BibTeX

    Changes in the sea-level annual cycle (SLAC) can have profound impacts on coastal areas, including increased flooding risk and ecosystem alteration, yet little is known about the magnitude and drivers of such changes. Here we show, using novel Bayesian methods, that there are significant decadal fluctuations in the amplitude of the SLAC along the United States Gulf and Southeast coasts, including an extreme event in 2008–2009 that is likely (probability ≥68%) unprecedented in the tide-gauge record. Such fluctuations are coherent along the coast but decoupled from deep-ocean changes. Through the use of numerical and analytical ocean models, we show that the primary driver of these fluctuations involves incident Rossby waves that generate fast western-boundary waves. These Rossby waves project onto the basin-wide upper mid-ocean transport (top 1000 m) leading to a link with the SLAC, wherein larger SLAC amplitudes coincide with enhanced transport variability.

    @article{CalafatWLWF:2018,
      author = {Calafat, Francisco M. and Wahl, Thomas and Lindsten, Fredrik and Williams, Joanne and Frajka-Williams, Eleanor},
      title = {Coherent modulation of the sea-level annual cycle in the {U}nited {S}tates by {A}tlantic {R}ossby waves},
      journal = {Nature Communications},
      year = {2018},
      volume = {9},
      number = {2571},
      doi = {10.1038/s41467-018-04898-y},
      pid = {J11}
    }
    
  10. [J10] T. B. Schön, A. Svensson, L. Murray, and F. Lindsten, Probabilistic learning of nonlinear dynamical systems using sequential Monte Carlo, Mechanical Systems and Signal Processing, vol. 104, pp. 866–883, 2018.

    Probabilistic modeling provides the capability to represent and manipulate uncertainty in data, models, predictions and decisions. We are concerned with the problem of learning probabilistic models of dynamical systems from measured data. Specifically, we consider learning of probabilistic nonlinear state-space models. There is no closed-form solution available for this problem, implying that we are forced to use approximations. In this tutorial we will provide a self-contained introduction to one of the state-of-the-art methods—the particle Metropolis–Hastings algorithm—which has proven to offer a practical approximation. This is a Monte Carlo based method, where the particle filter is used to guide a Markov chain Monte Carlo method through the parameter space. One of the key merits of the particle Metropolis–Hastings algorithm is that it is guaranteed to converge to the "true solution" under mild assumptions, despite being based on a particle filter with only a finite number of particles. We will also provide a motivating numerical example illustrating the method using a modeling language tailored for sequential Monte Carlo methods. The intention of modeling languages of this kind is to open up the power of sophisticated Monte Carlo methods—including particle Metropolis–Hastings—to a large group of users without requiring them to know all the underlying mathematical details.

    @article{SchonSML:2018,
      author = {Schön, Thomas B. and Svensson, Andreas and Murray, Lawrence and Lindsten, Fredrik},
      title = {Probabilistic learning of nonlinear dynamical systems using sequential {M}onte {C}arlo},
      journal = {Mechanical Systems and Signal Processing},
      year = {2018},
      volume = {104},
      pages = {866--883},
      url = {https://arxiv.org/abs/1703.02419},
      pid = {J10}
    }
    
  11. [J9] A. Svensson, T. B. Schön, and F. Lindsten, Learning of state-space models with highly informative observations: a tempered Sequential Monte Carlo solution, Mechanical Systems and Signal Processing, vol. 104, pp. 915–928, 2018.

    Probabilistic (or Bayesian) modeling and learning offers interesting possibilities for systematic representation of uncertainty using probability theory. However, probabilistic learning often leads to computationally challenging problems. Some problems of this type that were previously intractable can now be solved on standard personal computers thanks to recent advances in Monte Carlo methods. In particular, for learning of unknown parameters in nonlinear state-space models, methods based on the particle filter (a Monte Carlo method) have proven very useful. A notoriously challenging problem, however, still occurs when the observations in the state-space model are highly informative, i.e. when there is very little or no measurement noise present, relative to the amount of process noise. The particle filter will then struggle in estimating one of the basic components for probabilistic learning, namely the likelihood p(data|parameters). To this end we suggest an algorithm which initially assumes that there is substantial amount of artificial measurement noise present. The variance of this noise is sequentially decreased in an adaptive fashion such that we, in the end, recover the original problem or possibly a very close approximation of it. The main component in our algorithm is a sequential Monte Carlo (SMC) sampler, which gives our proposed method a clear resemblance to the SMC^2 method. Another natural link is also made to the ideas underlying the approximate Bayesian computation (ABC). We illustrate it with numerical examples, and in particular show promising results for a challenging Wiener-Hammerstein benchmark problem.

    @article{SvenssonSL:2018,
      author = {Svensson, Andreas and Schön, Thomas B. and Lindsten, Fredrik},
      title = {Learning of state-space models with highly informative observations: a tempered Sequential {M}onte {C}arlo solution},
      journal = {Mechanical Systems and Signal Processing},
      year = {2018},
      volume = {104},
      pages = {915--928},
      url = {https://arxiv.org/abs/1702.01618},
      pid = {J9}
    }
    
  12. [J8] S. S. Singh, F. Lindsten, and E. Moulines, Blocking Strategies and Stability of Particle Gibbs Samplers, Biometrika, vol. 104, no. 4, pp. 953–969, 2017.
    DOI  — arXiv  — Abstract — BibTeX

    Sampling from the conditional (or posterior) probability distribution of the latent states of a Hidden Markov Model, given the realization of the observed process, is a non-trivial problem in the context of Markov Chain Monte Carlo. To do this Andrieu et al. (2010) constructed a Markov kernel which leaves this conditional distribution invariant using a Particle Filter. From a practitioner’s point of view, this Markov kernel attempts to mimic the act of sampling all the latent state variables as one block from the posterior distribution but for models where exact simulation is not possible. There are some recent theoretical results that establish the uniform ergodicity of this Markov kernel and that the mixing rate does not diminish provided the number of particles grows at least linearly with the number of latent states in the posterior. This gives rise to a cost, per application of the kernel, that is quadratic in the number of latent states which could be prohibitive for long observation sequences. We seek to answer an obvious but important question: is there a different implementation with a cost per-iteration that grows linearly with the number of latent states, but which is still stable in the sense that its mixing rate does not deteriorate? We address this problem using blocking strategies, which are easily parallelizable, and prove stability of the resulting sampler.

    @article{SinghLM:2017,
      author = {Singh, Sumeetpal S. and Lindsten, Fredrik and Moulines, Eric},
      title = {Blocking Strategies and Stability of Particle {G}ibbs Samplers},
      journal = {Biometrika},
      year = {2017},
      volume = {104},
      number = {4},
      pages = {953--969},
      url = {https://arxiv.org/abs/1509.08362},
      doi = {10.1093/biomet/asx051},
      pid = {J8}
    }
    
  13. [J7] F. Lindsten, A. Johansen, C. A. Naesseth, B. Kirkpatrick, T. B. Schön, J. Aston, and A. Bouchard-Côté, Divide-and-Conquer with Sequential Monte Carlo, Journal of Computational and Graphical Statistics, vol. 26, no. 2, pp. 445–458, 2017.
    DOI  — arXiv  — Abstract — BibTeX

    We propose a novel class of Sequential Monte Carlo (SMC) algorithms, appropriate for inference in probabilistic graphical models. This class of algorithms adopts a divide-and-conquer approach based upon an auxiliary tree-structured decomposition of the model of interest, turning the overall inferential task into a collection of recursively solved sub-problems. The proposed method is applicable to a broad class of probabilistic graphical models, including models with loops. Unlike a standard SMC sampler, the proposed Divide-and-Conquer SMC employs multiple independent populations of weighted particles, which are resampled, merged, and propagated as the method progresses. We illustrate empirically that this approach can outperform standard methods in terms of the accuracy of the posterior expectation and marginal likelihood approximations. Divide-and-Conquer SMC also opens up novel parallel implementation options and the possibility of concentrating the computational effort on the most challenging sub-problems. We demonstrate its performance on a Markov random field and on a hierarchical logistic regression problem.

    @article{LindstenJNKSAB:2017,
      author = {Lindsten, Fredrik and Johansen, Adam and Naesseth, Christian A. and Kirkpatrick, Brent and Sch\"on, Thomas B. and Aston, John and Bouchard-C\^ot\'e, Alexandre},
      title = {Divide-and-Conquer with Sequential {M}onte {C}arlo},
      journal = {Journal of Computational and Graphical Statistics},
      year = {2017},
      volume = {26},
      number = {2},
      pages = {445--458},
      url = {https://arxiv.org/abs/1406.4993},
      doi = {10.1080/10618600.2016.1237363},
      pid = {J7}
    }
    
  14. [J6] F. Lindsten, P. Bunch, S. Särkkä, T. B. Schön, and S. J. Godsill, Rao-Blackwellized particle smoothers for conditionally linear Gaussian models, IEEE Journal of Selected Topics in Signal Processing, vol. 10, no. 2, pp. 353–365, 2016.

    Sequential Monte Carlo (SMC) methods, such as the particle filter, are by now one of the standard computational techniques for addressing the filtering problem in general state-space models. However, many applications require post-processing of data offline. In such scenarios the smoothing problem–in which all the available data is used to compute state estimates–is of central interest. We consider the smoothing problem for a class of conditionally linear Gaussian models. We present a forward-backward-type Rao-Blackwellized particle smoother (RBPS) that is able to exploit the tractable substructure present in these models. Akin to the well known Rao-Blackwellized particle filter, the proposed RBPS marginalizes out a conditionally tractable subset of state variables, effectively making use of SMC only for the "intractable part" of the model. Compared to existing RBPS, two key features of the proposed method are: (i) it does not require structural approximations of the model, and (ii) the aforementioned marginalization is done both in the forward direction and in the backward direction.

    @article{LindstenBSSG:2016,
      author = {Lindsten, Fredrik and Bunch, Pete and Särkkä, S. and Schön, Thomas B. and Godsill, Simon J.},
      title = {Rao-{B}lackwellized particle smoothers for conditionally linear {G}aussian models},
      journal = {IEEE Journal of Selected Topics in Signal Processing},
      year = {2016},
      volume = {10},
      number = {2},
      pages = {353--365},
      url = {https://arxiv.org/abs/1505.06357},
      pid = {J6}
    }
    
  15. [J5] F. Lindsten, R. Douc, and E. Moulines, Uniform ergodicity of the Particle Gibbs sampler, Scandinavian Journal of Statistics, vol. 42, no. 3, pp. 775–797, 2015.
    DOI  — arXiv  — Abstract — BibTeX

    The particle Gibbs (PG) sampler is a systematic way of using a particle filter within Markov chain Monte Carlo (MCMC). This results in an off-the-shelf Markov kernel on the space of state trajectories, which can be used to simulate from the full joint smoothing distribution for a state space model in an MCMC scheme. We show that the PG Markov kernel is uniformly ergodic under rather general assumptions, that we will carefully review and discuss. In particular, we provide an explicit rate of convergence which reveals that: (i) for fixed number of data points, the convergence rate can be made arbitrarily good by increasing the number of particles, and (ii) under general mixing assumptions, the convergence rate can be kept constant by increasing the number of particles superlinearly with the number of observations. We illustrate the applicability of our result by studying in detail two common state space models with non-compact state spaces.

    @article{LindstenDM:2015,
      author = {Lindsten, Fredrik and Douc, Randal and Moulines, Eric},
      title = {Uniform ergodicity of the Particle {G}ibbs sampler},
      journal = {Scandinavian Journal of Statistics},
      year = {2015},
      volume = {42},
      number = {3},
      pages = {775--797},
      pid = {J5},
      doi = {10.1111/sjos.12136},
      url = {https://arxiv.org/abs/1401.0683}
    }
    
  16. [J4] E. Özkan, F. Lindsten, C. Fritsche, and F. Gustafsson, Recursive maximum likelihood identification of jump Markov nonlinear systems, IEEE Transactions on Signal Processing, vol. 63, no. 3, pp. 754–765, 2015.

    In this contribution, we present an online method for joint state and parameter estimation in jump Markov non-linear systems (JMNLS). State inference is enabled via the use of particle filters which makes the method applicable to a wide range of non-linear models. To exploit the inherent structure of JMNLS, we design a Rao-Blackwellized particle filter (RBPF) where the discrete mode is marginalized out analytically. This results in an efficient implementation of the algorithm and reduces the estimation error variance. The proposed RBPF is then used to compute, recursively in time, smoothed estimates of complete data sufficient statistics. Together with the online expectation maximization algorithm, this enables recursive identification of unknown model parameters. The performance of the method is illustrated in simulations and on a localization problem in wireless networks using real data.

    @article{OezkanLFG:2015,
      author = {{\"O}zkan, Emre and Lindsten, Fredrik and Fritsche, Carsten and Gustafsson, Fredrik},
      title = {Recursive maximum likelihood identification of jump {M}arkov nonlinear systems},
      journal = {{IEEE} Transactions on Signal Processing},
      year = {2015},
      volume = {63},
      number = {3},
      pages = {754--765},
      url = {https://arxiv.org/abs/1312.0781},
      pid = {J4}
    }
    
  17. [J3] J. Dahlin, F. Lindsten, and T. B. Schön, Particle Metropolis-Hastings using gradient and Hessian information, Statistics and Computing, vol. 25, no. 1, pp. 81–92, 2015.

    Particle Metropolis-Hastings (PMH) allows for Bayesian parameter inference in nonlinear state space models by combining Markov chain Monte Carlo (MCMC) and particle filtering. The latter is used to estimate the intractable likelihood. In its original formulation, PMH makes use of a marginal MCMC proposal for the parameters, typically a Gaussian random walk. However, this can lead to a poor exploration of the parameter space and an inefficient use of the generated particles. We propose a number of alternative versions of PMH that incorporate gradient and Hessian information about the posterior into the proposal. This information is more or less obtained as a byproduct of the likelihood estimation. Indeed, we show how to estimate the required information using a fixed-lag particle smoother, with a computational cost growing linearly in the number of particles. We conclude that the proposed methods can: (i) decrease the length of the burn-in phase, (ii) increase the mixing of the Markov chain at the stationary phase, and (iii) make the proposal distribution scale invariant which simplifies tuning.

    @article{DahlinLS:2015,
      author = {Dahlin, Johan and Lindsten, Fredrik and Sch\"on, Thomas B.},
      title = {Particle {M}etropolis-{H}astings using gradient and {H}essian information},
      journal = {Statistics and Computing},
      year = {2015},
      volume = {25},
      number = {1},
      pages = {81--92},
      url = {https://arxiv.org/abs/1311.0686},
      pid = {J3}
    }
    
  18. [J2] F. Lindsten, M. I. Jordan, and T. B. Schön, Particle Gibbs with Ancestor Sampling, Journal of Machine Learning Research, vol. 15, pp. 2145–2184, 2014.
    URL  — Abstract — BibTeX

    Particle Markov chain Monte Carlo (PMCMC) is a systematic way of combining the two main tools used for Monte Carlo statistical inference: sequential Monte Carlo (SMC) and Markov chain Monte Carlo (MCMC). We present a novel PMCMC algorithm that we refer to as particle Gibbs with ancestor sampling (PGAS). PGAS provides the data analyst with an off-the-shelf class of Markov kernels that can be used to simulate the typically high-dimensional and highly autocorrelated state trajectory in a state-space model. The ancestor sampling procedure enables fast mixing of the PGAS kernel even when using seemingly few particles in the underlying SMC sampler. This is important as it can significantly reduce the computational burden that is typically associated with using SMC. PGAS is conceptually similar to the existing PG with backward simulation (PGBS) procedure. Instead of using separate forward and backward sweeps as in PGBS, however, we achieve the same effect in a single forward sweep. This makes PGAS well suited for addressing inference problems not only in state-space models, but also in models with more complex dependencies, such as non-Markovian, Bayesian nonparametric, and general probabilistic graphical models.

    @article{LindstenJS:2014,
      author = {Lindsten, Fredrik and Jordan, Michael I. and Sch\"on, Thomas B.},
      title = {Particle {G}ibbs with Ancestor Sampling},
      journal = {Journal of Machine Learning Research},
      year = {2014},
      volume = {15},
      pages = {2145--2184},
      url = {http://jmlr.org/papers/v15/lindsten14a.html},
      pid = {J2}
    }
    
  19. [J1] F. Lindsten, T. B. Schön, and M. I. Jordan, Bayesian semiparametric Wiener system identification, Automatica, vol. 49, no. 7, pp. 2053–2063, 2013.
    DOI  — Abstract — BibTeX

    We present a novel method for Wiener system identification. The method relies on a semiparametric, i.e. a mixed parametric/nonparametric, model of a Wiener system. We use a state-space model for the linear dynamical system and a nonparametric Gaussian process model for the static nonlinearity. We avoid making strong assumptions, such as monotonicity, on the nonlinear mapping. Stochastic disturbances, entering both as measurement noise and as process noise, are handled in a systematic manner. The nonparametric nature of the Gaussian process allows us to handle a wide range of nonlinearities without making problem-specific parameterizations. We also consider sparsity-promoting priors, based on generalized hyperbolic distributions, to automatically infer the order of the underlying dynamical system. We derive an inference algorithm based on an efficient particle Markov chain Monte Carlo method, referred to as particle Gibbs with ancestor sampling. The method is profiled on two challenging identification problems with good results. Blind Wiener system identification is handled as a special case.

    @article{LindstenSJ:2013,
      author = {Lindsten, Fredrik and Sch\"on, Thomas B. and Jordan, Michael I.},
      title = {Bayesian semiparametric {W}iener system identification},
      journal = {Automatica},
      year = {2013},
      volume = {49},
      number = {7},
      pages = {2053--2063},
      doi = {10.1016/j.automatica.2013.03.021},
      pid = {J1}
    }
    

Conference papers

  1. [C61] On Partial Prototype Collapse in the DINO Family of Self-Supervised Methods, in Proceedings of the 35th British Machine Vision Conference, Glasgow, UK, 2024.

    A prominent self-supervised learning paradigm is to model the representations as clusters, or more generally as a mixture model. Learning to map the data samples to compact representations and fitting the mixture model simultaneously leads to the representation collapse problem. Regularizing the distribution of data points over the clusters is the prevalent strategy to avoid this issue. While this is sufficient to prevent full representation collapse, we show that a partial prototype collapse problem still exists in the DINO family of methods, that leads to significant redundancies in the prototypes. Such prototype redundancies serve as shortcuts for the method to achieve a marginal latent class distribution that matches the prescribed prior. We show that by encouraging the model to use diverse prototypes, the partial prototype collapse can be mitigated. We study the downstream impact of effective utilization of the prototypes during pre-training. We show that it enables the methods to learn more fine-grained clusters, encouraging more informative representations. We demonstrate that this is especially beneficial when pre-training on a long-tailed fine-grained dataset.

    @inproceedings{GovindarajanSRL:2024,
      author = {},
      title = {On Partial Prototype Collapse in the DINO Family of Self-Supervised Methods},
      booktitle = {Proceedings of the  35th British Machine Vision Conference},
      year = {2024},
      address = {Glasgow, UK},
      month = nov,
      pid = {C61}
    }
    
  2. [C60] F. E. Kelvinius and F. Lindsten, Discriminator Guidance for Autoregressive Diffusion Models, in Proceedings of the 27th International Conference on Artificial Intelligence and Statistics, Valencia, Spain, 2024.

    We introduce discriminator guidance in the setting of Autoregressive Diffusion Models. The use of a discriminator to guide a diffusion process has previously been used for continuous diffusion models, and in this work we derive ways of using a discriminator together with a pretrained generative model in the discrete case. First, we show that using an optimal discriminator will correct the pretrained model and enable exact sampling from the underlying data distribution. Second, to account for the realistic scenario of using a sub-optimal discriminator, we derive a sequential Monte Carlo algorithm which iteratively takes the predictions from the discrimiator into account during the generation process. We test these approaches on the task of generating molecular graphs and show how the discriminator improves the generative performance over using only the pretrained model.

    @inproceedings{KelviniusL:2024,
      author = {Kelvinius, Filip Ekström and Lindsten, Fredrik},
      title = {Discriminator Guidance for Autoregressive Diffusion Models},
      booktitle = {Proceedings of the  27th International Conference on Artificial Intelligence and Statistics},
      year = {2024},
      address = {Valencia, Spain},
      month = may,
      pid = {C60}
    }
    
  3. [C59] A. Olmin, J. Lindqvist, L. Svensson, and F. Lindsten, On the connection between Noise-Contrastive Estimation and Contrastive Divergence, in Proceedings of the 27th International Conference on Artificial Intelligence and Statistics, Valencia, Spain, 2024.

    Noise-contrastive estimation (NCE) is a popular method for estimating unnormalised probabilistic models, such as energy-based models, which are effective for modelling complex data distributions. Unlike classical maximum likelihood (ML) estimation that relies on importance sampling (resulting in ML-IS) or MCMC resulting in contrastive divergence, CD), NCE uses a proxy criterion to avoid the need for evaluating an often intractable normalisation constant. Despite apparent conceptual differences, we show that two NCE criteria, ranking NCE (RNCE) and conditional NCE (CNCE), can be viewed as ML estimation methods. Specifically, RNCE is equivalent to ML estimation combined with conditional importance sampling, and both RNCE and CNCE are special cases of CD. These findings bridge the gap between the two method classes and allow us to apply techniques from the ML-IS and CD literature to NCE, offering several advantageous extensions.

    @inproceedings{OlminLSL:2024,
      author = {Olmin, Amanda and Lindqvist, Jakob and Svensson, Lennart and Lindsten, Fredrik},
      title = {On the connection between Noise-Contrastive Estimation and Contrastive Divergence},
      booktitle = {Proceedings of the  27th International Conference on Artificial Intelligence and Statistics},
      year = {2024},
      address = {Valencia, Spain},
      month = may,
      url = {https://arxiv.org/abs/2402.16688},
      pid = {C59}
    }
    
  4. [C58] A. Ahmadian, Y. Ding, G. Eilertsen, and F. Lindsten, Unsupervised Novelty Detection in Pretrained Representation Space with Locally Adapted Likelihood Ratio, in Proceedings of the 27th International Conference on Artificial Intelligence and Statistics, Valencia, Spain, 2024.

    Detecting novelties given unlabeled examples of normal data is a challenging task in machine learning, particularly when the novel and normal categories are semantically close. Large deep models pretrained on massive datasets can provide a rich representation space in which the simple k-nearest neighbor distance works as a novelty measure. However, as we show in this paper, the basic k-NN method might fail in this setup due to ignoring the ‘local geometry’ of the distribution over representations as well as the impact of irrelevant ’background features’. To address this, we propose a fully unsupervised novelty detection approach that integrates the flexibility of k-NN with a locally adapted scaling of dimensions based on the ’neighbors of nearest neighbor’ and the idea of ’likelihood ratio’ in pretrained (self-supervised) representation spaces. Our experiments with image data show the advantage of this method when off-the-shelf vision transformers (e.g., pretrained by DINO) are used as the feature extractor without any fine-tuning.

    @inproceedings{AhmadianDEL:2024,
      author = {Ahmadian, Amirhossein and Ding, Yifan and Eilertsen, Gabriel and Lindsten, Fredrik},
      title = {Unsupervised Novelty Detection in Pretrained Representation Space with Locally Adapted Likelihood Ratio},
      booktitle = {Proceedings of the  27th International Conference on Artificial Intelligence and Statistics},
      year = {2024},
      address = {Valencia, Spain},
      month = may,
      pid = {C58}
    }
    
  5. [C57] J. Oskarsson, T. Landelius, and F. Lindsten, Graph-based Neural Weather Prediction for Limited Area Modeling, in NeurIPS 2023 Workshop on Tackling Climate Change with Machine Learning, 2023.
    @inproceedings{OskarssonLL:2023,
      title = {Graph-based Neural Weather Prediction for Limited Area Modeling},
      author = {Oskarsson, Joel and Landelius, Tomas and Lindsten, Fredrik},
      booktitle = {NeurIPS 2023 Workshop on Tackling Climate Change with Machine Learning},
      year = {2023},
      month = dec,
      url = {https://arxiv.org/abs/2309.17370},
      pid = {C57}
    }
    
  6. [C56] J. Lindqvist, A. Olmin, L. Svensson, and F. Lindsten, Generalised Active Learning with Annotation Quality Selection, in Proceedings of the 33rd IEEE International Workshop on Machine Learning for Signal Processing (MLSP), Rome, Italy, 2023.

    In this paper we promote a general formulation of active learning (AL), wherein the typically binary decision to annotate a point or not is extended to selecting the qualities with which the points should be annotated. By linking the annotation quality to the cost of acquiring the label, we can trade a lower quality for a larger set of training samples, which may improve learning for the same annotation cost. To investigate this AL formulation, we introduce a concrete criterion, based on the mutual information (MI) between model parameters and noisy labels, for selecting annotation qualities for the entire dataset, before any labels are acquired. We illustrate the usefulness of our formulation with examples for both classification and regression and find that MI is a good candidate for a criterion, but its complexity limits its usefulness.

    @inproceedings{LindqvistOSL:2023,
      author = {Lindqvist, Jakob and Olmin, Amanda and Svensson, Lennart and Lindsten, Fredrik},
      title = {Generalised Active Learning with Annotation Quality Selection},
      booktitle = {Proceedings of the  33rd {IEEE} International Workshop on Machine Learning for Signal Processing ({MLSP})},
      year = {2023},
      address = {Rome, Italy},
      month = sep,
      pid = {C56}
    }
    
  7. [C55] J. Varga, E. Karlsson, G. Raidl, E. Rönnberg, F. Lindsten, and T. Rodemann, Speeding up Logic-Based Benders Decomposition by Strengthening Cuts with Graph Neural Networks, in Proceedings of the the 9th International Conference on Machine Learning, Optimization and Data science (LOD), Grasmere, UK, 2023.
    @inproceedings{VargaKRRLR:2023,
      author = {Varga, Johannes and Karlsson, Emil and Raidl, Günther and Rönnberg, Elina and Lindsten, Fredrik and Rodemann, Tobias},
      title = {Speeding up Logic-Based Benders Decomposition by Strengthening Cuts with Graph Neural Networks},
      booktitle = {Proceedings of the the 9th International Conference on Machine Learning, Optimization and Data science (LOD)},
      year = {2023},
      address = {Grasmere, UK},
      month = sep,
      pid = {C55}
    }
    
  8. [C54] P. Glaser, D. Widmann, F. Lindsten, and A. Gretton, Fast and Scalable Score-Based Kernel Calibration Tests, in Proceedings of the 39th Conference on Uncertainty in Artificial Intelligence (UAI), Pittsburgh, USA, 2023. [Spotlight]
    URL  — Abstract — BibTeX

    We introduce the Kernel Calibration Conditional Stein Discrepancy test (KCCSD test), a nonparametric, kernel-based test for assessing the calibration of probabilistic models with well-defined scores. In contrast to previous methods, our test avoids the need for possibly expensive expectation approximations while providing control over its type-I error. We achieve these improvements by using a new family of kernels for score-based probabilities that can be estimated without probability density samples, and by using a Conditional Goodness of Fit criterion for the KCCSD test’s U-statistic. We demonstrate the properties of our test on various synthetic settings.

    @inproceedings{GlaserWLG:2023,
      author = {Glaser, Pierre and Widmann, David and Lindsten, Fredrik and Gretton, Arthur},
      title = {Fast and Scalable Score-Based Kernel Calibration Tests},
      booktitle = {Proceedings of the  39th Conference on Uncertainty in Artificial Intelligence ({UAI})},
      year = {2023},
      address = {Pittsburgh, USA},
      month = aug,
      url = {https://proceedings.mlr.press/v216/glaser23a.html},
      pid = {C54},
      comment = {Spotlight}
    }
    
  9. [C53] A. Ahmadian and F. Lindsten, Enhancing Representation Learning with Deep Classifiers in Presence of Shortcut, in Proceedings of the 48th IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 2023.

    A deep neural classifier trained on an upstream task can be leveraged to boost the performance of another classifier in a related downstream task through the representations learned in hidden layers. However, presence of a shortcut (easy-to-learn feature) in the upstream task can considerably impair the versatility of intermediate representations and, in turn, the downstream performance. In this paper, we propose a method to improve the representations learned by deep neural image classifiers in spite of shortcuts in upstream data. In our method, the upstream classification objective is augmented with a type of adversarial training where an auxiliary network, so called lens, fools the classifier by exploiting the shortcut in reconstructing the images. Empirical comparisons in self-supervised and transfer learning tasks on three shortcut-biased datasets suggest the advantages of our method in terms of downstream performance and/or training time.

    @inproceedings{AhmadianL:2023,
      author = {Ahmadian, Amirhossein and Lindsten, Fredrik},
      title = {Enhancing Representation Learning with Deep Classifiers in Presence of Shortcut},
      booktitle = {Proceedings of the  48th {IEEE} International Conference on Acoustics, Speech and Signal Processing
      	({ICASSP})},
      address = {Rhodes Island, Greece},
      year = {2023},
      month = jun,
      pid = {C53}
    }
    
  10. [C52] H. Govindarajan, P. Sidén, J. Roll, and F. Lindsten, DINO as a von Mises-Fisher mixture model, in International Conference on Learning Representations, 2023. [ICLR Notable Paper Top 25 %]
    URL  — Abstract — BibTeX

    Self-distillation methods using Siamese networks are popular for self-supervised pre-training. DINO is one such method based on a cross-entropy loss between K-dimensional probability vectors, obtained by applying a softmax function to the dot product between representations and learnt prototypes. Given the fact that the learned representations are L2-normalized, we show that DINO can be interpreted as a mixture model of von Mises-Fisher components. With this interpretation, DINO assumes equal precision for all components when the prototypes are also L2-normalized. Using this insight we propose DINO-vMF, that adds appropriate normalization constants when computing the cluster assignment probabilities. Unlike DINO, DINO-vMF is stable also for the larger ViT-Base model with unnormalized prototypes. We show that the added flexibility of the mixture model is beneficial in terms of better image representations. The DINO-vMF pre-trained model consistently performs better than DINO on a range of downstream tasks.

    @inproceedings{GovindarajanSRL:2022,
      author = {Govindarajan, Hariprasath and Sidén, Per and Roll, Jacob and Lindsten, Fredrik},
      title = {{DINO} as a von {M}ises-{F}isher mixture model},
      booktitle = {International Conference on Learning Representations},
      year = {2023},
      url = {https://openreview.net/forum?id=cMJo1FTwBTQ},
      pid = {C52},
      comment = {ICLR Notable Paper Top 25 %}
    }
    
  11. [C51] J. Oskarsson, P. Sidén, and F. Lindsten, Temporal Graph Neural Networks for Irregular Data, in Proceedings of the 26th International Conference on Artificial Intelligence and Statistics, 2023.

    This paper proposes a temporal graph neural network model for forecasting of graph-structured irregularly observed time series. Our TGNN4I model is designed to handle both irregular time steps and partial observations of the graph. This is achieved by introducing a time-continuous latent state in each node, following a linear Ordinary Differential Equation (ODE) defined by the output of a Gated Recurrent Unit (GRU). The ODE has an explicit solution as a combination of exponential decay and periodic dynamics. Observations in the graph neighborhood are taken into account by integrating graph neural network layers in both the GRU state update and predictive model. The time-continuous dynamics additionally enable the model to make predictions at arbitrary time steps. We propose a loss function that leverages this and allows for training the model for forecasting over different time horizons. Experiments on simulated data and real-world data from traffic and climate modeling validate the usefulness of both the graph structure and time-continuous dynamics in settings with irregular observations.

    @inproceedings{OskarssonSL:2023,
      author = {Oskarsson, Joel and Sidén, Per and Lindsten, Fredrik},
      title = {Temporal Graph Neural Networks for Irregular Data},
      booktitle = {Proceedings of the  26th International Conference on Artificial Intelligence and Statistics},
      year = {2023},
      url = {https://arxiv.org/abs/2302.08415},
      pid = {C51}
    }
    
  12. [C50] A. Olmin, J. Lindqvist, L. Svensson, and F. Lindsten, Active Learning with Weak Supervision for Gaussian Processes, in Proceedings of the 29th International Conference on Neural Information Processing (ICONIP), 2022.

    Annotating data for supervised learning can be costly. When the annotation budget is limited, active learning can be used to select and annotate those observations that are likely to give the most gain in model performance. We propose an active learning algorithm that, in addition to selecting which observation to annotate, selects the precision of the annotation that is acquired. Assuming that annotations with low precision are cheaper to obtain, this allows the model to explore a larger part of the input space, with the same annotation costs. We build our acquisition function on the previously proposed BALD objective for Gaussian Processes, and empirically demonstrate the gains of being able to adjust the annotation precision in the active learning loop.

    @inproceedings{OlminLSL:2022,
      author = {Olmin, Amanda and Lindqvist, Jakob and Svensson, Lennart and Lindsten, Fredrik},
      title = {Active Learning with Weak Supervision for Gaussian Processes},
      booktitle = {Proceedings of the  29th International Conference on Neural Information Processing (ICONIP)},
      year = {2022},
      url = {https://arxiv.org/abs/2204.08335},
      pid = {C50}
    }
    
  13. [C49] J. Oskarsson, P. Sidén, and F. Lindsten, Scalable Deep Gaussian Markov Random Fields for General Graphs, in Proceedings of the 39th International Conference on Machine Learning, 2022.
    URL  — Abstract — BibTeX

    Machine learning methods on graphs have proven useful in many applications due to their ability to handle generally structured data. The framework of Gaussian Markov Random Fields (GMRFs) provides a principled way to define Gaussian models on graphs by utilizing their sparsity structure. We propose a flexible GMRF model for general graphs built on the multi-layer structure of Deep GMRFs, originally proposed for lattice graphs only. By designing a new type of layer we enable the model to scale to large graphs. The layer is constructed to allow for efficient training using variational inference and existing software frameworks for Graph Neural Networks. For a Gaussian likelihood, close to exact Bayesian inference is available for the latent field. This allows for making predictions with accompanying uncertainty estimates. The usefulness of the proposed model is verified by experiments on a number of synthetic and real world datasets, where it compares favorably to other both Bayesian and deep learning methods.

    @inproceedings{OskarssonSL:2022,
      author = {Oskarsson, Joel and Sidén, Per and Lindsten, Fredrik},
      title = {Scalable Deep {G}aussian {M}arkov Random Fields for General Graphs},
      booktitle = {Proceedings of the  39th International Conference on Machine Learning},
      year = {2022},
      url = {https://proceedings.mlr.press/v162/oskarsson22a.html},
      pid = {C49}
    }
    
  14. [C48] A. Olmin and F. Lindsten, Robustness and reliability when training with noisy labels, in Proceedings of the 25th International Conference on Artificial Intelligence and Statistics, 2022, vol. 151, pp. 922–942.
    URL  — Abstract — BibTeX

    Labelling of data for supervised learning can be costly and time-consuming and the risk of incorporating label noise in large data sets is imminent. When training a flexible discriminative model using a strictly proper loss, such noise will inevitably shift the solution towards the conditional distribution over noisy labels. Nevertheless, while deep neural networks have proven capable of fitting random labels, regularisation and the use of robust loss functions empirically mitigate the effects of label noise. However, such observations concern robustness in accuracy, which is insufficient if reliable uncertainty quantification is critical. We demonstrate this by analysing the properties of the conditional distribution over noisy labels for an input-dependent noise model. In addition, we evaluate the set of robust loss functions characterised by noise-insensitive, asymptotic risk minimisers. We find that strictly proper and robust loss functions both offer asymptotic robustness in accuracy, but neither guarantee that the final model is calibrated. Moreover, even with robust loss functions, overfitting is an issue in practice. With these results, we aim to explain observed robustness of common training practices, such as early stopping, to label noise. In addition, we aim to encourage the development of new noise-robust algorithms that not only preserve accuracy but that also ensure reliability.

    @inproceedings{OlminL:2022,
      title = {Robustness and reliability when training with noisy labels},
      author = {Olmin, Amanda and Lindsten, Fredrik},
      booktitle = {Proceedings of the  25th International Conference on Artificial Intelligence and Statistics},
      pages = {922--942},
      editor = {Camps-Valls, Gustau and Ruiz, Francisco J. R. and Valera, Isabel},
      volume = {151},
      series = {Proceedings of Machine Learning Research},
      month = {28--30 Mar},
      publisher = {PMLR},
      pdf = {https://proceedings.mlr.press/v151/olmin22a/olmin22a.pdf},
      url = {https://proceedings.mlr.press/v151/olmin22a.html},
      year = {2022},
      pid = {C48}
    }
    
  15. [C47] H. Govindarajan, P. Lindskog, D. Lundström, A. Olmin, J. Roll, and F. Lindsten, Self-Supervised Representation Learning for Content Based Image Retrieval of Complex Scenes, in 32nd IEEE Intelligent Vehicles Symposium – Workshop on Data Driven Intelligent Vehicle Applications, 2021.
    DOI  — Abstract — BibTeX

    Although Content Based Image Retrieval (CBIR) is an active research field, application to images simultaneously containing multiple objects has received limited research interest. For such complex images, it is difficult to precisely convey the query intention, to encode all the image aspects into one compact global feature representation and to unambiguously define label similarity or dissimilarity. Motivated by the recent success on many visual benchmark tasks, we propose a selfsupervised method to train a feature representation learning model. We propose usage of multiple query images, and use an attention based architecture to extract features from diverse image aspects that benefits from this. The method shows promising performance on road scene datasets, and, consistently improves when multiple query images are used instead of a single query image.

    @inproceedings{GovindarajanLLORL:2021,
      title = {Self-Supervised Representation Learning for Content Based Image Retrieval of Complex Scenes},
      author = {Govindarajan, Hariprasath and Lindskog, Peter and Lundstr\"om, Dennis and Olmin, Amanda and Roll, Jacob and Lindsten, Fredrik},
      booktitle = {32nd IEEE Intelligent Vehicles Symposium -- Workshop on Data Driven Intelligent Vehicle Applications},
      year = {2021},
      doi = {10.1109/IVWorkshops54471.2021.9669246},
      pid = {C47}
    }
    
  16. [C46] A. Ahmadian and F. Lindsten, Likelihood-free Out-of-Distribution Detection with Invertible Generative Models, in Proceedings of the 30th International Joint Conference on Artificial Intelligence (IJCAI), 2021.
    DOI  — Abstract — BibTeX

    Likelihood of generative models has been used traditionally as a score to detect atypical (Out-of-Distribution, OOD) inputs. However, several recent studies have found this approach to be highly unreliable, even with invertible generative models, where computing likelihood is feasible. In this paper, we present a different framework for generative model–based OOD detection that employs the model in constructing a new representation space, instead of using it directly in computing typicality scores, emphasizing that the score function should be interpretable as the similarity between the input and training data in the new space. In practice, with a focus on invertible models, we propose to extract low-dimensional features (statistics) based on the model encoder and complexity of input images, and then use a One-Class SVM to score the data. Contrary to recently proposed OOD detection methods for generative models, our method does not require computing likelihood values. Consequently, it is much faster when using invertible models with iteratively approximated likelihood (e.g. iResNet), while it still has a performance competitive with other related methods.

    @inproceedings{AhmadianL:2021,
      title = {Likelihood-free Out-of-Distribution Detection with Invertible Generative Models},
      author = {Ahmadian, Amirhossein and Lindsten, Fredrik},
      booktitle = {Proceedings of the  30th International Joint Conference on Artificial Intelligence ({IJCAI})},
      doi = {10.24963/ijcai.2021/292},
      year = {2021},
      pid = {C46}
    }
    
  17. [C45] D. Widmann, F. Lindsten, and D. Zachariah, Calibration tests beyond classification, in International Conference on Learning Representations, 2021.
    URL  — Abstract — BibTeX

    Most supervised machine learning tasks are subject to irreducible prediction errors. Probabilistic predictive models address this limitation by providing probability distributions that represent a belief over plausible targets, rather than point estimates. Such models can be a valuable tool in decision-making under uncertainty, provided that the model output is meaningful and interpretable. Calibrated models guarantee that the probabilistic predictions are neither over- nor under-confident. In the machine learning literature, different measures and statistical tests have been proposed and studied for evaluating the calibration of classification models. For regression problems, however, research has been focused on a weaker condition of calibration based on predicted quantiles for real-valued targets. In this paper, we propose the first framework that unifies calibration evaluation and tests for probabilistic predictive models. It applies to any such model, including classification and regression models of arbitrary dimension. Furthermore, the framework generalizes existing measures and provides a more intuitive reformulation of a recently proposed framework for calibration in multi-class classification.

    @inproceedings{WidmannLZ:2021,
      title = {Calibration tests beyond classification},
      author = {Widmann, David and Lindsten, Fredrik and Zachariah, Dave},
      booktitle = {International Conference on Learning Representations},
      year = {2021},
      url = {https://openreview.net/forum?id=-bxf89v3Nx},
      pid = {C45}
    }
    
  18. [C44] C. A. Naesseth, F. Lindsten, and D. M. Blei, Markovian Score Climbing: Variational Inference with KL(p||q), in Advances in Neural Information Processing Systems 33, 2020.
    URL  — Abstract — BibTeX

    Modern variational inference (VI) uses stochastic gradients to avoid intractable expectations, enabling large-scale probabilistic inference in complex models. VI posits a family of approximating distributions q and then finds the member of that family that is closest to the exact posterior p. Traditionally, VI algorithms minimize the “exclusive Kullback-Leibler (KL)” KL(q||p), often for computational convenience. Recent research, however, has also focused on the “inclusive KL” KL(p||q), which has good statistical properties that makes it more appropriate for certain inference problems. This paper develops a simple algorithm for reliably minimizing the inclusive KL. Consider a valid Markov chain Monte Carlo (MCMC) method, a Markov chain whose stationary distribution is p. The algorithm we develop iteratively samples the chain, and then uses those samples to follow the score function of the variational approximation with a Robbins-Monro stepsize schedule. This method, which we call Markovian score climbing (MSC), converges to a local optimum of the inclusive KL. It does not suffer from the systematic errors inherent in existing methods, such as Reweighted Wake-Sleep and Neural Adaptive Sequential Monte Carlo, which lead to bias in their final estimates. In a variant that ties the variational approximation directly to the Markov chain, MSC further provides a new algorithm that melds VI and MCMC. We illustrate convergence on a toy model and demonstrate the utility of MSC on Bayesian probit regression for classification as well as a stochastic volatility model for financial data.

    @incollection{NaessethLB:2020,
      author = {Naesseth, Christian A. and Lindsten, Fredrik and Blei, David M.},
      title = {Markovian Score Climbing: Variational Inference with KL(p||q)},
      booktitle = {Advances in Neural Information Processing Systems 33},
      url = {https://proceedings.neurips.cc/paper/2020/hash/b20706935de35bbe643733f856d9e5d6-Abstract.html},
      year = {2020},
      pid = {C44}
    }
    
  19. [C43] J. Lindqvist, A. Olmin, F. Lindsten, and L. Svensson, A general framework for ensemble distribution distillation, in Proceedings of the 30th IEEE International Workshop on Machine Learning for Signal Processing (MLSP), Virtual Conference, 2020.

    Ensembles of neural networks have been shown to give better performance than single networks, both in terms of predictions and uncertainty estimation. Additionally, ensembles allow the uncertainty to be decomposed into aleatoric (data) and epistemic (model) components, giving a more complete picture of the predictive uncertainty. Ensemble distillation is the process of compressing an ensemble into a single model, often resulting in a leaner model that still outperforms the individual ensemble members. Unfortunately, standard distillation erases the natural uncertainty decomposition of the ensemble. We present a general framework for distilling both regression and classification ensembles in a way that preserves the decomposition. We demonstrate the desired behaviour of our framework and show that its predictive performance is on par with standard distillation.

    @inproceedings{LindqvistOLS:2020,
      author = {Lindqvist, Jakob and Olmin, Amanda and Lindsten, Fredrik and Svensson, Lennart},
      title = {A general framework for ensemble distribution distillation},
      booktitle = {Proceedings of the  30th {IEEE} International Workshop on Machine Learning for Signal Processing ({MLSP})},
      url = {https://arxiv.org/abs/2002.11531},
      year = {2020},
      address = {Virtual Conference},
      month = sep,
      pid = {C43}
    }
    
  20. [C42] P. Sidén and F. Lindsten, Deep Gaussian Markov random fields, in Proceedings of the 37th International Conference on Machine Learning, Virtual Conference, 2020.
    URL  — Abstract — BibTeX

    Gaussian Markov random fields (GMRFs) are probabilistic graphical models widely used in spatial statistics and related fields to model dependencies over spatial structures. We establish a formal connection between GMRFs and convolutional neural networks (CNNs). Common GMRFs are special cases of a generative model where the inverse mapping from data to latent variables is given by a 1-layer linear CNN. This connection allows us to generalize GMRFs to multi-layer CNN architectures, effectively increasing the order of the corresponding GMRF in a way which has favorable computational scaling. We describe how well-established tools, such as autodiff and variational inference, can be used for simple and efficient inference and learning of the deep GMRF. We demonstrate the flexibility of the proposed model and show that it outperforms the state-of-the-art on a dataset of satellite temperatures, in terms of prediction and predictive uncertainty.

    @inproceedings{SidenL:2020,
      author = {Sidén, Per and Lindsten, Fredrik},
      title = {Deep {G}aussian {M}arkov random fields},
      booktitle = {Proceedings of the  37th International Conference on Machine Learning},
      url = {https://proceedings.mlr.press/v119/siden20a.html},
      year = {2020},
      address = {Virtual Conference},
      month = jul,
      pid = {C42}
    }
    
  21. [C41] J. Kudlicka, L. M. Murray, T. B. Schön, and F. Lindsten, Particle filter with rejection control and unbiased estimator of the marginal likelihood, in Proceedings of the 45th IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 2020.

    We consider the combined use of resampling and partial rejection control in sequential Monte Carlo methods, also known as particle filters. While the variance reducing properties of rejection control are known, there has not been (to the best of our knowledge) any work on unbiased estimation of the marginal likelihood (also known as the model evidence or the normalizing constant) in this type of particle filters. Being able to estimate the marginal likelihood without bias is highly relevant for model comparison, computation of interpretable and reliable confidence intervals, and in exact approximation methods, such as particle Markov chain Monte Carlo. In the paper we present a particle filter with rejection control that enables unbiased estimation of the marginal likelihood.

    @inproceedings{KudlickaMSL:2020,
      author = {Kudlicka, Jan and Murray, Lawrence M. and Schön, Thomas B. and Lindsten, Fredrik},
      title = {Particle filter with rejection control and unbiased estimator of the marginal likelihood},
      booktitle = {Proceedings of the  45th {IEEE} International Conference on Acoustics, Speech and Signal Processing
      	({ICASSP})},
      address = {Barcelona, Spain},
      year = {2020},
      month = dec,
      url = {https://arxiv.org/abs/1910.09527},
      pid = {C41}
    }
    
  22. [C40] A. Wigren, R. S. Risuleo, L. Murray, and F. Lindsten, Parameter elimination in particle Gibbs sampling, in Advances in Neural Information Processing Systems 32, H. Wallach, H. Larochelle, A. Beygelzimer, F. d’ Alché-Buc, E. Fox, and R. Garnett, Eds. Curran Associates, Inc., 2019, pp. 8916–8927. [NeurIPS Oral]
    URL  — Abstract — BibTeX

    Bayesian inference in state-space models is challenging due to high-dimensional state trajectories. A viable approach is particle Markov chain Monte Carlo (PMCMC), combining MCMC and sequential Monte Carlo to form exact approximations” to otherwise-intractable MCMC methods. The performance of the approximation is limited to that of the exact method. We focus on particle Gibbs (PG) and particle Gibbs with ancestor sampling (PGAS), improving their performance beyond that of the ideal Gibbs sampler (which they approximate) by marginalizing out one or more parameters. This is possible when the parameter(s) has a conjugate prior relationship with the complete data likelihood. Marginalization yields a non-Markov model for inference, but we show that, in contrast to the general case, the methods still scale linearly in time. While marginalization can be cumbersome to implement, recent advances in probabilistic programming have enabled its automation. We demonstrate how the marginalized methods are viable as efficient inference backends in probabilistic programming, and demonstrate with examples in ecology and epidemiology.

    @incollection{WigrenRML:2019,
      author = {Wigren, Anna and Risuleo, Riccardo Sven and Murray, Lawrence and Lindsten, Fredrik},
      title = {Parameter elimination in particle {G}ibbs sampling},
      booktitle = {Advances in Neural Information Processing Systems 32},
      publisher = {Curran Associates, Inc.},
      year = {2019},
      editor = {Wallach, H. and Larochelle, H. and Beygelzimer, A. and d\textquotesingle Alch\'{e}-Buc, F. and Fox, E. and Garnett, R.},
      pages = {8916--8927},
      comment = {NeurIPS Oral},
      pid = {C40},
      url = {http://papers.nips.cc/paper/9094-parameter-elimination-in-particle-gibbs-sampling}
    }
    
  23. [C39] D. Widmann, F. Lindsten, and D. Zachariah, Calibration tests in multi-class classification: A unifying framework, in Advances in Neural Information Processing Systems 32, H. Wallach, H. Larochelle, A. Beygelzimer, F. d’ Alché-Buc, E. Fox, and R. Garnett, Eds. Curran Associates, Inc., 2019, pp. 12236–12246. [NeurIPS Spotlight]
    URL  — Abstract — BibTeX

    In safety-critical applications a probabilistic model is usually required to be calibrated, i.e., to capture the uncertainty of its predictions accurately. In multi-class classification, calibration of the most confident predictions only is often not sufficient. We propose and study calibration measures for multi-class classification that generalize existing measures such as the expected calibration error, the maximum calibration error, and the maximum mean calibration error. We propose and evaluate empirically different consistent and unbiased estimators for a specific class of measures based on matrix-valued kernels. Importantly, these estimators can be interpreted as test statistics associated with well-defined bounds and approximations of the p-value under the null hypothesis that the model is calibrated, significantly improving the interpretability of calibration measures, which otherwise lack any meaningful unit or scale.

    @incollection{WidmannLZ:2019,
      author = {Widmann, David and Lindsten, Fredrik and Zachariah, Dave},
      title = {Calibration tests in multi-class classification: A unifying framework},
      booktitle = {Advances in Neural Information Processing Systems 32},
      publisher = {Curran Associates, Inc.},
      year = {2019},
      editor = {Wallach, H. and Larochelle, H. and Beygelzimer, A. and d\textquotesingle Alch\'{e}-Buc, F. and Fox, E. and Garnett, R.},
      pages = {12236--12246},
      comment = {NeurIPS Spotlight},
      pid = {C39},
      url = {http://papers.nips.cc/paper/9392-calibration-tests-in-multi-class-classification-a-unifying-framework}
    }
    
  24. [C38] C. Nemeth, F. Lindsten, M. Filippone, and J. Hensman, Pseudo-Extended Markov chain Monte Carlo, in Advances in Neural Information Processing Systems 32, H. Wallach, H. Larochelle, A. Beygelzimer, F. d’ Alché-Buc, E. Fox, and R. Garnett, Eds. Curran Associates, Inc., 2019, pp. 4314–4324.
    URL  — Abstract — BibTeX

    Sampling from posterior distributions using Markov chain Monte Carlo (MCMC) methods can require an exhaustive number of iterations, particularly when the posterior is multi-modal as the MCMC sampler can become trapped in a local mode for a large number of iterations. In this paper, we introduce the pseudo-extended MCMC method as a simple approach for improving the mixing of the MCMC sampler for multi-modal posterior distributions. The pseudo-extended method augments the state-space of the posterior using pseudo-samples as auxiliary variables. On the extended space, the modes of the posterior are connected, which allows the MCMC sampler to easily move between well-separated posterior modes. We demonstrate that the pseudo-extended approach delivers improved MCMC sampling over the Hamiltonian Monte Carlo algorithm on multi-modal posteriors, including Boltzmann machines and models with sparsity-inducing priors.

    @incollection{NemethLFH:2019,
      author = {Nemeth, Christopher and Lindsten, Fredrik and Filippone, Maurizio and Hensman, James},
      title = {Pseudo-Extended {M}arkov chain {M}onte {C}arlo},
      booktitle = {Advances in Neural Information Processing Systems 32},
      publisher = {Curran Associates, Inc.},
      year = {2019},
      editor = {Wallach, H. and Larochelle, H. and Beygelzimer, A. and d\textquotesingle Alch\'{e}-Buc, F. and Fox, E. and Garnett, R.},
      pages = {4314--4324},
      pid = {C38},
      url = {https://papers.nips.cc/paper/8683-pseudo-extended-markov-chain-monte-carlo}
    }
    
  25. [C37] J. Umenberger, T. B. Schön, and F. Lindsten, Bayesian identification of state-space models via adaptive thermostats, in Proceedings of the 58th IEEE Conference on Decision and Control (CDC), Nice, France, 2019.
    @inproceedings{UmenbergerSL:2019,
      author = {Umenberger, Jack and Schön, Thomas B. and Lindsten, Fredrik},
      title = {Bayesian identification of state-space models via adaptive thermostats},
      booktitle = {Proceedings of the  58th {IEEE} Conference on Decision and Control ({CDC})},
      year = {2019},
      address = {Nice, France},
      month = dec,
      pid = {C37}
    }
    
  26. [C36] J. Vaicenavicius, D. Widmann, C. Andersson, F. Lindsten, J. Roll, and T. B. Schön, Evaluating model calibration in classification, in Proceedings of the 22nd International Conference on Artificial Intelligence and Statistics, Naha, Okinawa, Japan, 2019.
    @inproceedings{VaicenaviciusWALRS:2019,
      author = {Vaicenavicius, Juozas and Widmann, David and Andersson, Carl and Lindsten, Fredrik and Roll, Jacob and Schön, Thomas B.},
      title = {Evaluating model calibration in classification},
      booktitle = {Proceedings of the  22nd International Conference on Artificial Intelligence and Statistics},
      year = {2019},
      address = {Naha, Okinawa, Japan},
      month = apr,
      pid = {C36}
    }
    
  27. [C35] F. Lindsten, J. Helske, and M. Vihola, Graphical model inference: Sequential Monte Carlo meets deterministic approximations, in Advances in Neural Information Processing Systems 31, S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett, Eds. Curran Associates, Inc., 2018, pp. 8190–8200. [NeurIPS Spotlight]
    @incollection{LindstenHV:2018,
      author = {Lindsten, Fredrik and Helske, Jouni and Vihola, Matti},
      title = {Graphical model inference: Sequential {M}onte {C}arlo meets deterministic approximations},
      booktitle = {Advances in Neural Information Processing Systems 31},
      publisher = {Curran Associates, Inc.},
      year = {2018},
      editor = {Bengio, S. and Wallach, H. and Larochelle, H. and Grauman, K. and Cesa-Bianchi, N. and Garnett, R.},
      pages = {8190--8200},
      comment = {NeurIPS Spotlight},
      pid = {C35},
      url = {http://papers.nips.cc/paper/8041-graphical-model-inference-sequential-monte-carlo-meets-deterministic-approximations.pdf}
    }
    
  28. [C34] R. S. Risuleo, F. Lindsten, and H. Hjalmarsson, Semi-parametric kernel-based identification of Wiener systems, in Proceedings of the 57th IEEE Conference on Decision and Control (CDC), Miami Beach, FL, USA, 2018.
    @inproceedings{RisuleoLH:2018,
      author = {Risuleo, Riccardo S. and Lindsten, Fredrik and Hjalmarsson, Håkan},
      title = {Semi-parametric kernel-based identification of {W}iener systems},
      booktitle = {Proceedings of the  57th {IEEE} Conference on Decision and Control ({CDC})},
      year = {2018},
      address = {Miami Beach, FL, USA},
      month = dec,
      pid = {C34}
    }
    
  29. [C33] A. Wigren, L. Murray, and F. Lindsten, Improving the particle filter in high dimensions using conjugate artificial process noise, in Proceedings of the 18th IFAC Symposium on System Identification (SYSID), Stockholm, Sweden, 2018.
    @inproceedings{WigrenML:2018,
      author = {Wigren, Anna and Murray, Lawrence and Lindsten, Fredrik},
      title = {Improving the particle filter in high dimensions using conjugate artificial process noise},
      booktitle = {Proceedings of the  18th {IFAC} Symposium on System Identification ({SYSID})},
      year = {2018},
      address = {Stockholm, Sweden},
      month = jul,
      pid = {C33}
    }
    
  30. [C32] A. Svensson, F. Lindsten, and T. B. Schön, Learning Nonlinear State-Space Models Using Smooth Particle-Filter-Based Likelihood Approximations, in Proceedings of the 18th IFAC Symposium on System Identification (SYSID), Stockholm, Sweden, 2018.
    @inproceedings{SvenssonLS:2018,
      author = {Svensson, Andreas and Lindsten, Fredrik and Schön, Thomas B.},
      title = {Learning Nonlinear State-Space Models Using Smooth Particle-Filter-Based Likelihood Approximations},
      booktitle = {Proceedings of the  18th {IFAC} Symposium on System Identification ({SYSID})},
      year = {2018},
      address = {Stockholm, Sweden},
      month = jul,
      pid = {C32}
    }
    
  31. [C31] T. Rainforth, C. A. Naesseth, F. Lindsten, B. Paige, J.-W. van de Meent, A. Doucet, and F. Wood, Interacting Particle Markov Chain Monte Carlo, in Proceedings of the 33rd International Conference on Machine Learning, New York, USA, 2016.
    @inproceedings{RainforthNLPMDW:2016,
      author = {Rainforth, Tom and Naesseth, Christian A. and Lindsten, Fredrik and Paige, Brooks and van de Meent, Jan-Willem and Doucet, Arnaud and Wood, Frank},
      title = {Interacting Particle {M}arkov Chain {M}onte Carlo},
      booktitle = {Proceedings of the  33rd International Conference on Machine Learning},
      year = {2016},
      address = {New York, USA},
      month = jun,
      url = {https://proceedings.mlr.press/v48/rainforth16.html},
      pid = {C31}
    }
    
  32. [C30] J. Wågberg, F. Lindsten, and T. B. Schön, Bayesian nonparametric identification of piecewise affine ARX systems, in Proceedings of the 17th IFAC Symposium on System Identification (SYSID), Beijing, China, 2015.
    @inproceedings{WagbergLS:2015,
      author = {Wågberg, Johan and Lindsten, Fredrik and Schön, Thomas B.},
      title = {Bayesian nonparametric identification of piecewise affine {ARX} systems},
      booktitle = {Proceedings of the  17th {IFAC} Symposium on System Identification ({SYSID})},
      year = {2015},
      address = {Beijing, China},
      month = oct,
      pid = {C30}
    }
    
  33. [C29] J. Dahlin, F. Lindsten, and T. B. Schön, Quasi-Newton particle Metropolis-Hastings applied to intractable likelihood models, in Proceedings of the 17th IFAC Symposium on System Identification (SYSID), Beijing, China, 2015.
    @inproceedings{DahlinLS:2015a,
      author = {Dahlin, Johan and Lindsten, Fredrik and Schön, Thomas B.},
      title = {Quasi-{N}ewton particle {M}etropolis-{H}astings applied to intractable likelihood models},
      booktitle = {Proceedings of the  17th {IFAC} Symposium on System Identification ({SYSID})},
      year = {2015},
      address = {Beijing, China},
      month = oct,
      pid = {C29}
    }
    
  34. [C28] M. Riabiz, F. Lindsten, and S. J. Godsill, Pseudo-Marginal MCMC for Parameter Estimation in Alpha-Stable Distributions, in Proceedings of the 17th IFAC Symposium on System Identification (SYSID), Beijing, China, 2015.
    @inproceedings{RiabizLG:2015,
      author = {Riabiz, Marina and Lindsten, Fredrik and Godsill, Simon J.},
      title = {Pseudo-Marginal {MCMC} for Parameter Estimation in Alpha-Stable Distributions},
      booktitle = {Proceedings of the  17th {IFAC} Symposium on System Identification ({SYSID})},
      year = {2015},
      address = {Beijing, China},
      month = oct,
      pid = {C28}
    }
    
  35. [C27] T. B. Schön, F. Lindsten, J. Dahlin, J. Wågberg, C. A. Naesseth, A. Svensson, and L. Dai, Sequential Monte Carlo Methods for System Identification, in Proceedings of the 17th IFAC Symposium on System Identification (SYSID), Beijing, China, 2015.
    @inproceedings{SchonLDWNSD:2105,
      author = {Schön, Thomas B. and Lindsten, Fredrik and Dahlin, Johan and Wågberg, Johan and Naesseth, Christian A. and Svensson, Andreas and Dai, Liang},
      title = {Sequential {M}onte {C}arlo Methods for System Identification},
      booktitle = {Proceedings of the  17th {IFAC} Symposium on System Identification ({SYSID})},
      year = {2015},
      address = {Beijing, China},
      month = oct,
      pid = {C27}
    }
    
  36. [C26] C. A. Naesseth, F. Lindsten, and T. B. Schön, Nested Sequential Monte Carlo Methods, in Proceedings of the 32nd International Conference on Machine Learning (ICML), Lille, France, 2015.
    @inproceedings{NaessethLS:2015,
      author = {Naesseth, Christian A. and Lindsten, Fredrik and Sch\"on, Thomas B.},
      title = {Nested Sequential {M}onte {C}arlo Methods},
      booktitle = {Proceedings of the  32nd International Conference on Machine Learning (ICML)},
      year = {2015},
      address = {Lille, France},
      month = jul,
      pid = {C26},
      howpublished = {arXiv.org, arXiv:1502.02536},
      owner = {lindsten},
      timestamp = {2015.03.13}
    }
    
  37. [C25] S. Lacoste-Julien, F. Lindsten, and F. Bach, Sequential Kernel Herding: Frank-Wolfe Optimization for Particle Filtering, in Proceedings of the 18th International Conference on Artificial Intelligence and Statistics, San Diego, USA, 2015.
    @inproceedings{Lacoste-JulienLB:2015,
      author = {Lacoste-Julien, Simon and Lindsten, Fredrik and Bach, Francis},
      title = {Sequential Kernel Herding: {F}rank-{W}olfe Optimization for Particle Filtering},
      booktitle = {Proceedings of the  18th International Conference on Artificial Intelligence and Statistics},
      year = {2015},
      address = {San Diego, USA},
      month = may,
      pid = {C25}
    }
    
  38. [C24] P. Bunch, F. Lindsten, and S. S. Singh, Particle Gibbs with refreshed backward simulation, in Proceedings of the 40th IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brisbane, Australia, 2015.
    @inproceedings{BunchLS:2015,
      author = {Bunch, Pete and Lindsten, Fredrik and Singh, Sumeetpal S.},
      title = {Particle {G}ibbs with refreshed backward simulation},
      booktitle = {Proceedings of the  40th {IEEE} International Conference on Acoustics, Speech and Signal Processing
      	({ICASSP})},
      year = {2015},
      address = {Brisbane, Australia},
      month = apr,
      pid = {C24},
      owner = {lindsten},
      timestamp = {2014.11.26}
    }
    
  39. [C23] C. A. Naesseth, F. Lindsten, and T. B. Schön, Sequential Monte Carlo for graphical models, in Advances in Neural Information Processing Systems (NIPS) 27, 2014, pp. 1862–1870.
    @incollection{NaessethLS:2014,
      author = {Naesseth, Christian A. and Lindsten, Fredrik and Sch\"on, Thomas B.},
      title = {Sequential {M}onte {C}arlo for graphical models},
      booktitle = {Advances in Neural Information Processing Systems ({NIPS}) 27},
      year = {2014},
      pages = {1862--1870},
      pid = {C23},
      owner = {lindsten},
      timestamp = {2014-02-18},
      url = {https://papers.nips.cc/paper/5570-sequential-monte-carlo-for-graphical-models}
    }
    
  40. [C22] C. A. Naesseth, F. Lindsten, and T. B. Schön, Capacity estimation of two-dimensional channels using Sequential Monte Carlo, in Proceedings of the 2014 IEEE Information Theory Workshop (ITW), Hobart, Tasmania, 2014.
    @inproceedings{NaessethLS:2014a,
      author = {Naesseth, Christian A. and Lindsten, Fredrik and Schön, Thomas B.},
      title = {Capacity estimation of two-dimensional channels using Sequential {M}onte {C}arlo},
      booktitle = {Proceedings of the  2014 IEEE Information Theory Workshop (ITW)},
      year = {2014},
      address = {Hobart, Tasmania},
      month = nov,
      pid = {C22}
    }
    
  41. [C21] A. Svensson, F. Lindsten, and T. B. Schön, Identification of jump Markov linear models using particle filters, in Proceedings of the 53rd IEEE Conference on Decision and Control (CDC), Los Angeles, USA, 2014.
    @inproceedings{SvenssonLS:2014,
      author = {Svensson, Andreas and Lindsten, Fredrik and Sch\"on, Thomas B.},
      title = {Identification of jump {M}arkov linear models using particle filters},
      booktitle = {Proceedings of the  53rd {IEEE} Conference on Decision and Control ({CDC})},
      year = {2014},
      address = {Los Angeles, USA},
      month = dec,
      url = {https://arxiv.org/abs/1409.7287},
      pid = {C21},
      owner = {lindsten},
      timestamp = {2014.09.02}
    }
    
  42. [C20] R. Frigola, F. Lindsten, T. B. Schön, and C. E. Rasmussen, Identification of Gaussian Process State-Space Models with Particle Stochastic Approximation EM, in Proceedings of the 19th IFAC World Congress, Cape Town, South Africa, 2014.
    @inproceedings{FrigolaLSR:2014,
      author = {Frigola, Roger and Lindsten, Fredrik and Sch\"on, Thomas B. and Rasmussen, Carl E.},
      title = {Identification of {G}aussian Process State-Space Models with Particle Stochastic Approximation {EM}},
      booktitle = {Proceedings of the  19th {IFAC} World Congress},
      year = {2014},
      address = {Cape Town, South Africa},
      month = aug,
      pid = {C20},
      owner = {lindsten},
      timestamp = {2014.02.19}
    }
    
  43. [C19] J. Dahlin and F. Lindsten, Particle filter-based Gaussian Process Optimisation for Parameter Inference, in Proceedings of the 19th IFAC World Congress, Cape Town, South Africa, 2014.
    @inproceedings{DahlinL:2014,
      author = {Dahlin, Johan and Lindsten, Fredrik},
      title = {Particle filter-based {G}aussian Process Optimisation for Parameter Inference},
      booktitle = {Proceedings of the  19th {IFAC} World Congress},
      year = {2014},
      address = {Cape Town, South Africa},
      month = aug,
      pid = {C19},
      owner = {lindsten},
      timestamp = {2014.02.13}
    }
    
  44. [C18] J. Dahlin, F. Lindsten, and T. B. Schön, Second-order Particle MCMC for Bayesian Parameter Inference, in Proceedings of the 19th IFAC World Congress, Cape Town, South Africa, 2014.
    @inproceedings{DahlinLS:2014,
      author = {Dahlin, Johan and Lindsten, Fredrik and Sch\"on, Thomas B.},
      title = {Second-order Particle {MCMC} for {B}ayesian Parameter Inference},
      booktitle = {Proceedings of the  19th {IFAC} World Congress},
      year = {2014},
      address = {Cape Town, South Africa},
      month = aug,
      pid = {C18},
      owner = {lindsten},
      timestamp = {2014.02.13}
    }
    
  45. [C17] F. Gunnarsson, F. Lindsten, and N. Carlsson, Particle Filtering for Network-Based Positioning Terrestrial Radio Networks, in Proceedings of the IET Conference on Data Fusion and Target Tracking, Liverpool, UK, 2014. [ISIF Best Paper Award]
    @inproceedings{GunnarssonLC:2014,
      author = {Gunnarsson, Fredrik and Lindsten, Fredrik and Carlsson, Niclas},
      title = {Particle Filtering for Network-Based Positioning Terrestrial Radio Networks},
      booktitle = {Proceedings of the  IET Conference on Data Fusion and Target Tracking},
      year = {2014},
      address = {Liverpool, UK},
      comment = {ISIF Best Paper Award},
      pid = {C17}
    }
    
  46. [C16] R. Frigola, F. Lindsten, T. B. Schön, and C. E. Rasmussen, Bayesian Inference and Learning in Gaussian Process State-Space Models with Particle MCMC, in Advances in Neural Information Processing Systems (NIPS) 26, 2013.
    @incollection{FrigolaLSR:2013,
      author = {Frigola, Roger and Lindsten, Fredrik and Sch\"on, Thomas B. and Rasmussen, Carl E.},
      title = {{B}ayesian Inference and Learning in {G}aussian Process State-Space Models with Particle {MCMC}},
      booktitle = {Advances in Neural Information Processing Systems ({NIPS}) 26},
      year = {2013},
      month = dec,
      pid = {C16},
      owner = {lindsten},
      timestamp = {2013.12.03}
    }
    
  47. [C15] F. Lindsten, An efficient stochastic approximation EM algorithm using conditional particle filters, in Proceedings of the 38th IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Vancouver, Canada, 2013.
    @inproceedings{Lindsten:2013a,
      author = {Lindsten, Fredrik},
      title = {An efficient stochastic approximation {EM} algorithm using conditional particle filters},
      booktitle = {Proceedings of the  38th {IEEE} International Conference on Acoustics, Speech and Signal Processing
      	({ICASSP})},
      year = {2013},
      address = {Vancouver, Canada},
      month = may,
      pid = {C15},
      owner = {lindsten},
      timestamp = {2012.12.17}
    }
    
  48. [C14] J. Dahlin, F. Lindsten, and T. B. Schön, Particle Metropolis Hastings using Langevin dynamics, in Proceedings of the 38th IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Vancouver, Canada, 2013.
    @inproceedings{DahlinLS:2013,
      author = {Dahlin, Johan and Lindsten, Fredrik and Sch\"on, Thomas B.},
      title = {Particle {M}etropolis {H}astings using {L}angevin dynamics},
      booktitle = {Proceedings of the  38th {IEEE} International Conference on Acoustics, Speech and Signal Processing
      	({ICASSP})},
      year = {2013},
      address = {Vancouver, Canada},
      month = may,
      pid = {C14},
      owner = {lindsten},
      timestamp = {2013.06.17}
    }
    
  49. [C13] F. Lindsten, P. Bunch, S. J. Godsill, and T. B. Schön, Rao-Blackwellized particle smoothers for mixed linear/nonlinear state-space models, in Proceedings of the 38th IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Vancouver, Canada, 2013.
    @inproceedings{LindstenBGS:2013,
      author = {Lindsten, Fredrik and Bunch, Pete and Godsill, Simon J. and Sch\"on, Thomas B.},
      title = {Rao-{B}lackwellized particle smoothers for mixed linear/nonlinear state-space models},
      booktitle = {Proceedings of the  38th {IEEE} International Conference on Acoustics, Speech and Signal Processing
      	({ICASSP})},
      year = {2013},
      address = {Vancouver, Canada},
      month = may,
      pid = {C13},
      owner = {lindsten},
      timestamp = {2012.12.08}
    }
    
  50. [C12] E. Taghavi, F. Lindsten, L. Svensson, and T. B. Schön, Adaptive stopping for fast particle smoothing, in Proceedings of the 38th IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Vancouver, Canada, 2013.
    @inproceedings{TaghaviLSS:2013,
      author = {Taghavi, Ehsan and Lindsten, Fredrik and Svensson, Lennart and Sch\"on, Thomas B.},
      title = {Adaptive stopping for fast particle smoothing},
      booktitle = {Proceedings of the  38th {IEEE} International Conference on Acoustics, Speech and Signal Processing
      	({ICASSP})},
      year = {2013},
      address = {Vancouver, Canada},
      month = may,
      pid = {C12},
      owner = {lindsten},
      timestamp = {2012.12.08}
    }
    
  51. [C11] F. Lindsten, M. I. Jordan, and T. B. Schön, Ancestor Sampling for Particle Gibbs, in Advances in Neural Information Processing Systems (NIPS) 25, P. Bartlett, F. C. N. Pereira, C. J. C. Burges, L. Bottou, and K. Q. Weinberger, Eds. 2012, pp. 2600–2608.
    @incollection{LindstenJS:2012,
      author = {Lindsten, Fredrik and Jordan, Michael I. and Sch\"on, Thomas B.},
      title = {Ancestor Sampling for Particle {G}ibbs},
      booktitle = {Advances in Neural Information Processing Systems ({NIPS}) 25},
      year = {2012},
      editor = {Bartlett, P. and Pereira, F. C. N. and Burges, C. J. C. and Bottou, L. and Weinberger, K. Q.},
      pages = {2600--2608},
      pid = {C11},
      owner = {lindsten},
      timestamp = {2013.06.25}
    }
    
  52. [C10] F. Lindsten, T. B. Schön, and M. I. Jordan, A semiparametric Bayesian approach to Wiener system identification, in Proceedings of the 16th IFAC Symposium on System Identification (SYSID), Brussels, Belgium, 2012.
    @inproceedings{LindstenSJ:2012,
      author = {Lindsten, Fredrik and Sch\"on, Thomas B. and Jordan, Michael I.},
      title = {A semiparametric {B}ayesian approach to {W}iener system identification},
      booktitle = {Proceedings of the  16th {IFAC} Symposium on System Identification ({SYSID})},
      year = {2012},
      address = {Brussels, Belgium},
      month = jul,
      pid = {C10},
      owner = {lindsten},
      timestamp = {2011.12.01}
    }
    
  53. [C9] F. Lindsten, T. B. Schön, and L. Svensson, A non-degenerate Rao-Black\-wellised particle filter for estimating static parameters in dynamical models, in Proceedings of the 16th IFAC Symposium on System Identification (SYSID), Brussels, Belgium, 2012.
    @inproceedings{LindstenSS:2012,
      author = {Lindsten, Fredrik and Sch\"on, Thomas B. and Svensson, Lennart},
      title = {A non-degenerate {R}ao-{B}lack\-wellised particle filter for estimating static parameters in dynamical models},
      booktitle = {Proceedings of the  16th {IFAC} Symposium on System Identification ({SYSID})},
      year = {2012},
      address = {Brussels, Belgium},
      month = jul,
      pid = {C9},
      owner = {lindsten},
      timestamp = {2013.06.17}
    }
    
  54. [C8] J. Dahlin, F. Lindsten, T. B. Schön, and A. Wills, Hierarchical Bayesian ARX models for robust inference, in Proceedings of the 16th IFAC Symposium on System Identification (SYSID), Brussels, Belgium, 2012.
    @inproceedings{DahlinLSW:2012,
      author = {Dahlin, Johan and Lindsten, Fredrik and Sch\"on, Thomas B. and Wills, Adrian},
      title = {Hierarchical {B}ayesian {ARX} models for robust inference},
      booktitle = {Proceedings of the  16th {IFAC} Symposium on System Identification ({SYSID})},
      year = {2012},
      address = {Brussels, Belgium},
      month = jul,
      pid = {C8},
      owner = {lindsten},
      timestamp = {2013.06.17}
    }
    
  55. [C7] A. Wills, T. B. Schön, F. Lindsten, and B. Ninness, Estimation of Linear Systems using a Gibbs Sampler, in Proceedings of the 16th IFAC Symposium on System Identification (SYSID), Brussels, Belgium, 2012.
    @inproceedings{WillsSLN:2012,
      author = {Wills, Adrian and Sch\"on, Thomas B. and Lindsten, Fredrik and Ninness, Brett},
      title = {Estimation of Linear Systems using a {G}ibbs Sampler},
      booktitle = {Proceedings of the  16th {IFAC} Symposium on System Identification ({SYSID})},
      year = {2012},
      address = {Brussels, Belgium},
      month = jul,
      pid = {C7},
      owner = {lindsten},
      timestamp = {2011.12.01}
    }
    
  56. [C6] F. Lindsten and T. B. Schön, On the use of backward simulation in the particle Gibbs sampler, in Proceedings of the 37th IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Kyoto, Japan, 2012.
    @inproceedings{LindstenS:2012,
      author = {Lindsten, Fredrik and Sch\"{o}n, Thomas B.},
      title = {On the use of backward simulation in the particle {G}ibbs sampler},
      booktitle = {Proceedings of the  37th {IEEE} International Conference on Acoustics, Speech and Signal Processing
      	({ICASSP})},
      year = {2012},
      address = {Kyoto, Japan},
      month = mar,
      pid = {C6},
      owner = {lindsten},
      timestamp = {2011.10.04}
    }
    
  57. [C5] F. Lindsten, H. Ohlsson, and L. Ljung, Clustering using sum-of-norms regularization; with application to particle filter output computation, in Proceedings of the IEEE Workshop on Statistical Signal Processing (SSP), Nice, France, 2011.
    @inproceedings{LindstenOL:2011,
      author = {Lindsten, Fredrik and Ohlsson, Henrik and Ljung, Lennart},
      title = {Clustering using sum-of-norms regularization; with application to particle filter output computation},
      booktitle = {Proceedings of the  {IEEE} Workshop on Statistical Signal Processing ({SSP})},
      year = {2011},
      address = {Nice, France},
      month = jun,
      pid = {C5},
      owner = {lindsten},
      timestamp = {2011.02.07}
    }
    
  58. [C4] F. Lindsten, T. B. Schön, and J. Olsson, An explicit variance reduction expression for the Rao-Blackwellised particle filter, in Proceedings of the 18th IFAC World Congress, Milan, Italy, 2011.
    @inproceedings{LindstenSO:2011,
      author = {Lindsten, Fredrik and Sch\"on, Thomas B. and Olsson, Jimmy},
      title = {An explicit variance reduction expression for the {R}ao-{B}lackwellised particle filter},
      booktitle = {Proceedings of the  18th {IFAC} World Congress},
      year = {2011},
      address = {Milan, Italy},
      month = aug,
      pid = {C4},
      owner = {lindsten},
      timestamp = {2011.03.17}
    }
    
  59. [C3] F. Lindsten and T. B. Schön, Identification of Mixed Linear/Nonlinear State-Space Models, in Proceedings of the 49th IEEE Conference on Decision and Control (CDC), Atlanta, USA, 2010.
    @inproceedings{LindstenS:2010a,
      author = {Lindsten, Fredrik and Sch\"{o}n, Thomas B.},
      title = {Identification of Mixed Linear/Nonlinear State-Space Models},
      booktitle = {Proceedings of the  49th {IEEE} Conference on Decision and Control ({CDC})},
      year = {2010},
      address = {Atlanta, USA},
      month = dec,
      pid = {C3},
      owner = {lindsten},
      timestamp = {2010.07.06}
    }
    
  60. [C2] F. Lindsten, J. Callmer, H. Ohlsson, D. Törnqvist, T. B. Schön, and F. Gustafsson, Geo-referencing for UAV Navigation using Environmental Classification, in Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Anchorage, USA, 2010.
    @inproceedings{LindstenCOTSG:2010,
      author = {Lindsten, Fredrik and Callmer, Jonas and Ohlsson, Henrik and T{\"{o}}rnqvist, David and Sch{\"{o}}n, Thomas B. and Gustafsson, Fredrik},
      title = {Geo-referencing for {UAV} Navigation using Environmental Classification},
      booktitle = {Proceedings of the  {IEEE} International Conference on Robotics and Automation ({ICRA})},
      year = {2010},
      address = {Anchorage, USA},
      month = may,
      pid = {C2}
    }
    
  61. [C1] F. Lindsten, P.-J. Nordlund, and F. Gustafsson, Conflict Detection Metrics for Aircraft Sense and Avoid Systems, in Proceedings of the 7th IFAC Symposium on Fault Detection, Supervision and Safety of Technical Processes (SafeProcess), Barcelona, Spain, 2009.
    @inproceedings{LindstenNG:2009,
      author = {Lindsten, Fredrik and Nordlund, Per-Johan and Gustafsson, Fredrik},
      title = {Conflict Detection Metrics for Aircraft Sense and Avoid Systems},
      booktitle = {Proceedings of the 7th {IFAC} Symposium on Fault Detection, Supervision and Safety of Technical
      	Processes (SafeProcess)},
      year = {2009},
      address = {Barcelona, Spain},
      month = jul,
      pid = {C1}
    }
    

Theses

  1. [PhD] F. Lindsten, Particle Filters and Markov Chains for Learning of Dynamical Systems, Linköping Studies in Science and Technology. Dissertations, No. 1530, Department of Electrical Engineering, Linköping University, SE-581 83 Linköping, Sweden, 2013.
    @phdthesis{Lindsten:2013,
      author = {Lindsten, Fredrik},
      title = {Particle Filters and Markov Chains for Learning of Dynamical Systems},
      school = {Department of Electrical Engineering, Link\"oping University},
      year = {2013},
      type = {Linköping Studies in Science and Technology. Dissertations, No. 1530},
      address = {SE-581 83 Link{\"o}ping, Sweden},
      month = oct,
      pid = {PhD}
    }
    
  2. [Lic] F. Lindsten, Rao-Blackwellised particle methods for inference and identification, Licentiate Thesis no. 1480, Department of Electrical Engineering, Linköping University, SE-581 83 Linköping, Sweden, 2011.
    @phdthesis{Lindsten:2011,
      author = {Lindsten, Fredrik},
      title = {Rao-{B}lackwellised particle methods for inference and identification},
      school = {Department of Electrical Engineering, Link\"oping University},
      year = {2011},
      type = {Licentiate Thesis no. 1480},
      address = {SE-581 83 Link{\"o}ping, Sweden},
      month = jun,
      pid = {Lic}
    }