Research reports
Years: 2024 2023 2022 2021 2020 2019 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2004 2003 2002 2001 2000 1999 1998 1997 1996 1995 1994 1993 1992 1991
Scaling Limits in Computational Bayesian Inversion
by C. Schillings and Ch. Schwab
(Report number 2014-26)
Abstract
Computational Bayesian inversion of operator equations with distributed uncertain input parameters is based on an infinite-dimensional version of Bayes' formula established in [31] and its numerical realization in [27,28]. Based on the sparsity of the posterior density shown in [29],
dimension-adaptive Smolyak quadratures afford higher convergence rates than MCMC in terms of the number \(M\) of solutions of the forward (parametric operator) equation [27,28]
The error bounds and convergence rates obtained in [27,28] are independent of the parameter dimension (in particular free from the curse of dimensionality) but depend on the (co)variance
\(\Gamma > 0\) of the additive, Gaussian observation noise as \(\exp(b \Gamma^{-1})\) for some constant \(b>0\). It is proved that the Bayesian estimates admit asymptotic expansions as \(\Gamma \downarrow 0\). Sufficient (nondegeneracy) conditions for the existence of finite limits as \(\Gamma \downarrow 0\) are presented. For Gaussian priors, these limits are related to MAP estimators obtained from Tikhonov regularized least-squares functionals. Non-intrusive identification of concentration points and curvature information of the posterior density at these points by Quasi-Newton (QN) minimization of the Bayesian
potential with SR1 updates from [7,14] is proposed. Two Bayesian estimation algorithms with robust in \(\Gamma\) performance are developed: first, dimension-adaptive Smolyak
quadrature from [27,28] combined with a novel, curvature-based reparametrization of the
parametric Bayesian posterior density near the (assumed unique) global maximum of the posterior density and, second, extrapolation to the limit of vanishing observation noise variance. For either approach, we prove convergence with rates independent of the number of parameters as well as of the observation noise variance {\(\Gamma\)}. The generalized Richardson extrapolation to the limit \(\Gamma \downarrow 0\) due to A. Sidi [30] is justified by establishing asymptotic expansions wr. to \(\Gamma\downarrow 0\) of the Bayesian estimates. Numerical experiments are presented which indicate a performance independent of \(\Gamma\) on the curvature-rescaled, adaptive Smolyak algorithm.
Keywords: Bayesian inverse problems, Parametric operator equations, Smolyak quadrature, Sparsity, Non-Gaussian prior, Quasi-Newton methods, SR1 update, Posterior reparametrization, Richardson extrapolation
BibTeX@Techreport{SS14_576, author = {C. Schillings and Ch. Schwab}, title = {Scaling Limits in Computational Bayesian Inversion}, institution = {Seminar for Applied Mathematics, ETH Z{\"u}rich}, number = {2014-26}, address = {Switzerland}, url = {https://www.sam.math.ethz.ch/sam_reports/reports_final/reports2014/2014-26.pdf }, year = {2014} }
Disclaimer
© Copyright for documents on this server remains with the authors.
Copies of these documents made by electronic or mechanical means including
information storage and retrieval systems, may only be employed for
personal use. The administrators respectfully request that authors
inform them when any paper is published to avoid copyright infringement.
Note that unauthorised copying of copyright material is illegal and may
lead to prosecution. Neither the administrators nor the Seminar for
Applied Mathematics (SAM) accept any liability in this respect.
The most recent version of a SAM report may differ in formatting and style
from published journal version. Do reference the published version if
possible (see SAM
Publications).