Bayesian tomography with prior-knowledge-based parametrization and surrogate modelling
Détails
Télécharger: 2201.02444.pdf (9186.26 [Ko])
Etat: Public
Version: de l'auteur⸱e
Licence: Non spécifiée
Etat: Public
Version: de l'auteur⸱e
Licence: Non spécifiée
ID Serval
serval:BIB_F6461418CA5D
Type
Article: article d'un périodique ou d'un magazine.
Collection
Publications
Institution
Titre
Bayesian tomography with prior-knowledge-based parametrization and surrogate modelling
Périodique
Geophysical Journal International
ISSN
0956-540X
Statut éditorial
Publié
Date de publication
2022
Volume
231
Numéro
1
Pages
673-691
Langue
anglais
Résumé
We present a Bayesian tomography framework operating with prior-knowledge-based parametrization that is accelerated by surrogate models. Standard high-fidelity forward solvers (e.g. finite-difference time-domain schemes) solve wave equations with natural spatial parametrizations based on fine discretization. Similar parametrizations, typically involving tens of thousand of variables, are usually employed to parametrize the subsurface in tomography applications. When the data do not allow to resolve details at such finely parametrized scales, it is often beneficial to instead rely on a prior-knowledge-based parametrization defined on a lower dimension domain (or manifold). Due to the increased identifiability in the reduced domain, the concomitant inversion is better constrained and generally faster. We illustrate the potential of a prior-knowledge-based approach by considering ground penetrating radar (GPR) traveltime tomography in a crosshole configuration with synthetic data. An effective parametrization of the input (i.e. the permittivity distributions determining the slowness field) and compression in the output (i.e. the traveltime gathers) spaces are achieved via data-driven principal component decomposition based on random realizations of the prior Gaussian-process model with a truncation determined by the performances of the standard solver on the full and reduced model domains. To accelerate the inversion process, we employ a high-fidelity polynomial chaos expansion (PCE) surrogate model. We investigate the impact of the size of the training set on the performance of the PCE and show that a few hundreds design data sets is sufficient to provide reliable Markov chain Monte Carlo inversion at a fraction of the cost associated with a standard approach involving a fine discretization and physics-based forward solvers. Appropriate uncertainty quantification is achieved by reintroducing the truncated higher order principle components in the original model space after inversion on the manifold and by adapting a likelihood function that accounts for the fact that the truncated higher order components are not completely located in the null space.
Site de l'éditeur
Création de la notice
30/06/2023 10:33
Dernière modification de la notice
24/07/2023 6:17