• Keine Ergebnisse gefunden

Surrogate models for forward and inverse uncertainty quantification

N/A
N/A
Protected

Academic year: 2021

Aktie "Surrogate models for forward and inverse uncertainty quantification"

Copied!
65
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

Research Collection

Presentation

Surrogate models for forward and inverse uncertainty quantification

Author(s):

Sudret, Bruno Publication Date:

2021-01-11 Permanent Link:

https://doi.org/10.3929/ethz-b-000469582

Rights / License:

In Copyright - Non-Commercial Use Permitted

This page was generated automatically upon download from the ETH Zurich Research Collection. For more information please consult the Terms of use.

ETH Library

(2)

Surrogate models for forward and inverse uncertainty quantification

Bruno Sudret

(3)

How to cite?

This presentation is an invited lecture given at the RWTH Aachen University (Germany) in the International Research Training Group “Modern Inverse Problems” (IRTG-2379) on January 11, 2021.

How to cite

Sudret, B.Surrogate models for forward and inverse uncertainty quantification, International Research Training Group "Modern Inverse Problems", RWTH Aachen University (Germany), January 11th, 2021.

(4)

Chair of Risk, Safety and Uncertainty quantification

The Chair carries out research projects in the field of uncertainty quantification for engineering problems with applications in structural reliability, sensitivity analysis, model

calibration and reliability-based design optimization

Research topics

• Uncertainty modelling for engineering systems

• Structural reliability analysis

• Surrogate models (polynomial chaos expansions, Kriging, support vector machines)

• Bayesian model calibration and stochastic inverse problems

• Global sensitivity analysis

(5)

Computational models in engineering

Complex engineering systems are designed and assessed using computational models, a.k.a simulators A computational model combines:

A mathematical description of the physical phenomena (governing equations),

e.g.

mechanics, electromagnetism, fluid dynamics, etc.

divσ+f=0 σ=D·ε

ε=1 2

∇u+T∇u

• Discretization techniques which transform continuous equations into linear algebra problems

• Algorithms to solve the discretized equations

(6)

Computational models in engineering

Computational models are used:

• To explore the design space (“virtual prototypes”)

• To optimize the system (e.g. minimize the mass) under performance constraints

• To assess its robustness w.r.t uncertainty and its reliability

• Together with experimental data for calibration purposes

(7)

Computational models: the abstract viewpoint

A computational model may be seen as a black box program that computes quantities of interest (QoI) (a.k.a. model responses) as a function of input parameters

Computational modelM Vector of input

parameters x∈RM

Model response y=M(x)∈RQ

• Geometry

• Material properties

Loading

• Analytical formula

• Finite element model

• Comput. workflow

• Displacements

• Strains, stresses

• Temperature, etc.

(8)

Real world is uncertain

• Differences between the designed and the real system:

Dimensions (tolerances in manufacturing)

Material properties (e.g.variability of the stiffness or resistance)

• Unforecast exposures: exceptional service loads, natural hazards (earthquakes, floods, landslides),

climate loads (hurricanes, snow storms, etc.), accidental human actions (explosions, fire, etc.)

(9)

Outline

Introduction

Uncertainty quantification: why surrogate models?

Polynomial chaos expansions PCE basis

Computing the coefficients Sparse PCE

Post-processing Bayesian inversion

Introduction

Stochastic spectral likelihood embedding

Application

(10)

Global framework for uncertainty quantification

Step A

Model(s) of the system Assessment criteria

Step B

Quantification of sources of uncertainty

Step C

Uncertainty propagation

Random variables

Computational model Moments

Probability of failure Response PDF

Step C’

Sensitivity analysis

Step C’

Sensitivity analysis

(11)

Step B: Quantification of the sources of uncertainty

Goal:

represent the uncertain parameters based on the available

data and information

Probabilistic modelfX

Experimental data is available

• What is the distribution of each parameter ?

• What is the dependence structure ? Copula theory

0 2 4 6

0 2 4 6 8 10

0 100 200 300 400

Data Normal LN Gamma

?

No data is available: expert judgment

• Engineering knowledge (e.g. reasonable bounds and uniform distributions)

• Statistical arguments and literature (e.g. extreme value distributions for climatic events)

Scarce data + expert information

0.1 0.2 0.3 0.4 0.5 0.6 0.7

PriorpΘ LikelihoodL(θ|X) Posteriorp(θ|X)

Bayesian statistics

(12)

Step C: uncertainty propagation

Goal:

estimate the uncertainty / variability of the quantities of interest (QoI) Y = M(X) due to the input uncertainty f

X

• Output statistics,

i.e.

mean, standard deviation, etc.

µ

Y

= E

X

[M(X)]

σ

Y2

= E

X

(M(X) − µ

Y

)

2

Mean/std.

deviation

µ σ

• Distribution of the QoI

Response PDF

• Probability of exceeding an admissible threshold y

adm Probability

of P

(13)

Step C’: sensitivity analysis

Goal:

determine what are the input parameters (or combinations thereof) whose uncertainty explains the variability of the quantities of interest

• Screening: detect input parameters whose uncertainty has no impact on the output variability

• Feature setting: detect input parameters which allow one to best decrease the output variability when set to a deterministic value

• Exploration: detect interactions between parameters,

i.e.

joint effects not detected when varying

parameters one-at-a-time

0.01 0.2 0.4 0.6 0.8

φD4 φC3abφL1bφL1a φC1 ∇H2φL2a φD1 AD4 K AC3ab

a

Sobol’ Indices Order 1

S(1)i

Variance decomposition (Sobol’ indices)

(14)

Uncertainty propagation using Monte Carlo simulation

Principle:

Generate virtual prototypes of the system using random numbers

• A sample set X = {x

1

, . . . , x

n

} is drawn according to the input distribution f

X

• For each sample the quantity of interest (resp. performance criterion) is evaluated, say Y = {M(x

1

), . . . , M(x

n

)}

• The set of model outputs is used for moments-, distribution- or reliability analysis

(15)

Uncertainty propagation using Monte Carlo simulation

• •

• •• •

• • • •

X1

• • •

• • • • • • •

X2

• • •

• •

• •

• • •

X3

Computational model

• • •

Y

• •• • •

• •

(16)

Advantages/Drawbacks of Monte Carlo simulation

Advantages

• Universal method: only rely upon sampling random numbers and running repeatedly the computational model

• Sound statistical foundations: convergence when n → ∞

Suited to High Performance Computing:

“embarrassingly parallel”

Drawbacks

• Statistical uncertainty: results are not exactly reproducible when a new analysis is carried out (handled by computing confidence intervals)

• Low efficiency: convergence rate ∝ n

−1/2

(17)

Surrogate models for uncertainty quantification

A surrogate model M ˜ is an approximation of the original computational model M with the following features:

• It assumes some regularity of the model M and some general functional shape

• It is built from a limited set of runs of the original model M called the experimental design X =

x

(i)

, i = 1, . . . , N

Simulated data

It is fast to evaluate!

(18)

Surrogate models for uncertainty quantification

Name Shape Parameters

Polynomial chaos expansions M(x) = ˜ X

α∈A

a

α

Ψ

α

(x) a

α

Low-rank tensor approximations M(x) = ˜

R

X

l=1

b

l M

Y

i=1

v

(i)l

(x

i

)

!

b

l

, z

k,l(i)

Kriging (a.k.a Gaussian processes) M(x) = ˜ β

T

· f (x) + Z(x, ω) β , σ

2Z

, θ

Support vector machines M(x) = ˜

m

X

i=1

a

i

K(x

i

, x) + b a , b

(Deep) Neural networks M(x) = ˜ f

n

(· · · f

2

(b

2

+ f

1

(b

1

+ w

1

· x) · w

2

)) w, b

(19)

Ingredients for building a surrogate model

Select an experimental design X that covers at best the domain of input parameters:

(Monte Carlo simulation) Latin hypercube sampling(LHS) Low-discrepancy sequences

• Run the computational model M onto X exactly as in Monte Carlo simulation

(20)

Ingredients for building a surrogate model

• Smartly post-process the data {X , M(X )} through a learning algorithm

Name Learning method

Polynomial chaos expansions sparse grid integration, least-squares, compressive sensing Low-rank tensor approximations alternate least squares

Kriging maximum likelihood, Bayesian inference

Support vector machines quadratic programming

• Validate the surrogate model,

e.g.

estimate a global error ε = E

h M(X) − M(X) ˜

2

i

(21)

Advantages of surrogate models

Usage

M(x) ≈ M(x) ˜

hours per run seconds for

10

6runs

Advantages

• Non-intrusive methods: based on runs of the computational model, exactly as in Monte Carlo simulation

• Suited to high performance computing:

“embarrassingly parallel”

Challenges

• Need for rigorous validation

• Communication: advanced mathematical background

Efficiency: 2-3 orders of magnitude less runs compared to Monte Carlo

(22)

Surrogate modelling vs. machine learning

Features Supervised learning Surrogate modelling

Computational model M

7 4

Probabilistic model of the input Xf

X

7 4

Training data: X = {(x

i

, y

i

), i = 1, . . . , n}

4 4

Training data set Experimental design

(big data) (small data)

Prediction goal: for a new x ∈ X / , y(x) ?

m

X

i=1

y

i

K(x

i

, x) + b X

α∈A

y

α

Ψ

α

(x)

Validation (resp. cross-validation)

4 4

(23)

Outline

Introduction

Uncertainty quantification: why surrogate models?

Polynomial chaos expansions PCE basis

Computing the coefficients Sparse PCE

Post-processing

Bayesian inversion

(24)

Polynomial chaos expansions in a nutshell

Ghanem & Spanos (1991; 2003); Xiu & Karniadakis (2002); Soize & Ghanem (2004)

• We assume here for simplicity that the input parameters are independent with XifXi, i= 1, . . . , d

• PCE is also applicable in the general case using an isoprobabilistic transformX7→Ξ

The polynomial chaos expansion of the (random) model response reads:

Y = X

α∈Nd

y

α

Ψ

α

(X)

where:

• Ψ

α

(X) are basis functions (multivariate orthonormal polynomials)

y are coefficients to be computed (coordinates)

(25)

Sampling (MCS) vs. spectral expansion (PCE)

Whereas MCS explores the output space /distribution point-by-point, the polynomial chaos expansion assumes a generic structure (polynomial function), which better exploits the available information (runs of the original model)

Example: load bearing capacity as a function of (c, ϕ)

Thousands (resp. millions) of points are needed to grasp the structure of the response (resp.

capture the rare events)

(26)

Visualization of the PCE construction

= “Sum of coefficients × basic surfaces”

(27)

Visualization of the PCE construction

= y

0,0

× + y

0,1

×

+ y

1,0

× + y

1,1

× + y

2,0

×

+ · · · + y

0,2

× + y

3,3

× + y

4,2

×

(28)

Polynomial chaos expansion: procedure

Y

PCE

= X

α∈A

y

α

Ψ

α

(X)

Four steps

• How to construct the polynomial basis Ψ

α

(X) for given X

i

f

Xi

?

• How to compute the coefficients y

α

?

• How to check the accuracy of the expansion ?

• How to answer the engineering questions:

Mean, standard deviation PDF, quantiles

Sensitivity indices

(29)

Multivariate polynomial basis

Univariate polynomials

• For each input variable X

i

, univariate orthogonal polynomials {P

k(i)

, k ∈ N } are built:

D

P

j(i)

, P

k(i)

E

= Z

P

j(i)

(u) P

k(i)

(u) f

Xi

(u) du = γ

j(i)

δ

jk

e.g.,Legendre polynomialsifXi∼ U(−1,1),Hermite polynomialsifXi∼ N(0,1)

• Normalization: Ψ

(i)j

= P

j(i)

/ q

γ

j(i)

i = 1, . . . , M, j ∈ N Tensor product construction

Ψ

α

(x)

def

=

M

Y

i=1

Ψ

(i)αi

(x

i

) E [Ψ

α

(X)Ψ

β

(X)] = δ

αβ

where α = (α

1

, . . . , α

M

) are multi-indices (partial degree in each dimension)

(30)

Dealing with complex input distributions

Independent variables

Input parameters with given marginal CDFs X

i

F

Xi

, i = 1, . . . , M

• Arbitrary PCE: orthogonal polynomial computed numerically on-the-fly

• Isoprobabilistic transform through a one-to-one mapping to reduced variables,

e.g.

: X

i

= F

X−1i

ξ

i

+ 1 2

if ξ

i

∼ U(−1 , 1) X

i

= F

X−1i

(Φ(ξ

i

)) if ξ

i

∼ N (0, 1)

General case: addressing dependence

Sklar’s theorem (1959)

• The joint CDF is defined through its marginals and copula

F

X

(x) = C (F

X1

(x

1

), . . . , F

XM

(x

M

))

(31)

Outline

Introduction

Uncertainty quantification: why surrogate models?

Polynomial chaos expansions

PCE basis

Computing the coefficients

Sparse PCE

Post-processing

Bayesian inversion

(32)

Computing the coefficients by least-square minimization

Isukapalli (1999); Berveiller, Sudret & Lemaire (2006)

Principle

The exact (infinite) series expansion is considered as the sum of a truncated series and a residual:

Y = M(X) = X

α∈A

y

α

Ψ

α

(X) + ε

P

Y

T

Ψ(X) + ε

P

(X)

where : Y = {y

α

, α ∈ A} ≡ {y

0

, . . . , y

P−1

} ( P unknown coefficients) Ψ(x) =

0

(x), . . . , Ψ

P−1

(x)}

Least-square minimization

The unknown coefficients are estimated by minimizing the mean square residual error:

Y ˆ = arg min E h

Y

T

Ψ(X) − M(X)

2

i

(33)

Discrete (ordinary) least-square minimization

An estimate of the mean square error (sample average) is minimized:

Y ˆ = arg min

Y∈RP

1 n

n

X

i=1

Y

T

Ψ(x

(i)

) − M(x

(i)

)

2

Procedure

• Select a truncation scheme,

e.g.

A

M,p

=

α ∈ N

M

: |α|

1

p

• Select an experimental design and evaluate the model response M =

M(x

(1)

), . . . , M(x

(n)

)

T

• Compute the experimental matrix A

ij

= Ψ

j

x

(i)

i = 1, . . . , n ; j = 0, . . . , P − 1

• Solve the resulting linear system

Y ˆ = (A

T

A)

−1

A

T

M Simple is beautiful !

(34)

Error estimators

• In least-squares analysis, the generalization error is defined as:

E

gen

= E

h M(X) − M

PC

(X)

2

i

M

PC

(X) = X

α∈A

y

α

Ψ

α

(X)

The empirical error based on the experimental design X is a poor estimator in case of overfitting

E

emp

= 1 n

n

X

i=1

M(x

(i)

) − M

PC

(x

(i)

)

2

Leave-one-out cross validation

• From statistical learning theory, model validation shall be carried out using independent data E

LOO

= 1

n

n

X

i=1

M(x

(i)

) − M

P C

(x

(i)

) 1 − h

i

2

(35)

Outline

Introduction

Uncertainty quantification: why surrogate models?

Polynomial chaos expansions

PCE basis

Computing the coefficients

Sparse PCE

Post-processing

Bayesian inversion

(36)

Curse of dimensionality

• The cardinality of the truncation scheme A

M,p

is P = (M + p)!

M ! p!

• Typical computational requirements: n = OSR · P where the oversampling rate is OSR = 2 − 3

However ... most coefficients are close to zero !

Example

• Elastic truss structure with M = 10 independent input

variables

10

−8 10−6 10−4 10−2 100

|aα/a0|

Mean p= 1 p= 2 p= 3 p >3

(37)

Hyperbolic truncation sets

Sparsity-of-effects principle

Blatman & Sudret, Prob. Eng. Mech (2010); J. Comp. Phys (2011)

In most engineering problems, only low-order interactions between the input variables are relevant

q− norm of a multi-index α :

||α||q

M

X

i=1

αqi

!

1/q

, 0< q≤1

• Hyperbolic truncation sets:

AM,pq ={α∈NM : ||α||qp}

Dim. input vectorM

2 5 10 20 50

|AM,p q|

101 103 105 107 109

p= 3 p= 3, q= 0.5 p= 5 p= 5, q= 0.5 p= 7 p= 7, q= 0.5

(38)

Compressive sensing approaches

Blatman & Sudret (2011); Doostan & Owhadi (2011); Sargsyanet al.(2014); Jakemanet al.(2015)

• Sparsity in the solution can be induced by `

1

-regularization:

y

α

= arg min 1 n

n

X

i=1

Y

T

Ψ(x

(i)

) − M(x

(i)

)

2

+ λ k y

α

k

1

• Different algorithms: LASSO, orthogonal matching pursuit, Bayesian compressive sensing

Least Angle Regression

Efronet al.(2004)

Blatman & Sudret (2011)

• Least Angle Regression (LAR) solves the LASSO problem for different values of the penalty constant in a single run without matrix inversion

• Leave-one-out cross validation error allows one to select the best model

(39)

Outline

Introduction

Uncertainty quantification: why surrogate models?

Polynomial chaos expansions

PCE basis

Computing the coefficients Sparse PCE

Post-processing

Bayesian inversion

(40)

Post-processing sparse PC expansions

Statistical moments

• Due to the orthogonality of the basis functions ( E [Ψ

α

(X)Ψ

β

(X)] = δ

αβ

) and using E [Ψ

α6=0

] = 0 the statistical moments read:

Mean: µ ˆ

Y

= y

0

Variance: σ ˆ

2Y

= X

α∈A\0

y

α2

Distribution of the QoI

• The PCE can be used as a response surface for sampling:

y

j

= X

α∈A

y

α

Ψ

α

(x

j

) j = 1, . . . , n

big

The PDF of the response is estimated by histograms or kernel

0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5

PDF

Data Kernel density

(41)

Sensitivity analysis

Goal

Sobol’ (1993); Saltelliet al.(2008)

Global sensitivity analysis aims at quantifying which input parameter(s) (or combinations thereof) influence the most the response variability (variance decomposition)

Hoeffding-Sobol’ decomposition (X ∼ U([0, 1]

M

))

M(x) = M

0

+

M

X

i=1

M

i

(x

i

) + X

1≤i<j≤M

M

ij

(x

i

, x

j

) + · · · + M

12...M

(x)

= M

0

+ X

u⊂{1, ... ,M}

M

u

(x

u

) (x

u

def

= {x

i1

, . . . , x

is

})

The summands satisfy the orthogonality condition:

Z

[0,1]M

M

u

(x

u

) M

v

(x

v

) dx = 0 ∀ u 6= v

(42)

Sobol’ indices

Total variance: D ≡ Var [M(X)] = X

u⊂{1, ... ,M}

Var [M

u

(X

u

)]

• Sobol’ indices:

S

u

def

= Var [M

u

(X

u

)]

D

• First-order Sobol’ indices:

S

i

= D

i

D = Var [M

i

(X

i

)]

D Quantify the additive effect of each input parameter separately

• Total Sobol’ indices:

S

Ti def

= X

u⊃i

S

u

Quantify the total effect of X

i

, including interactions with the other variables.

(43)

Link with PC expansions

Sobol decomposition of a PC expansion

Sudret, CSM (2006); RESS (2008)

Obtained by reordering the terms of the (truncated) PC expansion M

PC

(X)

def

= P

α∈A

y

α

Ψ

α

(X) Interaction sets

For a given u

def

= {i

1

, . . . , i

s

} : A

u

= {α ∈ A : kuα

k

6= 0}

M

PC

(x) = M

0

+ X

u⊂{1, ... ,M}

M

u

(x

u

) where M

u

(x

u

)

def

= X

α∈Au

y

α

Ψ

α

(x)

PC-based Sobol’ indices

S

u

= D

u

/D = X

α∈Au

y

α2

/ X

α∈A\0

y

2α

The Sobol’ indices are obtained analytically, at any order from the coefficients of the PC

expansion

(44)

Example: sensitivity analysis in hydrogeology

Source: http://www.futura-sciences.com/

• When assessing a nuclear waste repository, the Mean Lifetime Expectancy MLE(x) is the time required for a molecule of water at point x to get out of the boundaries of the system

• Computational models have numerous

input parameters (in each geological layer)

that are difficult to measure, and that show

scattering

(45)

Geological model

Joint work with University of Neuchâtel

Deman, Konakli, Sudret, Kerrou, Perrochet & Benabderrahmane, Reliab. Eng. Sys. Safety (2016)

• Two-dimensional idealized model of the Paris Basin (25 km long / 1,040 m depth) with 5 × 5 m mesh ( 10

6

elements)

• Steady-state flow simulation with Dirichlet boundary conditions:

∇ · (K · ∇H) = 0

• 15 homogeneous layers with uncertainties in:

Porosity (resp. hydraulic conductivity)

Anisotropy of the layer properties (inc. dispersivity) Boundary conditions (hydraulic gradients)

78 input parameters

(46)

Sensitivity analysis

10−12 10−10 10−8 10−6 10−4 10−2

T D1 D2 D3 D4 C1 C2 C3ab L1a L1b L2a L2b L2c K1K2 K3

Kbx[m/s]

Geometry of the layers Conductivity of the layers

Question

What are the parameters (out of 78) whose uncertainty drives the uncertainty of the

prediction of the mean life-time expectancy?

(47)

Sensitivity analysis: results

Technique: Sobol’indices computed from polynomial chaos expansions

0.01 0.2 0.4 0.6 0.8

φD4 φC3ab φL1b φL1a φC1 ∇H2φL2a φD1 AD4K AC3aba Total Sobol’ Indices

SToti

Parameter

P

jSj φ(resp.Kx) 0.8664

AK 0.0088

θ 0.0029

αL 0.0076

Aα 0.0000

∇H 0.0057

Conclusions

Only 200 model runs allow one to detect the 10 important parameters out of 78

• Uncertainty in the porosity/conductivity of 5 layers explain 86% of the variability

• Small interactions between parameters detected

(48)

Outline

Introduction

Uncertainty quantification: why surrogate models?

Polynomial chaos expansions

Bayesian inversion Introduction

Stochastic spectral likelihood embedding

Application

(49)

Framework

Consider a computational model M with input parameters Xπ(x) and measurements Y , the Bayesian inverse problem reads:

π(x|Y) = L(x; Y)π(x)

Z where Z =

Z

DX

L(x; Y)π(x)dx with:

• L : D

X

→ R

+

: likelihood function (measure of how well the model fits the data)

π(x|Y) : posterior density function

(50)

Markov-chain Monte Carlo

• Generally no analytical expression for π(x|Y) (exception: conjugate distributions)

MCMC based approaches to generate posterior sample:

X|Y ∼ π(x|Y)

• Quantities of interest E [h(X)|Y ]) estimated with this sample (e.g.

posterior moments).

• Large number of MCMC algorithms (e.g. Metropolis-Hastings, Hamiltonian, affine invariant ensemble sampler)

Problems

Require tuning & post-processing

No clear convergence criterion

• Does not work well with multimodal posteriors

• Overall extremely computationally expensive.

(51)

MCMC + Surrogate models

• Forward model evaluations are the expensive part in solving inverse problems

Solution:

1. Train surrogateM

c

ofM

2. Formulate likelihood functionLwithM

c

3. Use conventional MCMC algorithms

• Speed up by orders of magnitude

• With PCE surrogate, additional sensitivity analysis for free

Problems

• Globally accurate surrogate might be inaccurate in posterior domain

• Still suffers from most MCMC problems (tuning, no convergence criterion, multimodality)

Towards a sampling-free inversion approach

(52)

Spectral likelihood embedding

Basic idea: expand the likelihood function onto a PCE L(X) ≈ X

α∈A

y

α

Ψ

α

(X)

The full posterior distribution (resp. quantities of interest) can be computed analytically:

Nagel et al. (2016)

Z ˆ = E [L(X)] = y

0

π(x|Y ˆ ) = π(x) Z

X

α∈A

y

α

Ψ

α

(x)

E [h(X)|Y] = 1 Z

X

α∈Ak

y

α

a

α

with h(x) = X

α∈A

a

α

Ψ

α

(x)

Requires extremely large truncated bases A to be accurate

(53)

Outline

Introduction

Uncertainty quantification: why surrogate models?

Polynomial chaos expansions

Bayesian inversion

Introduction

Stochastic spectral likelihood embedding

Application

(54)

Stochastic spectral embedding - adaptive enrichment

L(X) ≈ X

k∈K

1

Dk

X

(X)R

kS

(X), where R

kS

(X)

def

= X

α∈Ak

y

αk

Ψ

kα

(X) ≈ R

k

.

Sequential partitioning approach

Adaptive experimental design enrichment

Algorithm:

1. initializeK={(0,1)}

2. forrefinement domaink= arg maxk∈T{Ek} do:

2.1 Partition domainDkin half:Dk{1,2}

2.2 Enrich experimental designX 2.3 ExpandRk{1,2}toRkS{1,2}

2.4 Addk{1,2}toK

(55)

Stochastic spectral likelihood embedding

After expanding the likelihood with SSLE as L(X) ≈ P

k∈K

1

Dk

X

(X)R

kS

(X) , the full posterior distribution or the following quantities of interest can be computed analytically:

Wagner et al. (2020)

Z = E [L(X)] ≈ X

k∈K

V

k

y

k0

π(x|Y ) ≈ π(x) Z

X

k∈K

1

Dk

X

(x)R

kS

(x) E [h(X)|Y] ≈ 1

Z X

k∈K

V

k

· X

α∈Ak

y

αk

a

kα

after h(x) ≈ X

α∈Ak

a

kα

Ψ

kα

(x)

(56)

Example: Heat transfer problem

• Temperature measurements at 20 locations Y = {T

1

, . . . , T

N

}

• Computational forward model solves the steady-state heat equation (FE-method):

∇(κ∇T ) = 0

Likelihood with κ

def

= (κ

1

, . . . , κ

6

) :

L(κ; Y) =

N

Y

i=1

N (T

i

|M(κ), σ

2

)

• Prior distributions:

π(κ) =

M

Y LN(µ = 30, σ = 6 W/mK)

(57)

Heat transfer problem

A reference solution is obtained by MCMC (AIES, 10

5

L evaluations)

(58)

Heat transfer problem

. . . and compared to the SSLE solution ( 10

4

L evaluations)

(59)

Posterior moments and correlations

SSE ( 10

4

L evaluations) vs.

MCMC

( 10

5

L evaluations) Moments

κ1 κ2 κ3 κ4 κ5 κ6

E[·|Y] (W/mK) 30.1(29.8) 32.5(32.3) 20.7(20.7) 32.4(32.4) 36(36.4) 26.4(26.2) Var[·|Y] (W2/mK2) 12.5(10.5) 17.4(17.9) 6.51(6.12) 27.9(26.8) 13.6(14.7) 12.8(9.02)

Correlations

(60)

Conclusions

• Surrogate models are unavoidable for solving uncertainty quantification problems involving costly computational models (e.g. finite element models)

• Depending on the analysis, specific surrogates are most suitable: polynomial chaos expansions for distribution- and sensitivity analysis, Kriging for reliability analysis

• Bayesian inverse problems can be solved with surrogate modeling (stochastic spectral embedding), without the need for MCMC simulations

• Techniques for constructing surrogates are versatile, general-purpose and field-independent

• All the presented algorithms are available in the general-purpose uncertainty quantification software

UQLab

(61)

www.uqlab.com

(62)

UQLab: The Uncertainty Quantification Software http://www.uqlab.com

• ETH license:

+ free access to academia + yearly fee for non-academic usage

• 3,285 registered users

• 1,350+ active users from 87 countries

• About 37% license renewal after one year

Country # Users

United States 529

China 440

France 321

Switzerland 255

Germany 244

United Kingdom 146

Italy 125

Brazil 116

India 107

Canada 81

As of January 11, 2021

(63)
(64)

UQWorld: the community of UQ https://uqworld.org/

(65)

Questions ?

Chair of Risk, Safety & Uncertainty Quantification www.rsuq.ethz.ch

The Uncertainty Quantification Software

www.uqlab.com

The Uncertainty Quantification Community

www.uqworld.org

Surrogate models for forward and inverse UQ RWTH Aachen – January 11, 2021 B. Sudret 50 / 50

Referenzen

ÄHNLICHE DOKUMENTE

The quantification of CVC intensity was conducted with two methods: (i) using a descriptive rating scale with four ratings, in which 0 represents plants without symptoms; 1

Starting fromknownvalue ranges for the parameters, obtained from earlier measurements in other Gelderland regions, and with additional information from a dark and light

The descrioed indicators would be prepared at the national level as per capita values for models on the basis of specific national population structures. The dietary allowances

To test the hypothesis of earlier stopping we model and solve a best choice problem first introduced by Gardner (1960) in the multiple prior framework: An ambiguity averse

Lin, Selection of polynomial chaos bases via Bayesian model uncertainty methods with applications to sparse ap- proximation of PDEs with stochastic inputs, Journal of

various characteristics of a distribution such as asymmetries or heavy tails. A further advantage in our setting is, that the characteristic function is known and, therefore,

To assess the results of the Expectation Propagation methods, we employ a Markov chain Monte Carlo (MCMC) method to approximate the first two moments, mean and variance, of

In the next simulation (c.f. Table 2), we use the analytical solution for the computations in MCS, Stochastic Collocations and Galerkin methods, thus eliminating the error caused by