• Keine Ergebnisse gefunden

Sepsis Survival Data:

K- M Estimator & Survival Curves from Posterior Predictive

7. Models and algorithms

7.5. MCMC sampler

7.5.1. Posterior and full conditionals

For the hierarchy given for the reversible jump model, the full posterior with given hyperparameterssmean,smax,az,bz,aσ andbσ can be written as

p(B,s,l,z,τ2,β,σε2,y) = 1−

smax

i

=1

simean

i! esmean

!1

sBmean

B! esmean· (B−1)! (J−d−2)B1· bzBaz

Γ(az)B

B b=1

zbaz1exp(−bzzb

B b=1

zbexp

−zbτb2

·

Bb=1q τb2

lb

(2π)(Jd)/2 exp

1

0(d)0T(τ2,l)1(d)β

· bσaσ

Γ(aσ)σ

2(−aσ1)

ε exp

bσ σε2

1

(2πσε2)n/2 exp

−ky−Xβk2ε2

. Accordingly, the full conditionals are:

p(zb|az,bz,τb2) ∝zabzexp

−(bz+τb2)zb

⇒ zb|· ∼Γ(az+1,bz+τb2) p(τb|s,l,β,zb) ∝q

τb2

lb

exp − 1b2

sb+11 k

=sb

((d)β)2k−zbτb2

!

= (τb2)lb/2exp −1 2

sb+11 k

=sb

((d)β)2k(τb2)1+2zbτb2

!!

τb2|· ∼GIG χ=

sb+11 k

=sb

((d)β)2k;ψ=2zb;λ =1−lb 2

!

GIG(χ,ψ,λ)denotes the generalized inverse Gaussian distribution with den-sity

f(x) = (ψ/χ)λ/2 2Kλ(√

ψχ)x

λ1exp

1 2

χx1+ψx

for x > 0, where Kλ(·) is the modified Bessel function of the third kind of (fractional) order λ (Jørgensen, 1982). We use our own C-code implementing the algorithm given by Dagpunar (1989) to sample from this distribution.

p(β|τ2,l,σε2) ∝exp −ky−Xβk2ε2β

0(d)0T(τ2,l)1(d)β 2

!

⇒β|· ∼ NJ µ=σε2V X0y;Σ=V

;

V =σε2X0X+(d)0T(τ2,l)1(d)1

p(σε2|aσ,bσ,β) ∝σε2(−aσn/21)exp

−ky−Xβk2+2bσε2

σε2|· ∼ IG(aσ+n/2,bσ+ky−Xβk2/2)

7.5.2. Performance

MCMC sampling for high-dimensional hierarchical models such as the ones we consider can run into a number of difficulties such as failure to visit all the relevant modes in cases of multimodality, non-convergence, slow mix-ing and strong sensitivity to startmix-ing values. To address these concerns, we investigated the behavior of the proposed samplers for multiple runs initial-ized with highly overdispersed starting values for bz and τ2 generated from their respective diffuse priors. Sensitivity to hyperparameters is discussed in section 8.7.2.

We evaluated convergence of the runs by convergence in β (and σε2, for Gaussian responses), as the parameters (z,τ2)change in meaning due to the changing shape ofsfor FlexNEG and RJ.NEG. We found that all the samplers for Gaussian responses converge quickly (<1000 iterations) even if initialized in highly improbable regions of the posterior and mix very well. Some

quali-fications apply for the basic NEG model: it is more sensitive to starting values due to its less flexible parametrization and occasionally gets stuck in local modes with too much or too little regularization of the spline coefficients in some regions if unfortunate starting values are chosen, especially if the initial value of bz is extreme. This is not the case for FlexNEG and RJ.NEG, which are able to move away from inferior basins of attraction quickly due to the more flexible shape of the variance function.

Generally speaking, the quick convergence and excellent mixing is due to the availability of blockwise Gibbs sampling steps in the relevant levels of the hierarchy, which obviate the manual tuning of proposal densities entirely.

We achieved stable results for the NEG for starting values of τ2 = 100 and 0.1<bz <10, which is the configuration we used in the simulation study.

The performance of the P-IWLS sampler for βfor non-Gaussian responses is highly dependent on the starting values: If unsuitable starting values are provided, the sampler will get stuck in the initial configuration and fail to update, because the local approximation of the posterior used for the pro-posal density is unsuitable. In our implementation, suitable starting values forβ to initialize the chain are found by performing a number of Fisher scor-ing steps for fixed values of τ2 starting from the unregularized estimate of β. Chains initialized in this way converge quickly regardless of the starting values forbz andτ2, with satisfactory acceptance rates and good mixing due to the automatic adaptation of the proposal density to the mode and curva-ture of the full conditional. In our simulation study, acceptance rates for the IWLS proposals are between 26% and 88% and usually around 60% for Pois-son responses. Acceptance rates for binary Binomial responses are between 13% and 42%, and usually around 25%. Acceptance rates tend to decrease with increasing J since we update all elements of β simultaneously, which is necessary to achieve good mixing. Note that our implementation is therefore not well suited for very heavily parameterized models using more than 100 basis functions. This will rarely be an issue in practical applications, however.

We did encounter some numerical problems in the fitting of very challeng-ing functional forms: On the one hand, the samplchalleng-ing of variates from the generalized inverse Gaussian distribution (the full conditional density ofτ2) can occasionally fail for extreme combinations of parameter values. In this case, we simply keep the previous iterates of the respective elements of τ2. If, in the case of RJ.NEG, this is not possible because the dimension of the τ2-proposal to be drawn is different from the dimension of the current τ2, we calculate the expected values of the full conditional distributions for the problematic elements ofτ2and use those as the updated values. The software gives out a warning if sampling from the generalized inverse Gaussian fails.

In our experience, this ad-hoc fix works well in practice and the resulting samples are indistinguishable from regularly obtained samples, because, if at all, only ever a small fraction of elements inτ2 fails to update in the regular

fashion so that the convergence of the chain is not affected.

A second type of problem – which we were unable to fix – occurs (rarely) in the sampling ofβ:

As the expressions above show, generating a new β-proposal requires the inversion of the J×J matrix

σε2X0X+(d)0T(τ2,l)1(d)

. Occasionally, despite its construction ensuring positive definiteness, this matrix will be nu-merically not positive definite (or not even semi-definite). In this case the Cholesky-root based matrix inversion we employ fails and we switch to an inversion based on the singular value decomposition (SVD). When the under-lying BLAS routine for SVD fails as well, the program crashes. This caused the 12 aborted fits for RJ.NEG for m3(x) in Section 8.3. Using the numeri-cally more stable QR- or LU-decompositions for matrix inversion is unfortu-nately not effective in this case since we also require the matrix root V1/2 to generate the multivariate normal proposal vector βprop = µ+V1/2η, where η ∼ N(0, 1) . This matrix root can be obtained, however, from the Cholesky or the SVD of V1.

7.5.3. Alternative proposals for the birth and death moves

We also experimented with a more complex proposal scheme for the birth and death moves. Specifically, for the birth step we select the interval

b? ∈ {1, . . . ,B−1} \ {b : lb =1}with probability

p(b) ∝lb2Var

((d)β)sb−1,...,sb1

skb=s1b1|((d)β)k| ,

placing a higher proposal density on selecting long intervals with a large variation coefficient of the increments of the random walk. This increases the chance of splitting intervals in which both the proportion of small changes in β and the variability in the entries of ((d)β) are large. Intervals with those properties are not homogeneous and can potentially benefit from at least one additional changepoint separating the small changes, which may war-rant stronger regularization, from the larger ones responsible for the larger variation which potentially reflect jumps or curvature changes in the function to be fitted. The location of the new changepoints?b? is then drawn uniformly from

{sb? +1, . . . ,sb?+1−1}.

In the death step, we select the changepoint sb?; b? ∈ {1, . . . ,B−1} to be

removed with probability

p(b) ∝ 1 lb+lb+1

skb=s1b−1((d)β)k

lb

sb+11

k=sb ((d)β)k lb+1

.

This increases the chance of removing a changepoint sb with short adjacent intervals and small difference between the neighboring local means of(d)β.

The fitted functions based on these proposals and a uniform prior for the number of knotsBwere practically identical to fitted functions for the simpler algorithm with a truncated Poisson prior for B. We did not observe any im-provement in the sense of a more parsimonious representation of the variance function of the random walk and acceptance probabilities for the dimension changing moves were unreasonably low (0.1−0.2) in most cases.