MCMC Model Composition Search Strategy in a Hierarchical Model

Download Presentation

MCMC Model Composition Search Strategy in a Hierarchical Model

Loading in 2 Seconds...

- 102 Views
- Uploaded on
- Presentation posted in: General

MCMC Model Composition Search Strategy in a Hierarchical Model

Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author.While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server.

- - - - - - - - - - - - - - - - - - - - - - - - - - E N D - - - - - - - - - - - - - - - - - - - - - - - - - -

MCMC Model Composition Search Strategy in a Hierarchical Model

by Susan J. Simmons

University of North Carolina Wilmington

- In many hierarchical settings, it is of interest to be able to identify important features or variables
- We propose using an MC3 search strategy to identify these important features
- We denote the response variable by yij where i=1,…,L (number of clusters or groups) and j=1,…,ni (number of observations within group i).
yij ~ N(qi, si2) and qi ~ N(Xi´b,t2)

Where X is P x L matrix of explanatory variables (P=#variables and L=# obsn)

- We set prior distributions on each of the parameters:
bj ~ N(0,100)t2 ~ Inv-c2(1)

si2 ~ Inv-c2(1)

Combining the data and prior distributions give us an implicit posterior distribution, but the full conditional posterior distributions have a nice form

.

where

- With nice parametric full conditional distributions, we can use the Gibbs sampler to get posterior samples from joint posterior distribution p(q,s2,t2,b |y).
- Using these samples, we can estimate

- The model selection vector is a binary vector that indicates which features are in the model. The vector is of length PFor example, say there are 5 variables, then one possibility is [1 0 0 1 0].
- (0) The search begins by randomly selecting the first model selection vector, and using this model, calculate P(D|M(1)).
- (1) A position along the model selection vector is randomly chosen, and then switched (0 to 1, or 1 to 0).
- (2) The probability of the data given this model is calculated (P(D|M(2)).
- (3) A transition probability aij is calculated as min (1, P(D|M(j))/ P(D|M(i))).
- (4) This transition probability is the success probability in a randomly chosen Bernoulli trial.
- (5) If the value of the Bernoulli trial is 1, the chain moves to model j and this becomes the new model selection vector. If the value is 0, the old model selection vector is maintained.
Repeat (1) – (5).

- To ensure that the chain does not get stuck in a local minimum, we generated 10 chains (each of length 2000). Total of 20,000 models investigated.
- Once all the chains have completed, we can calculate the activation probability for each feature as
- P(bj ≠0|D) = ∑P(bj ≠0|D,M(k))P(M(k)|D)
Where P(bj ≠0|D,M(k)) = 0 if feature j is not in the model and 1 if feature jis in the model, and P(M(k)|D) is calculated as

Line 1

Clones

Line 2

There are 165 different lines (or clusters) and in this simulation, ni=10 for i=1,…,165. We generated 60 different simulations scenarios.

- The 60 different simulations had effect sizes (1 through 9) and with two different error distributions. (gamma(0.5,1) and gamma(3,1)). In addition, we simulated models between 1 and 6 QTL.
- The model used to simulate the data was yij=m+∑ajxij + eij
where aj is the effect of marker j, m is the overall mean and eij is the error (gamma)

- Results so far are promising.
- Try more complex models.
- Need to make model flexible to deal with p>>n (restricted parameter search).