This document discusses Bayesian variable selection methods for regression models. It begins by reviewing traditional ANOVA tables and their limitations for modern applications with many variables, such as GWAS studies. It then introduces Bayesian approaches using priors to perform variable selection by building it into the regression model. Several variable selection methods are described that use different prior distributions, such as slab and spike priors, the stochastic search variable selection (SSVS) method, and the normal-exponential-gamma (NEG) distribution. The document discusses how these methods can be implemented using MCMC sampling and compares their performance. It also discusses some extensions like using random effects and polynomial terms.
Time Series Foundation Models - current state and future directions
Bayesian Variable Selection Using Priors
1. Bayesian Variable Selection and the
(Ab)use of Priors
Bob O'Hara
BiK-F
Frankfurt am Main
Germany
blogs.nature.com/boboh/2012/07/16/abusing_a_prior
(this is mainly a review of work by other people)
3. Not Useful for Modern Applications
GWAS: 105 variables
Ikram MK et al (2010) Four Novel Loci (19q13, 6q24, 12q24, and 5q14) Influence the Microcirculation In Vivo. PLoS Genet. 2010 Oct 28;6(10):e1001184.
4. Anyway, we want to be Bayesian
Could use DIC, but same problems
So, let's build variable
selection into the model
22. Indicators
Ik – indicator that variable k is in the model
P(Ik=1) = p
θ ~ N(0,σβ2)
P(β) = (1-I) 0 + I θ
And integrate over P(I=1) by MCMC
Gibbs sampling should work nicely
23. A problem with Gibbs Sampling
P(β) = (1-I) 0 + I θ
When I = 0, θ only depends on its prior
So MCMC draws wide values of θ
Only rarely will it draw
“sensible” values
24. A Better Version: GVS
θ ~ N(0,σβ2(I)) Pseudo-prior
P(β) = I θ
Now if I=0, generate from a pseudo-
prior, tuned to propose sensible
values
i.e. select σβ2(0) to cover
likely values of the posterior
31. Normal Exponential Gamma
Integrate µ from Lasso over a Gamma
β ~ N(0, σβ2)
σβ2 ~ Exp(µ)
µ ~ Γ(λ,γ 2)
NEG Exponential
32. NEG & Lasso
GWAS too big for MCMC
Use quicker algorithms & only estimate
posterior modes
33. How do they compare?
Want good
separation
Good
Bad
34. Comparison
Laplace – awful. Shrinks
everything
GVS – works well (when
tuned), but slower
SSVS – works well
Jeffrey's – works very well
35. Fixed and Random Effects
Rather than fixing parameters, can
treat as a random effect to tune it
e.g. SSVS
β ~ N(0, σI2)
σ12 ~ Γ(), σ02 = c σ12 (c<<1)
Useful with many variables, can learn about scale of
response
36. Random Effects
Useful with many variables, can learn about scale of
response
Variables not in model get P(I=1|data) = P(I=1)
Random Effect
59. Response: respiration
Predictors:
Climate (PCA)
Slope, latitude, longitude etc.
Number of shrub species
Most probably have some effect
(or are correlated with something that
does)
61. If we want to do variable selection...
Should first think about priors
If our subjective prior doesn't
shrink properly, either don't
select variables, or admit to
yourself you're abusing your
priors
62. Thank you for not abusing me
blogs.nature.com/boboh/2012/07/16/abusing_a_prior