180 | | * |
| 180 | === Thursday PM1 === |
| 181 | * Linear regression models |
| 182 | * y = a0 + a1Xa + e (e=random variation between individuals) |
| 183 | * linear in the parameters |
| 184 | * I = covariance of outcomes. In the simple case you assume there is no covariance, so I is the identity matrix, but if there is covariance between the data or the animals or experiments then it can be incorporated with this variable. |
| 185 | * Ordinary least squares estimation |
| 186 | * SSR(Beta) sum of squared residuals |
| 187 | * Find values for Beta such that sum of squared residuals is small to increase the likelihood value |
| 188 | * R - solve(), lm() "linear model" |
| 189 | * Bayesian regression |
| 190 | * You can make probability statements |
| 191 | * probability some beta value >0 given |
| 192 | * OLS overfits when # of predictors is large |
| 193 | * can do model selection and averaging |
| 194 | * Beta is a vector |
| 195 | * If prior variance is very small, then posterior mean will concentrate around your prior mean. |
| 196 | * If you have a lot of data, as sample size grows mean will concentrate on OLS estimate. |
| 197 | * Similar logic to Bayesian inference |
| 198 | * How to select prior for a vector of parameters |
| 199 | * g-prior |
| 200 | * Uncertainty in my prior is the same as uncertainy from n/g observations. |
| 201 | * Posterior calculations are relatively similar |
| 202 | * Posterior mean estimate is OLS estimate shrunken a little bit toward zero |
| 203 | * "It's bad form not to have a picture at the end." |
| 204 | === Thursday PM2 === |
| 205 | * Changing significance level as a function of ''n'' |
| 206 | * a/(1-b) * prior odds of H0/H1 |
| 207 | * a = alpha level, significance |
| 208 | * b = power |
| 209 | * R is the ratio of costs |
| 210 | * Significance level should decrease as n increases. |
| 211 | * False discovery rate B/K |
| 212 | * Bayesian False Discoveries |
| 213 | * Posterior odds = bayes factor * prior odds < R |
| 214 | * '''Depends on the sample size, but not on the number of test''' |
| 215 | * '''Don't use bonferroni, use false discovery rate, but better still use bayesian methods.''' |
| 216 | * Bayesian approaches in GWAS (Stephens & Balding 2009) |