Introduction to Bayesian inference and computation for social .ppt
《Introduction to Bayesian inference and computation for social .ppt》由会员分享,可在线阅读,更多相关《Introduction to Bayesian inference and computation for social .ppt(53页珍藏版)》请在麦多课文档分享上搜索。
1、Introduction to Bayesian inference and computation for social science data analysisNicky Best Imperial College, Londonwww.bias-project.org.uk,Outline,Overview of Bayesian methods Illustration of conjugate Bayesian inference MCMC methods Examples illustrating: Analysis using informative priors Hierar
2、chical priors, meta-analysis and evidence synthesis Adjusting for data quality Model uncertainty Discussion,Overview of Bayesian inference and computation,Overview of Bayesian methods,Bayesian methods have been widely applied in many areas medicine / epidemiology / genetics ecology / environmental s
3、ciences finance archaeology political and social sciences, Motivations for adopting Bayesian approach vary natural and coherent way of thinking about science and learning pragmatic choice that is suitable for the problem in hand,Overview of Bayesian methods,Medical context: FDA draft guidance www.fd
4、a.gov/cdrh/meetings/072706-bayesian.html: “Bayesian statisticsprovides a coherent method for learning from evidence as it accumulates” Evidence can accumulate in various ways: Sequentially Measurement of many similar units (individuals, centres, sub-groups, areas, periods) Measurement of different a
5、spects of a problem Evidence can take different forms: Data Expert judgement,Overview of Bayesian methods,Bayesian approach also provides formal framework for propagating uncertainty Well suited to building complex models by linking together multiple sub-models Can obtain estimates and uncertainty i
6、ntervals for any parameter, function of parameters or predictive quantity of interest Bayesian inference doesnt rely on asymptotics or analytic approximations Arbitrarily wide range of models can be handled using same inferential framework Focus on specifying realistic models, not on choosing analyt
7、ically tractable approximation,Bayesian inference,Distinguish between x : known quantities (data) q : unknown quantities (e.g. regression coefficients, future outcomes, missing observations) Fundamental idea: use probability distributions to represent uncertainty about unknowns Likelihood model for
8、the data: p( x | q ) Prior distribution representing current uncertainty about unknowns: p(q ) Applying Bayes theorem gives posterior distribution,Conjugate Bayesian inference,Example: election poll (from Franklin, 2004*) Imagine an election campaign where (for simplicity) we have just a Government/
9、Opposition vote choice. We enter the campaign with a prior distribution for the proportion supporting Government. This is p(q ) As the campaign begins, we get polling data. How should we change our estimate of Governments support?,*Adapted from Charles Franklins Essex Summer School course slides: ht
10、tp:/www.polisci.wisc.edu/users/franklin/Content/Essex/Lecs/BayesLec01p6up.pdf,Conjugate Bayesian inference,Data and likelihood Each poll consists of n voters, x of whom say they will vote for Government and n - x will vote for the opposition. If we assume we have no information to distinguish voters
11、 in their probability of supporting government then we have a binomial distribution for x,This binomial distribution is the likelihood p(x | q ),Conjugate Bayesian inference,Prior We need to specify a prior that expresses our uncertainty about the election (before it begins) conforms to the nature o
12、f the q parameter, i.e. is continuous but bounded between 0 and 1 A convenient choice is the Beta distribution,Conjugate Bayesian inference,Beta(a,b) distribution can take a variety of shapes depending on its two parameters a and b,Mean of Beta(a, b) distribution = a/(a+b)Variance of Beta(a,b) distr
13、ibution = ab(a+b+1)/(a+b)2,Conjugate Bayesian inference,Posterior Combining a beta prior with the binomial likelihood gives a posterior distribution,When prior and posterior come from same family, the prior is said to be conjugate to the likelihood Occurs when prior and likelihood have the same kern
14、el,Conjugate Bayesian inference,Suppose I believe that Government only has the support of half the population, and I think that estimate has a standard deviation of about 0.07 This is approximately a Beta(50, 50) distribution We observe a poll with 200 respondents, 120 of whom (60%) say they will vo
15、te for Government This produces a posterior which is a Beta(120+50, 80+50) = Beta(170, 130) distribution,Conjugate Bayesian inference,Prior mean, E(q ) = 50/100 = 0.5 Posterior mean, E(q | x, n) = 170/300 = 0.57 Posterior SD, Var(q | x, n) = 0.029 Frequentist estimate is based only on the data:,Conj
16、ugate Bayesian inference,A harder problem What is the probability that Government wins? It is not .57 or .60. Those are expected votes but not the probability of winning. How to answer this? Frequentists have a hard time with this one. They can obtain a p-value for testing H0: q 0.5, but this isnt t
17、he same as the probability that Government wins (its actually the probability of observing data more extreme than 120 out of 200 if H0 is true),Easy from Bayesian perspective calculate Pr(q 0.5 | x, n), the posterior probability that q 0.5,Bayesian computation,All Bayesian inference is based on the
18、posterior distribution Summarising posterior distributions involves integration,Except for conjugate models, integrals are usually analytically intractable Use Monte Carlo (simulation) integration (MCMC),Bayesian computation,Suppose we didnt know how to analytically integrate the Beta(170, 130) post
19、erior .but we do know how to simulate from a Beta,Bayesian computation,Can also use samples to estimate posterior tail area probabilities, percentiles, variances etc. Difficult to generate independent samples when posterior is complex and high dimensional Instead, generate dependent samples from a M
20、arkov chain having p(q | x ) as its stationary distribution Markov chain Monte Carlo (MCMC),Illustrative Examples,Borrowing strength,Bayesian learning borrowing “strength” (precision) from other sources of information Informative prior is one such source “todays posterior is tomorrows prior” relevan
21、ce of prior information to current study must be justified,Informative priors,Example 1: Western and Jackman (1994)* Example of regression analysis in comparative research What explains cross-national variation in union density? Union density is defined as the percentage of the work force who belong
22、s to a labour union Two issues Philosophical: data represent all available observations from a population conventional (frequentist) analysis based on long-run behaviour of repeatable data mechanism not appropriate Practical: small, collinear dataset yields imprecise estimates of regression effects,
23、* Slides adapted from Jeff Grynaviski: http:/home.uchicago.edu/grynav/bayes/abs03.htm,Informative priors,Competing theories Wallerstein: union density depends on the size of the civilian labour force (LabF) Stephens: union density depends on industrial concentration (IndC) Note: These two predictors
24、 correlate at -0.92. Control variable: presence of a left-wing government (LeftG) Sample: n = 20 countries with a continuous history of democracy since World War II Fit linear regression model to compare theoriesunion densityi N(mi, s2)mi = b0 + b1LeftG + b2LabF + b3IndC,Informative priors,Results w
- 1.请仔细阅读文档,确保文档完整性,对于不预览、不比对内容而直接下载带来的问题本站不予受理。
- 2.下载的文档,不会出现我们的网址水印。
- 3、该文档所得收入(下载+内容+预览)归上传者、原创作者;如果您是本文档原作者,请点此认领!既往收益都归您。
下载文档到电脑,查找使用更方便
2000 积分 0人已下载
下载 | 加入VIP,交流精品资源 |
- 配套讲稿:
如PPT文件的首页显示word图标,表示该PPT已包含配套word讲稿。双击word图标可打开word文档。
- 特殊限制:
部分文档作品中含有的国旗、国徽等图片,仅作为作品整体效果示例展示,禁止商用。设计者仅对作品中独创性部分享有著作权。
- 关 键 词:
- INTRODUCTIONTOBAYESIANINFERENCEANDCOMPUTATIONFORSOCIALPPT

链接地址:http://www.mydoc123.com/p-376631.html