




已阅读5页,还剩7页未读, 继续免费阅读
版权说明:本文档由用户提供并上传,收益归属内容提供方,若内容存在侵权,请进行举报或认领
文档简介
Sample Adaptive MCMC Michael H. Zhu Department of Computer Science Stanford University Stanford, CA 94305 Abstract For MCMC methods like Metropolis-Hastings, tuning the proposal distribution is important in practice for effective sampling from the target distribution. In this paper, we present Sample Adaptive MCMC (SA-MCMC), a MCMC method based on a reversible Markov chain forNthat uses an adaptive proposal distribution based on the current state ofNpoints and a sequential substitution procedure with one new likelihood evaluation per iteration and at most one updated point each iteration. The SA-MCMC proposal distribution automatically adapts within its parametric family to best approximate the target distribution, so in contrast to many existing MCMC methods, SA-MCMC does not require any tuning of the proposal distribution. Instead, SA-MCMC only requires specifying the initial state ofNpoints, which can often be chosen a priori, thereby automating the entire sampling procedure with no tuning required. Experimental results demonstrate the fast adaptation and effective sampling of SA-MCMC. 1Introduction Markov Chain Monte Carlo (MCMC) methods are a large class of sampling-based algorithms that can be applied to solve integration problems in high-dimensional spaces. The goal of MCMC methods is to sample from a probability distributionp()(known up to some normalization constant) by constructing a Markov chain that visits pointswith a frequency proportional to the corresponding probabilityp() . The Markov chain is constructed so that the Markov chain satisfi es detailed balance with respect top(), and assuming the Markov chain is also ergodic, the limiting distribution isp(). For MCMC methods like Metropolis-Hastings 1,2, the choice of the proposal distributionq(|(k) is important in practice for effective sampling from the target distribution. MH is generally used with random walk proposals where local moves based onq(|(k)are used to globally simulate the target distributionp() . A suboptimal choice for the scale or shape of the proposal can lead to ineffi cient sampling, yet the design of an optimal proposal distribution is challenging when the properties of the target distribution are unknown, especially in high-dimensional spaces. Gelman et al.3,4recommend a two-phase approach where the covariance matrix of the proposal distribution in phase 2 is proportional to the covariance of the posterior samples from phase 1. Adaptive MCMC methods such as Adaptive Metropolis 5 continually adapt the proposal distribution based on the entire history of past states. However, the method is no longer based on a valid Markov chain, so the usual MCMC convergence theorems do not apply and the validity of the sampler and an ergodic theorem must be proved for each specifi c algorithm under certain technical assumptions 6,7. In this paper, we propose Sample Adaptive MCMC, a method that is adaptive based on the current state ofNpoints and uses an adaptive proposal which is an adaptive approximation of the target distribution. 33rd Conference on Neural Information Processing Systems (NeurIPS 2019), Vancouver, Canada. 2Related work Our substitution procedure is related to the Sample Metropolis-Hastings algorithm by Liang et al.8, ch. 5and Lewandowski9. The SMH algorithm reduces to Metropolis-Hastings forN = 1, and our substitution procedure reduces to the method of Barker10. The SMH algorithm has also been used by Martino et al.11in the context of adaptive importance sampling and with independent SMH proposals by Martino et al.12who propose a family of orthogonal parallel MCMC methods where “vertical” MCMC chains are run in parallel using random-walk proposals and share information using “horizontal” MCMC steps encompassing all of the chains using independent proposals. Parallel tempering 13 runs parallel MCMC chains targeting the posterior distribution at different temperatures. Many previous works have studied MCMC methods which simulate fromN(a sample of sizeNfrom). Early works include the Adaptive Direction Sampler by Gilks et al.14, the Normal Kernel Coupler by Warnes15, and the pinball sampler by Mengersen and Robert16. Warnes15 fi rst selects one of theNpoints in the state to update, uses a kernel density estimate constructed from the state ofN points to propose a new point, and fi nally accepts or rejects the proposed swap according to the Metropolis-Hasting acceptance probability. Goodman and Weare17 propose an ensemble MCMC sampler with affi ne invariance. Griffi n and Walker18present a method for adaptation in MH by letting the joint density be the product of a proposal density andNand then sampling this augmented density using a Gibbs sampler including a Metropolis step. Their work builds on the earlier works by Cai et al.19, Keith et al.20. Leimkuhler et al.21propose an Ensemble Quasi-Newton sampler using gradient information based on time discretization of an SDE that can incorporate covariance information from the other walkers. The Multiple-Try Metropolis method 22,23 fi rst proposesMpotential candidates, randomly chooses one of the best candidates based on the weights to be the potential move, and fi nally accepts or rejects the move according to a generalized MH ratio. Neal et al.24propose a new Markov chain method for sampling from the posterior of a hidden state sequence in a non-linear dynamical system by fi rst proposing a “pool” of candidate states and then using DP with an embedded HMM. Tjelmeland25describe a general framework for running MCMC with multiple proposals in each iteration and using all proposed states to estimate mean values. Neal26propose a MCMC scheme which fi rst stochastically maps the current stateto an ensemble(1,.,N), applies a MCMC update to the ensemble, and fi nally stochastically selects a single state. Calderhead27presents a general construction for parallelizing MH algorithms. Population Monte Carlo 28 is an iterated importance sampling scheme with a state ofNpoints where the proposal distribution can be adapted for each point and at each iteration in any way and a resampling step based on the importance sampling weights is used to update the state. Adaptive Importance Sampling 29 represents a class of methods, including PMC, based on importance sampling with adaptive proposals. Our work is also inspired by particle fi lters 30,31 and PMCMC 32 which combines standard MCMC methods with a particle fi lter based inner loop for joint parameter and state estimation in state-space models. 3Sample Adaptive MCMC We now present the Sample Adaptive MCMC algorithm. Letp()be the target probability density known up to some normalization constant, and let() = p()/ R p(0)d0. The state of the SA- MCMC Markov Chain consists ofNpoints at each iteration. We denote the state at iterationkby S(k)= (k) 1 ,(k) 2 ,.,(k) N ) . Defi ne(S) = 1 N PN n=1n to be the mean of theNpoints in the stateS . Defi ne(S)to be the sample covariance matrix of theNpoints in the stateS. Optionally, we can also consider a diagonal approximation of(S)that is non-zero along the diagonals and zero elsewhere. When proposing a new point0, the proposal distributionq(|(S(k),(S(k)is a function of the mean and covariance of allNpoints in the current stateS(k). In our experiments, we use a Gaussian or Gaussian scale-mixture family as our adaptive family of proposal distributions. After proposing0, the algorithm might reject the proposed point0or substitute any of theN current points with0. For example, the algorithm might substitute(k) 1 with0so that the new state becomesS(k+1)= (0,(k) 2 ,.,(k) N ). The probabilities of substituting each of theNpoints with 2 123 S N+1 q(|(S),(S) N+123 S1 1 1 q(1|(S1),(S1)/p(1) 1N+13 S2 2 2 q(2|(S2),(S2)/p(2) 12N+1 S3 3 3 q(3|(S3),(S3)/p(3) 123 S(N+1) N+1 4 q(4|(S4),(S4)/p(4) Figure 1: Illustration of one iteration of SA-MCMC forN = 3.After the proposed point N+1 q(|(S),(S)is sampled, the setsS1,.,S(N+1)are used to calculate the sub- stitution probabilities1,.,(N+1). One of the setsS1,.,S(N+1)is chosen to be the next state with probability proportional to n. Algorithm 1 Sample Adaptive MCMC Require: p(), q0(), q(|(S),(S), N, , K 1:Initialize S(0) (1,.,N) where n q0() for n = 1,.,N 2:for k = 1 to + K do 3:Let S = (1,.,N) S(k1) 4:Sample N+1 q(|(S),(S) 5:Let Sn (S with nreplaced by N+1) for n = 1,.,N. Let S(N+1) S. 6:Let n q(n|(Sn),(Sn)/p(n) for n = 1,.,N + 1 7:Sample j J with PJ = n = n .PN+1 i=1 i,1 n N + 1 8:Let S(k) Sj 9:end for 10:Return S k=+1,.,+KS (k) the proposed point and the probability of rejecting the proposed point are all constructed so that the stationary distribution of the SA-MCMC Markov chain is N(1,.,N) = QN n=1(n). The SA-MCMC algorithm is presented in Algorithm 1 and illustrated in Figure 1. The initialization distribution for initializing theNpoints isq0(). The setsSn, withnreplaced by the proposed point, are theN + 1possibilities for the next state depending on which of the currentNpoints gets replaced, if any. One of the setsS1,.,S(N+1)is chosen to be the next state with probability proportional ton. The number of burn-in iterations is, and the number of estimation iterations isK. For any functionh()satisfying R |h()|()d 0such that ifa j band|xj j| for j 1,.,d, then ?1 q(x | ,diag(2) 0such that ifkx yk ?35,34. (A2) is a generalization for a family of proposal distributions with different means and scales (e.g. location-scale families). 4 Our next theorem is stated in a more general form for a proposal distributionq(|(S). We prove the uniform ergodicity of SA-MCMC assumingq(|)/()is bounded above and below. The proof is based on the proof of Lemma 1 in the working paper by Chan and Lai 36 and is in Appendix 3. Theorem 2.Letbe a positive target density on the parameter space, and letq(|) : be a family of positive proposal densities, witha convex Euclidean set. Let(|) = q(|)/() and let the proposal density beq(|(S), where(S) = N1 P S() for some continuous : . Letfkdenote the joint densities of(k) 1 ,.,(k) N )andNthe product density of onN. If there exist constants0 a 0such thatqk(|hk1)/() for all,hk1,k. Uniform ergodicity is proven by lower bounding the one-step probability of transitioning to the target density each iteration. TheconditionsaboveforIMHandAIMHessentiallyrequirethattheproposaldensitieshaveuniformly heavier tails than the target. We note that a mixture proposal distribution, with the main distribution and a fat-tailed distribution with a small mixing proportion, can be used as a safeguard to guarantee this lower bound 39,40,41. For our algorithm, in practice, we observe that this condition (i.e. (|) a) is also crucial for SA-MCMC. In the proof and the corresponding bound, this condition corresponds to the termaNinC. Formally, our proof also requires the assumption(|) bto lower bound the acceptance probability ofNsubstitutions by the term( a (N+1)b) N inCcorresponding to Line 7 in Algorithm 1. In practice, we fi nd that this condition is not necessary for effective sampling. A proposed point with signifi cantly larger is unlikely to be accepted in the fi rst place, and if accepted, the point is likely to be replaced quickly, so the worst-case bound of(N + 1)bin the denominator likely understates the practical performance of the algorithm. To support this conclusion, we conduct extensive experiments on t-distributions with different degrees of freedom in Appendix 4 and observe that only the assumption (|) a is necessary in practice. While IMH and AIMH have weaker assumptions for uniform ergodicity in theory, we note that IMH and AIMH fail to work for any of the examples in our paper since they are not adaptive enough for an independent proposal distribution to work in high-dimensional spaces. We elaborate in Appendix 5. ImplementationA fast, numerically stable implementation of SA-MCMC is given in Appendix 6. 4Experimental
温馨提示
- 1. 本站所有资源如无特殊说明,都需要本地电脑安装OFFICE2007和PDF阅读器。图纸软件为CAD,CAXA,PROE,UG,SolidWorks等.压缩文件请下载最新的WinRAR软件解压。
- 2. 本站的文档不包含任何第三方提供的附件图纸等,如果需要附件,请联系上传者。文件的所有权益归上传用户所有。
- 3. 本站RAR压缩包中若带图纸,网页内容里面会有图纸预览,若没有图纸预览就没有图纸。
- 4. 未经权益所有人同意不得将文件中的内容挪作商业或盈利用途。
- 5. 人人文库网仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对用户上传分享的文档内容本身不做任何修改或编辑,并不能对任何下载内容负责。
- 6. 下载文件中如有侵权或不适当内容,请与我们联系,我们立即纠正。
- 7. 本站不保证下载资源的准确性、安全性和完整性, 同时也不承担用户因使用这些下载资源对自己和他人造成任何形式的伤害或损失。
最新文档
- 1.3正方形的性质与判定 预习练(含解析)北师大版数学九年级上册
- 2025温福高铁(连江段)项目指挥部招聘2人笔试备考试题及答案解析
- 2025浙江丽水市青田县卫健系统招聘编外聘用人员28人考试备考试题及答案解析
- 2025四川九州电子科技股份有限公司招聘PQE等3人考试备考题库及答案解析
- 2025四川绵阳市华丰科技股份有限公司招聘生产管理等岗位14人笔试参考题库附答案解析
- 出租车行业急救知识培训课件
- 2025云南省红河州个旧市锡城街道卫生院招聘编外工作人员(1人)笔试备考试题及答案解析
- 2025中国建筑一局(集团)有限公司基础设施事业部海外商务管理岗招聘1人考试备考题库及答案解析
- 2025年广东阳江市阳西县中小学校选调教师91人笔试参考题库附答案解析
- 2025浙江丽水市松阳县“乡村运营师”人才引进1人笔试参考题库附答案解析
- 旋风分离器效率计算
- 温硝化制硝基苯装置的改进
- 保教知识与能力幼儿园课件
- 财务部半年度述职汇报PPT模板
- 药品种类清单
- 公共基础知识(社区工作者基础知识)试题(附答案)
- GB/T 37915-2019社区商业设施设置与功能要求
- GB/T 31298-2014TC4钛合金厚板
- 《电业安全工作规程》
- 卡西欧gw5600说明书
- 中兴NGN培训教材 MSG9000结构原理介绍课件
评论
0/150
提交评论