版权说明:本文档由用户提供并上传,收益归属内容提供方,若内容存在侵权,请进行举报或认领
文档简介
1、Bayesian Optimization under Heavy-tailed PayoffsSayak Ray Chowdhury Department of ECE Indian Institute of Science Bangalore, India 560012 sayakiisc.ac.inAditya Gopalan Department of ECE Indian Institute of Science Bangalore, India 560012 adityaiisc.ac.inAbstractWe consider black box optimization of
2、an unknown function in the nonparametric Gaussian process setting when the noise in the observed function values can be heavy tailed. This is in contrast to existing literature that typically assumes sub- Gaussian noise distributions for queries. Under the assumption that the unknown function belong
3、s to the Reproducing Kernel Hilbert Space (RKHS) induced by a kernel, we first show that an adaptation of the well-known GP-UCB algorithm 2+2(1+)with reward truncation enjoys sublinear O Tregret even with only the(1 + )-th moments, (0, 1, of the reward distribution being bounded (Ohides logarithmic
4、factors). However, for the common squared exponential (SE) and Matrn kernels, this is seen to be significantly larger than a fundamental(T 1+ ) lower bound on regret. We resolve this gap by developing novel Bayesian optimization algorithms, based on kernel approximation techniques, with regret bound
5、s matching the lower bound in order for the SE kernel. We numerically benchmark the algorithms on environments based on both synthetic models and real-world data sets. 11IntroductionBlack-box optimization of an unknown function f : Rd Rwith expensive, noisy queries is ageneric problem arising in dom
6、ains such as hyper-parameter tuning for complex machine learningmodels 3, sensor selection 14, synthetic gene design 15, experimental design etc. The popular Bayesian optimization (BO) approach, towards solving this problem, starts with a prior distribution, typically a nonparametric Gaussian proces
7、s (GP), over a function class, uses function evaluations to compute the posterior distribution over functions, and chooses the next function evaluation adaptively using a sampling strategy towards reaching the optimum. Popular sampling strategies include expected improvement 25, probability of impro
8、vement 40, upper confidence bounds 35, Thompson sampling 11, predictive-entropy search 17, etc.The design and analysis of adaptive sampling strategies for BO typically involves the assumption of bounded, or at worst sub-Gaussian, distributions for rewards (or losses) observed by the learner, which i
9、s quite light-tailed. Yet, many real-world environments are known to exhibit heavy-tailed behavior, e.g., the distribution of delays in data networks is inherently heavy-tailed especially with highly variable or bursty traffic flow distributions that are well-modeled with heavy tails 20, heavy-taile
10、d price fluctuations are common in finance and insurance data 29, properties of complex networks often exhibit heavy tails such as degree distribution 37, etc. This motivates studying methods for Bayesian optimization when observations are significantly heavy tailed compared toGaussian.A simple vers
11、ion of black box optimization in the form of online learning in finite multi-armed bandits (MABs) with heavy-tailed payoffs, was first studied rigorously by Bubeck et al. 8, where the payoffs are assumed to have bounded (1 + )-th moment for (0, 1. They showed that for33rd Conference on Neural Inform
12、ation Processing Systems (NeurIPS 2019), Vancouver, Canada.MABs with only finite variances (i.e., = 1), by using statistical estimators that are more robust than the empirical mean, one can still recover the optimal regret rate for MAB under the sub-Gaussian assumption. Moving further, Medina and Ya
13、ng 24 consider these estimators for the problem of linear (parametric) stochastic bandits under heavy-tailed rewards and Shao et al. 34 show that almost optimal algorithms can be designed by using an optimistic, data-adaptive truncation of rewards. Some other important works include pure exploration
14、 under heavy-tailed noise 43, payoffs with bounded kurtosis 23, extreme bandits 10, heavy tailed payoffs with (0, ) 38.Against this backdrop, we consider regret minimization with heavy-tailed reward distributions in bandits with a potentially continuous arm set, and whose (unknown) expected reward f
15、unction is nonparametric and assumed to have smoothness compatible with a kernel on the arm set. Here, it is unclear if existing BO techniques relying on statistical confidence sets based on sub-Gaussian observations can be made to work to attain nontrivial regret, since it is unlikely that these co
16、nfidence sets will at all be correct. It is worth mentioning that in the finite dimensional setting, Shao et al. 34 solve the problem almost optimally, but their results do not carry over to the general nonparametric kernelized setup since their algorithms and regret bounds depend crucially on the f
17、inite feature dimension. We answer this affirmatively in this work, and formalize and solve BO under heavy tailed noise almost optimally. Specifically, this paper makes the following contributions.We adapt the GP-UCB algorithm to heavy-tailed payoffs by a truncation step, and show 2+that it enjoys a
18、 regret bound of O( T) where depends on the kernel associated2(1+)TTwith the RKHS and is generally sub-linear in T . This regret rate, however, is potentially 1sub-optimal due to a (T 1+ ) fundamental lower bound on regret that we show for two specific kernels, namely the squared exponential (SE) ke
19、rnel and the Matrn kernel.We develop a new Bayesian optimization algorithm by truncating rewards in each direction of an approximate, finite-dimensional feature space. We show that the feature approximation can be carried out by two popular kernel approximation techniques: Quadrature Fourier feature
20、s 26 and Nystrm approximation 9. The new algorithmunder eitherapproximationscheme gets regret O( T), which is optimal upto log factors for the SE kernel. 11+TFinally, we report numerical results based on experiments on synthetic as well as real-world based datasets, for which the algorithms we devel
21、op are seen to perform favorably in the harsher heavy-tailed environments.Related work. An alternative line of work uses approaches for black box optimization based on Lipschitz-type smoothness structure 22, 7, 2, 33, which is qualitatively different from RKHS smoothness type assumptions. Recently,
22、Bogunovic et al. 5 consider GP optimization under an adversarial perturbation of the query points. But, the observation noise is assumed to be Gaussian unlike our heavy-tailed environments. Kernel approximation schemes in the context of BO usually focuses on reducing the cubic cost of gram matrix in
23、version 39, 41, 26, 9. However, we crucially use these approximations to achieve optimal regret for BO under heavy tailed noise, which, we believe, might not be possible without resorting to the kernel approximations.2Problem formulationLet f : X R be a fixed but unknown function over a domain X Rd
24、for some d N. At every round, a learner queries f at a single point xt X , and observes a noisy payoff yt = f(xt) + t.Here the noise sequenhce ,t t 1 areiassumed to be zero mean i.i.d. random variables such that1+the payoffs satisfy E |yt|Ft1 v for some (0, 1 and v (0, ), where Ft1 =t1, x t1. Observ
25、e that this bound(x , y ) denotes the -algebra generated by the events so far=1on the (1 + )-th moment at best yields bounded variance for yt, and does not necessarily meanthat yt (or t) is sub-Gaussian as is assumed typically. The query point xt at round t is chosent1causally depending upon the his
26、tory (x s,sy )of query and payoff sequences available up tos=1PTround t 1. The learners goal is to maximize its (expected) cumulativerewardt=1 f(xt) over aPTt=1 (f(x?) f(xt), wheretime horizon T or equivalently minimize its cumulative regret RT =1If instead the moment bound holds for each t then thi
27、s can be translated to a moment bound for each ytusing, say, a bound on f (x).2x? argmaxxX f(x) is a maximum point of f (assuming the maximum is attained; not necessarily unique). A sublinear growth of RT with T implies the time-average regret RT /T 0 as T .Regularity assumptions: Attaining sub-line
28、ar regret is impossible in general for arbitrary reward functions f, andthussomeregularityassumptionsareneeded. In thispaper, weassumesmoothness for f induced by the structure of a kernel on X. Specifically, we make the standard assumption of ap.s.d. kernel k : X X R such that k(x, x) 1 for all x X
29、, and f being an element of thereproducing kernel Hilbert space (RKHS) Hk(X ) of smooth real valued functions on X . Moreover, the RKHS norm of f is assumed to be bounded, i.e., kf kH B for some B 0 and 0 are hyperparameters of the kernels, r = kx x0k2 is the distance between xand x0, and B is the m
30、odified Bessel function.3Warm-up: the first algorithmTowards designing a BO algorithm for heavy tailed observations, we briefly recall the stan- dard GP-UCB algorithm for the sub-Gaussian setting. GP-UCB at time t chooses the pointxt = argmaxxX t1(x) + tt1(x) where t(x) = kt(x)T (Kt + It)1Yt and 2(t
31、x) = k(x, x) kt(x)T (Kt + It)1kt(x) are the posterior mean and variance functions after t obser- vations from a function drawn from the GP prior GPX (0, k), with additive i.i.d. Gaussian noiseN(0, ). Here Yt = y1,., ytT is the vector formed by observations, Kt = k(u, v)u,vX tisthe kernel matrix comp
32、uted at the set Xt = x1, . . . , xt, kt(x) = k(x1, x), . . . , k(xt, x)T andIt is the identity matrix of order t. If the noise t is assumed conditionally R-sub-Gaussian, i.e.,p 2R2F t1 expfor all R, then usingt+1 = ORln |It + 1 K t| ensuresE et2O(T ) regret 11, as the posterior GP concentrates rapid
33、ly on the true function f . However, whenthe sub-Gaussian assumption does not hold, we cannot expect the posterior GP to have such niceconcentrationproperty. In fact, it isknownthattheridgeregressionestimator t Hk(X) of f is notrobust when the noise exhibits heavy fluctuations 19. So, in order to ta
34、ckle heavy tailed noise, oneneeds more robust estimates bt of f along with suitable confidence sets. A natural idea to curb the effects of heavy fluctuations is to truncate high rewards 8. Our first algorithm Truncated GP-UCB (Algorithm 1) is based on this idea.Truncated GP-UCB (TGP-UCB) algorithm:A
35、t each time t, we truncate the reward yt to zero ifAlgorithm 1 Truncated GP-UCB(TGP-UCB)Input: Parameters 0, b, t11and (x) = k(x, x)x Xtttit is larger than a suitably chosen truncation level bt,2Set b0(x) = 01.b0i.e., we set the truncated reward yb = yt |yfor t = 1, 2, 3 . . . dott | tThen, we const
36、ruct the truncated version of thePlay xt = argmaxxX bt1(x) + tt1 (x)posterior mean as bt (x) = kt(x)T (K t + It)1 btand observe payoff ytand simply run GP-UCB1 ywhere Yb = bTand Yb = b , . . . , btT, . . . , b Set bt = y1ttinstead of . The truncation level b canbt1t | |ttwith bCompute bt (x) = kt (x
37、)T (Kt + It)1 bttttbe adapted with time t. We choose an increasing sequence of bts, i.e., as time progresses and confi-and t2(x) = kt(x)T (Kt + It)1kt(x)end fordence interval shrinks, we truncate less and lessaggressively. Finally, in order to account for the bias introduced by truncation, we blow u
38、p the confidence width t of GP-UCB by a multiplicative factor of bt so that f(x) is contained in the interval t1(x) tt1(x) with high probability. This helps us to obtain a sub-linear regret boundfor TGP-UCB given in the Theorem 1, with a full proof deferred to appendix B. Hk (X ), kf kH B and k(x, x
39、) 1 for allTheorem 1 (Rehgret bound foriTGP-UCB) Let f1+x X . Let E |yt|Ft1 v 0, T N, (0, 1 and v 0 . Given any algohrithm, there exiists a function1+ Hk (X ) with kf k H B, and a reward distribution satisfying E|yt|Ft1 v for allft T := 1, 2, . . . , T , such that when the algorithm is run with this
40、 f and reward distribution, itsregret satisfies d 1 1 11+1+1. ERT = v 1+ln vBTT 1+if k = k ,SE d +d 2. ER T = v(1+)+d B (1+)+d T (1+)+dif k = kMatrn .The proof argument is inspired by that of Scarlett et al. 31, which provides the lower bound of BO under i.i.d. Gaussian noise, but with nontrivial ch
41、anges to account for heavy tailed observations. The proof is based on constructing a finite subset of “difficult” functions in Hk(X ). Specifically,as a uniformly sampled function from a finite set f1 , . . . , fM , where each fjwe choose fisobtained by shifting a common function g Hk(Rd) by a diffe
42、rent amount such that each of these has a unique maximum, and then cropping to X = 0, 1d. g takes values in 2, 2 with themaximum attained at x = 0. The function g is constructed properly, and the parameters , Mare chosen appropriately based on the kernel k, fixed constants B, T, , v such that any -o
43、ptimalpoint for fj fails to be -optimal point for any other fj0 and that kfj kH B for all j M . The1 vsgn (f(x), 0reward function takes values in, with the former occurring with probability212 |f (x)|, such that, for every x X , the expected reward is f (x) and (1 + )-th raw momentvis upper bounded
44、by v. Now, if we can lower bound the regret averaged over j M , then theremust exist some fj for which the bound holds. The formal proof is deferred to Appendix C.Remark 2. Theorem 2 suggests that (a) TGP-UCB may be suboptimal, and (b) for the SE kernel, itmay be possible to design algorithms recove
45、ring O( T ) regret bound under finite variances ( = 1).5An optimal algorithm under heavy tailed rewardsIn view of the gap between the regret bound for TGP-UCB and the fundamental lower bound, it is possible that TGP-UCB (Algorithm 1) does not completely mitigate the effect of heavy-tailed fluctuatio
46、ns, and perhaps that truncation in a different domain may work better. In fact, for parametric linear bandits (i.e., BO with finite dimensional linear kernels), it has been shown that appropriate truncation in feature space improves regret performance as opposed to truncating raw observations 34, an
47、d in this case the feature dimension explicitly appears in the regret bound. However, the main challenge in the more general nonparametric setting is that the feature space is infinite dimensional,4which would yield a trivial regret upper bound. If we can find an approximate feature map : X Rm in a
48、low-dimensional Euclidean inner product space Rm such that k(x, y) (x)T (y), thenwe can perform the above feature adaptive truncation effectively as well as keep the error introduced due to approximation in control. Such a kernel approximation can be done efficiently either in a data independent way
49、 (Fourier features approximation 28) or in a data dependent way (Nystrm approximation 12) and has been used in the context of BO to reduce the time complexity of GP-UCB 26, 9. But in this work, the approximations are crucial to obtain optimal theoretical guarantees. We now describe our algorithm Ada
50、ptively Truncated Approximate GP-UCB (Algorithm 2).5.1 Adaptively Truncated Approximate GP-UCB (ATA-GP-UCB) algorithmAt each round t, we select an arm xt which maximizes the approximate (under kernel approximation)GP-UCB score t1(x) + tt1(x), where t1(x) and 2 t(1x) denote approximate posteriormean and variance from the previous round, respectively and t is an appropriately chosen confidencewidth. Then, we update t(x) and t2(x) as follows. First, we find a feature embedding t R mt ,of some appropriate dimension mt, which approximates the kern
温馨提示
- 1. 本站所有资源如无特殊说明,都需要本地电脑安装OFFICE2007和PDF阅读器。图纸软件为CAD,CAXA,PROE,UG,SolidWorks等.压缩文件请下载最新的WinRAR软件解压。
- 2. 本站的文档不包含任何第三方提供的附件图纸等,如果需要附件,请联系上传者。文件的所有权益归上传用户所有。
- 3. 本站RAR压缩包中若带图纸,网页内容里面会有图纸预览,若没有图纸预览就没有图纸。
- 4. 未经权益所有人同意不得将文件中的内容挪作商业或盈利用途。
- 5. 人人文库网仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对用户上传分享的文档内容本身不做任何修改或编辑,并不能对任何下载内容负责。
- 6. 下载文件中如有侵权或不适当内容,请与我们联系,我们立即纠正。
- 7. 本站不保证下载资源的准确性、安全性和完整性, 同时也不承担用户因使用这些下载资源对自己和他人造成任何形式的伤害或损失。
最新文档
- 护理分级效果评价
- 贵州省黔东南苗族侗族自治州2026年中考模拟考试数学试卷附答案
- 2026年工业软件投资回报周期与经济效益测算方法
- 移植舱病人护理信息化管理
- 2025年前台服务礼仪冲刺卷
- 管网专项施工方案样本
- 2025年前台服务技巧冲刺卷
- 宠物美容新趋势:新媒体护理技巧分享
- 2026年智慧园区通感算控一体化建设解决方案
- 2026年广东首批数据经纪人试点:电力 金融 电商领域项目落地经验复盘
- 2026广东深圳医学科学院科研职能岗位招聘笔试备考试题及答案解析
- 山东大众报业集团有限公司招聘笔试题库2026
- 2026年国网江苏省电力有限公司高校毕业生招聘约825人(第二批)笔试模拟试题及答案解析
- 2026上半年新疆维吾尔自治区招聘事业单位工作人员分类考试4474人笔试备考题库及答案解析
- GB/T 20151-2026光度学CIE物理光度系统
- 高中实验室安全教育课件
- 2026年甘肃省交通运输厅所属事业单位招聘笔试易考易错模拟试题(共500题)试卷后附参考答案
- 电信公司客户服务部门员工绩效考评表
- 安徽合肥市人力资源服务有限公司招聘笔试题库2026
- GB/T 1883.1-2025往复式内燃机词汇第1部分:发动机设计和运行术语
- 过境公路改建工程施工组织设计
评论
0/150
提交评论