版权说明:本文档由用户提供并上传,收益归属内容提供方,若内容存在侵权,请进行举报或认领
文档简介
1、Shadowing Properties of Optimization Algorithms Antonio Orvieto Department of Computer Science ETH Zurich, Switzerland Aurelien Lucchi Department of Computer Science ETH Zurich, Switzerland Abstract Ordinary differential equation (ODE) models of gradient-based optimization meth- ods can provide insi
2、ghts into the dynamics of learning and inspire the design of new algorithms. Unfortunately, this thought-provoking perspective is weakened by the fact that in the worst case the error between the algorithm steps and its ODE approximation grows exponentially with the number of iterations. In an attem
3、pt to encourage the use of continuous-time methods in optimization, we show that, if some additional regularity on the objective is assumed, the ODE representations of Gradient Descent and Heavy-ball do not suffer from the aforementioned problem, once we allow for a small perturbation on the algorit
4、hm initial condition. In the dynamical systems literature, this phenomenon is called shadowing. Our analysis relies on the concept of hyperbolicity, as well as on tools from numerical analysis. 1Introduction We consider the problem of minimizing a smooth functionf : Rd! R. This is commonly solved us
5、ing gradient-based algorithms due to their simplicity and provable convergence guarantees. Two of these approaches that prevail in machine learning are: 1) Gradient Descent (GD), that computes a sequence (xk)1 k=0of approximations to the minimizer recursively: xk+1= xk? rf(xk),(GD) given an initial
6、pointx0and a learning rate 0; and 2) Heavy-ball (HB)43 that computes a different sequence of approximations (zk)1 k=0such that zk+1= zk+ ?(zk? zk?1) ? rf(zk),(HB) given an initial pointz0= z?1and a momentum? 2 0,1). A method related to HB is Nesterovs accelerated gradient (NAG), for which rf is eval
7、uated at a different point (see Eq. 2.2.22 in 38). Analyzing the convergence properties of these algorithms can be complex, especially for NAG whose convergence proof relies on algebraic tricks that reveal little detail about the acceleration phenomenon, i.e. the celebrated optimality of NAG in conv
8、ex smooth optimization. Instead, an alternative approach is to view these methods as numerical integrators of some ordinary differential equations (ODEs). For instance, GD performs the explicit Euler method on y = ?rf(y)and HB the semi-implicit Euler method on q + q + rf(q) = 043,47. This connection
9、 goes beyond the study of the dynamics of learning algorithms, and has recently been used to get a useful and thought-provoking viewpoint on the computations performed by residual neural networks 10, 12. In optimization, the relation between discrete algorithms and ordinary differential equations is
10、 not new at all: the fi rst contribution in this direction dates back to, at least, 1958 18. This connection has recently been revived by the work of 49,28, where the authors show that the continuous-time limit of NAG is a second order differential equation with vanishing viscosity. This approach pr
11、ovides an interesting perspective on the somewhat mysterious acceleration phenomenon, connecting it to Correspondence to orvietoaethz.ch 33rd Conference on Neural Information Processing Systems (NeurIPS 2019), Vancouver, Canada. the theory of damped nonlinear oscillators and to Bessel functions. Bas
12、ed on the prior work of 52,27 and follow-up works, it has become clear that the analysis of ODE models can provide simple intuitive proofs of (known) convergence rates and can also lead to the design of new discrete algorithms 55,4,53,54. Hence, one area of particular interest has been to study the
13、relation between continuous-time models and their discrete analogs, specifi cally understanding the error resulting from the discretization of an ODE. The conceptual challenge is that, in the worst case, the approximation error of any numerical integrator grows exponentially2as a function of the int
14、egration interval 11,21; therefore, convergence rates derived for ODEs can not be straightforwardly translated to the corresponding algorithms. In particular, obtaining convergence guarantees for the discrete case requires analyzing a discrete-time Lyapunov function, which often cannot be easily rec
15、overed from the one used for the continuous-time analysis 46,47. Alternatively, (sophisticated) numerical arguments can be used to get a rate through an approximation of such Lyapunov function 55. In this work, we follow a different approach and directly study conditions under which the fl ow of an
16、ODE model is shadowed by (i.e. is uniformly close3to) the iterates of an optimization algorithm. The key difference with previous work, which makes our analysis possible, is that we allow the algorithm i.e. the shadow to start from a slightly perturbed initial condition compared to the ODE (see Fig.
17、 2 for an illustration of this point). We rely on tools from numerical analysis 21 as well as concepts from dynamical systems 7, where solutions to ODEs and iterations of algorithm are viewed as the same object, namely maps in a topological space 7 . Specifi cally, our analysis builds on the theory
18、of hyperbolic sets, which grew out of the works of Anosov 1 and Smale 48 in the 1960s and plays a fundamental role in several branches of the area of dynamical systems but has not yet been seen to have a relationship with optimization for machine learning. In this work we pioneer the use of the theo
19、ry of shadowing in optimization. In particular, we show that, if the objective is strongly-convex or if we are close to an hyperbolic saddle point, GD and HB are a shadow of (i.e. follow closely) the corresponding ODE models. We back-up and complement our theory with experiments on machine learning
20、problems. To the best of our knowledge, our work is the fi rst to focus on a (Lyapunov function independent) systematicandquantitativecomparisonofODEsandalgorithmsforoptimization. Also, webelievethe tools we describe in this work can be used to advance our understanding of related machine learning p
21、roblems, perhaps to better characterize the attractors of neural ordinary differential equations 10. 2Background This section provides a comprehensive overview of some fundamental concepts in the theory of dynamical systems, which we will use heavily in the rest of the paper. 2.1 Differential equati
22、ons and fl ows Consider the autonomous differential equation y = g(y). Everyyrepresents a point inRn(a.k.a phase space) andg : Rn! Rn is a vector fi eld which, at any point, prescribes the velocity of the solutionythat passes through that point. Formally, the curvey : R ! Rnis a solution passing thr
23、oughy0at time0if y(t) = g(y(t)fort 2 Randy(0) = y0. We call this the solution to the initial value problem (IVP) associated withg(starting aty0). The following results can be found in 42,26. Theorem 1. Assume g is Lipschitz continuous and Ck. The IVP has a unique Ck+1solution in R. 00.20.40.60.811.2
24、1.41.61.82 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 y0 g t1+t2(y0) g t1(y0) This fundamental theorem tells us that, if we integrate forttime units from positiony0 , the fi nal positiony(t)is uniquely determined. Therefore, we can defi ne the familyg tt2Rof maps the fl owofg such thatg t(y0)is the solution at t
25、imetof the IVP. Intuitively, we can think ofg t(y0)as determining the location of a particle starting at y0 and moving via the velocity fi eldgfortseconds. Since the vector fi eld is static, we can move along the solution (in both directions) by iteratively applying this map (or its inverse). This i
26、s formalized below. 2The error between the numerical approximation and the actual trajectory with the same initial conditions is, for a p-th order method, eCthpat time t with C ? 0. 3 Formally defi ned in Sec. 2. 2 Proposition 1.Assumegis Lipschitz continuous andCk. For anyt 2 R,g t 2 Ck+1and, for a
27、ny t1,t22 R, g t1+t2 = g t1? g t2. In particular, g t is a diffeomorphism4with inverse g ?t. In a way, the fl ow allows us to exactly discretize the trajectory of an ODE. Indeed, let us fi x a stepsize h 0; the associatedtime-h map g hintegrates the ODE y = g(y)forhseconds starting from some y0, and
28、 we can apply this map recursively to compute a sampled ODE solution. In this paper, we study how fl ows can approximate the iterations of some optimization algorithm using the language of dynamical systems and the concept of shadowing. 2.2Dynamical systems and shadowing A dynamical system onRnis a
29、map : Rn! Rn. Forp 2 N , we defi ne p= ? (ptimes). If is invertible, then ?p= ?1? ? ?1(ptimes). Since p+m= p? m, these iterates form a group if is invertible, and a semigroup otherwise. We proceed with more defi nitions. Defi nition 1. A sequence (xk)1 k=0is a (positive) orbit of if, for all k 2 N,
30、xk+1 = (xk). For the rest of this subsection, the reader may think of as an optimization algorithm (such as GD, which mapsxtox ? rf(x) and of the orbit(xk)1 k=0 as its iterates.Also, the reader may think of(yk)1 k=0 as the sequence of points derived from the iterative application ofg h, which is its
31、elf a dynamical system, from some y0. The latter sequence represents our ODE approximation of the algorithm .Our goal in this subsection is to understand when a sequence(yk)1 k=0 is close to an orbit of . The fi rst notion of such similarity is local. x0 x1 x2 x3 x4 x5 y0 y1 y2 y3 y4 y5 Defi nition
32、2.The sequence(yk)1 k=0is a?pseudo-orbitof if, for all k 2 N, kyk+1? (yk)k ?. If(yk)1 k=0is locally similar to an orbit of (i.e. it is a pseudo- orbit of ), then we may hope that such similarity extends globally. This is captured by the concept of shadowing. Defi nition 3.A pseudo-orbit(yk)1 k=0of i
33、s?shadowedif there exists an orbit(xk)1 k=0 of such that, for allk 2 N, kxk? ykk . It is crucial to notice that, as depicted in the fi gure above, we allow the shadowing orbit (a.k.a. the shadow) to start at a perturbed pointx06= y0. A natural question is the following: which properties must have su
34、ch that a general pseudo-orbit is shadowed? A lot of research has been carried out in the last decades on this topic (see e.g. 41, 31 for a comprehensive survey). Shadowing for contractive/expanding maps. A straight-forward suffi cient condition is related to contraction. is said to be uniformly con
35、tracting if there exists 0there exists? 0such that every?pseudo-orbit(yk)1 k=0of is?shadowed by the orbit (xk)1 k=0of starting at x0 = y0, that is xk:= k(x0). Moreover, ? (1 ? ). The idea behind this result is simple: since all distances are contracted, errors that are made along the pseudo-orbit va
36、nish asymptotically. For instructive purposes, we report the full proof. Proof: We proceed by induction: the proposition is trivially true atk = 0, sincekx0? y0k ; next, we assume the proposition holds atk 2 Nand we show validity fork + 1. We have xk yk+1 yk (xk) () yk ? kxk+1? yk+1k subadditivity k
37、 (xk) ? (yk)k + k (yk) ? yk+1k ?-pseudo-orbit k (xk) ? (yk)k + ? contraction kxk? ykk + ? induction + ?. 4A C1 map with well-defi ned and C1inverse. 3 Finally, since ? (1 ? ), + ? = . Next, assume is invertible. If isuniformly expanding(i.e. 1), then ?1is uniformly contracting with contraction facto
38、r1/and we can apply the same reasoning backwards: consider the pseudo-orbity0,y1, ,yKup to iterationK, and setxK= yK(before, we hadx0= y0); then, apply the same reasoning as above using the map ?1 and fi nd the initial conditionx0= ?K(yK). In App. B.2 we show that this reasoning holds in the limit,
39、i.e. that ?K(yK)converges to a suitable x0to construct a shadowing orbit under ? (1 ? 1/) (precise statement in Thm. B.3). Shadowing in hyperbolic sets.In general, for machine learning problems, the algorithm map can be a combination of the two cases above: an example is the pathological contraction
40、-expansion behavior of Gradient Descent around a saddle point 16. To illustrate the shadowing properties of such systems, we shall start by taking to be linear5, that is (x) = Axfor someA 2 Rnn. is calledlinear hyperbolicif its spectrum?(A)does not intersect the unit circleS1. If this is the case, w
41、e call?s(A)the part of?(A)insideS1and?u(A)the part of?(A)outsideS1. The spectral decomposition theorem (see e.g. Corollary 4.56 in 24) ensures thatRndecomposes into two -invariant closed subspaces (a.k.a the stable and unstable manifolds) in direct sum :Rn= Es?Eu. We call sand uthe restrictions of t
42、o these subspaces andAsandAuthe corresponding matrix representations; the theorem also ensures that?(As) = ?s(A)and?(Au) = ?u(A). Moreover, using standard results in spectral theory (see e.g. App. 4 in 24) we can equipEsandEuwith norms equivalent to the standard Euclidean normk ksuch that, w.r.t. th
43、e new norms, uis uniformly expanding and suniformly contracting. IfAis symmetric, then it is diagonalizable and the norms above can be taken to be Euclidean. To wrap up this paragraph we can think of a linear hyperbolic system as a map that allows us to decouple its stable and unstable components co
44、nsistently acrossRn. Therefore, a shadowing result directly follows from a combination of Thm. 2 and B.3 23. An important further question is whether the result above for linear maps can be generalized. From the classic theory of nonlinear systems 26,2, we know that in a neighborhood of anhyperbolic
45、 pointp for , that isD (p)has no eigenvalues onS1, behaves like6a linear system. Similarly, in the analysis of optimization algorithms, it is often used that, near a saddle point, an Hessian- smooth function can be approximated by a quadratic 35,15,19. Hence, it should not be surprising that pseudo-
46、orbits of are shadowed in a neighborhood of an hyperbolic point, if such set is -invariant 31: this happens for instance if p is a stable equilibrium point (see defi nition in 26). Starting from the reasoning above, the celebrated shadowing theorem, which has its foundations in the work of Ansosov 1
47、 and was originally proved in 6, provides the strongest known result of this line of research: near anhyperbolic setof , pseudo-orbits are shadowed. Informally, Rn is called hyperbolic if it is -invariant and has clearly marked local directions of expansion and contraction which make behave similarl
48、y to a linear hyperbolic system.We provide the precise defi nition of hyperbolic set and the statement of the shadowing theorem in App. B.1. Unfortunately, despite the ubiquity of hyperbolic sets, it is practically infeasible to establish this property analytically 3. Therefore, an important part of
49、 the literature 14,44,11,51 is concerned with the numerical verifi cation of the existence of a shadowing orbit a posteriori, i.e. given a particular pseudo-orbit. 3The Gradient Descent ODE We assume some regularity on the objective function f : Rd! R which we seek to minimize. (H1) f 2 C2(Rd,R) is
50、coercive7, bounded from below and L-smooth (8a 2 Rd, kr2f(a)k L). In this chapter, we study the well-known continuous-time model for GD : y = rf(y)(GD-ODE). Under(H1), Thm. 1 ensures that the solution to GD-ODE exists and is unique. We denote byGD h the corresponding time-hmap. We show that, under s
51、ome additional assumptions, the orbit ofGD h (which we indicate as (yk)1 k=0) is shadowed by an orbit of the GD map with learning rate h: GD h . 5This case is restrictive, yet it includes the important cases of the dynamics of GD and HB when fis a quadratic (which is the case in, for instance, least
52、 squares regression). 6For a precise description of this similarity, we invite the reader to read on topological conjugacy in 2. 7f(x) ! 1 as kxk ! 1 4 Remark 1(Bound on the ODE gradients).Under(H1)letGy= p : p = krf(y(t)k,t ? 0 be the set of gradient magnitudes experienced along the GD-ODE solution
53、 starting at anyy0. It is easy to prove, using an argument similar to Prop. 2.2 in 8, that coercivity impliessupGy0! Rdis the approximation error as a function ofh, which can be bounded as follows: kR(2,h)k h2 2 sup 0?1 k y(t + ?h)k = h2 2 sup 0?1 kr2f(y(t + ?h)rf(y(t + ?h)k h2 2 sup 0?1 kr2f(y(t +
54、?h)k sup 0?1 krf(y(t + ?h)k h2 2 L. Hence, since y(kh) ? y(kh)h = GD h (yk), we have kyk+1? GD h (yk)k L 2 h2. 3.1Shadowing under strong convexity As seen in Sec. 2.2, the last proposition provides the fi rst step towards a shadowing result. We also discussed that if, in addition, GD h is a contract
55、ion, we directly have shadowing (Thm. 1). Therefore, we start with the following assumption that will be shown to imply contraction. (H2) f is -strongly-convex: for all a 2 Rd, kr2f(a)k ? . The next result follows from standard techniques in convex optimization (see e.g. 25). Proposition 3.Assume(H1
56、),(H2). If0 h 1 L, GD h is uniformly contracting with = 1 ? h. We provide the proof in App. D.1 and sketch the idea using a quadratic form: letf(x) = hx,Hxi withHsymmetric s.t.I ? H ? LI; ifh 1 L then(1 ? Lh) kI ? hHk (1 ? h). Prop. 3 follows directly: k GD h (x) ? GD h (y)k = k(I ? hH)(x ? y)k kI ?
57、 hHkkx ? yk kx ? yk. The shadowing result for strongly-convex functions is then a simple application of Thm. 2. Theorem 3.Assume(H1),(H2)and letbe the desired accuracy. Fix0 h min2 L , 1 L; the orbit(yk)1 k=0of GD h is-shadowed by any orbit(xk)1 k=0of GD h withx0such thatkx0?y0k . Proof.From Thm. 2,
58、 we need(yk)1 k=0 to be a?-pseudo-orbit of GD h with? (1 ? ). From Prop. 2 we know? = L 2 h2, while from Prop. 3 we have (1?h). Putting it all together, we get L 2 h2 h, which holds if and only if h 2 L . Figure 1:Orbit of GD h ,GD h (sampled ODE sol.) on strongly-convex quadratic. h = 0.2. Notice that we can formulate the theorem in a dual way: namely, for every learning rate we can bound the ODE approximation error (i.e. fi nd the shadowing radius). Corollary 1.Assume(H1),(H2). If0 h 1 L, (yk)1 k=0 is-shadowed by any orbit(xk)1 k=0 of GD h starting at x0with kx0? y0k
温馨提示
- 1. 本站所有资源如无特殊说明,都需要本地电脑安装OFFICE2007和PDF阅读器。图纸软件为CAD,CAXA,PROE,UG,SolidWorks等.压缩文件请下载最新的WinRAR软件解压。
- 2. 本站的文档不包含任何第三方提供的附件图纸等,如果需要附件,请联系上传者。文件的所有权益归上传用户所有。
- 3. 本站RAR压缩包中若带图纸,网页内容里面会有图纸预览,若没有图纸预览就没有图纸。
- 4. 未经权益所有人同意不得将文件中的内容挪作商业或盈利用途。
- 5. 人人文库网仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对用户上传分享的文档内容本身不做任何修改或编辑,并不能对任何下载内容负责。
- 6. 下载文件中如有侵权或不适当内容,请与我们联系,我们立即纠正。
- 7. 本站不保证下载资源的准确性、安全性和完整性, 同时也不承担用户因使用这些下载资源对自己和他人造成任何形式的伤害或损失。
最新文档
- 2026浙江宁波市鄞州区教育系统招聘事业编制教师79人考试备考题库及答案解析
- 2026中国农业大学康绍忠院士团队“区域遥感+水资源”方向全球招聘博士后考试备考题库及答案解析
- 江苏省苏南五市联考2026年初三3月联合调研考试英语试题含解析
- 2026届安徽省淮南市潘集区重点名校下学期初三英语试题毕业班调研考试试卷含解析
- 吉林省长春市汽开区达标名校2026年初三下-第八次质量检测试题英语试题试卷含解析
- 四川省成都市青羊区2026年初三第二学期期末试卷英语试题模拟试题含解析
- 宁夏固原市泾源县市级名校2026届初三下学期第一次月考(英语试题-理)试卷含解析
- 福建省闽侯县重点中学2026届初三英语试题下学期第二次月考试题含解析
- 产品安全检验承诺书8篇
- 项目管理团队建设沟通协调预案
- DZ/T 0221-2006崩塌、滑坡、泥石流监测规范
- 科技论文写作 第2版 课件 第1-5章 科技论文写作概述-英文科技论文的写作
- 2025年春苏教版生物七年级下册教学课件 4.11.1 植物的光合作用
- T/CECS 10270-2023混凝土抑温抗裂防水剂
- 中华护理学会团体标准|2024 针刺伤预防与处理
- 2024法院书记员招聘笔试历年真题答案
- 加盟店管理制度
- 《自动化生产线安装与调试》课件-项目二 供料单元安装与调试
- 成语故事-南辕北辙-课件
- 2025届安徽省示范高中皖北协作区高三下学期3月联考(一模)历史试题(含答案)
- 人教版五年级数学下册各单元知识点总结
评论
0/150
提交评论