免费预览已结束,剩余1页可下载查看
下载本文档
版权说明:本文档由用户提供并上传,收益归属内容提供方,若内容存在侵权,请进行举报或认领
文档简介
Entropic Risk Measure in Policy Search David Nass Boris Belousov and Jan Peters Abstract With the increasing pace of automation mod ern robotic systems need to act in stochastic non stationary partially observable environments A range of algorithms for fi nding parameterized policies that optimize for long term average performance have been proposed in the past However the majority of the proposed approaches does not explicitly take into account the variability of the performance metric which may lead to fi nding policies that although performing well on average can perform spectacularly bad in a particular run or over a period of time To address this shortcoming we study an approach to policy optimization that explicitly takes into account higher order statistics of the reward function In this paper we extend policy gradient methods to include the entropic risk measure in the objective function and evaluate their performance in simulation experiments and on a real robot task of learning a hitting motion in robot badminton I INTRODUCTION Applying reinforcement learning RL to robotics is noto riously hard due to the curse of dimensionality 1 Robots operate in continuous state action spaces and visiting every state quickly becomes infeasible Therefore function approx imation has become essential to limit the number of param eters that need to be learned Policy search methods that employ pre structured parameterized policies to deal with continuous action spaces have been successfully applied in robotics 2 These methods include policy gradient 3 4 natural policy gradient 5 expectation maximization EM policy search 6 7 and information theoretic ap proaches 8 A common feature of the aforementioned policy search methods is that they all aim to maximize the expected reward Therefore they do not take into account the variability and uncertainty of the performance measure However robotic systems need to act in stochastic non stationary partially observable environments To account for these challenges the objective function should include an additional variance related criteria This paper contributes to the fi eld of rein forcement learning for robotics by extending the range of applicability of policy search methods to problems with risk sensitive optimization criteria where risk is given by the entropic risk measure 9 II RELATED WORK Howard and Matheson 10 along with Jacobson 11 were the fi rst to consider risk sensitivity in optimal control both The authors are with Intelligent Autonomous Systems Lab Depart ment of Computer Science Technische Universit at Darmstadt Germany surname ias tu darmstadt de This project has received funding from the European Union s Horizon 2020 research and innovation programme under grant agreement No 640554 in discrete and continuous settings Jacobson attempted to solve the linear exponential quadratic Gaussian problem that is analogous to the linear quadratic Gaussian but with an exponentially transformed quadratic cost Later connections between risk sensitive optimal control and H theory 12 as well as differential games 13 were found In the recent years there have been some advances in risk sensitive policy search using policy gradients In 14 a policy gradient algorithm was developed that accounted for the variance in the objective either through a penalty or as a constraint The Conditional value at risk criterion was combined with policy gradients in 15 and 16 In this paper we study properties of policy gradient methods with the entropic risk measure in the objective Employing this particular type of risk measure reveals tight links to popular policy search algorithms such as reward weighted regression RWR 7 and relative entropy policy search REPS 8 III BACKGROUND ANDNOTATION In this section we provide the necessary background information on risk measures and policy search methods A Risk Sensitive Objectives and Measures The goal of an RL agent is to learn a mapping from states to actions that maximizes a performance measure 17 In the episodic RL setting 2 the episode return R also called reward is a random variable and the standard objective is to maximize the expected return Our approach considers the case where the performance measure is risk sensitive That is instead of optimizing for the long term system performance we account for higher order moments of R by performing a risk sensitive transformation of the return Risk sensitivity can be described in terms of utility 18 Namely the utility function defi nes a transformation of the reward U R The performance measure is then given by the expected utility E U R Clearly depending on the choice of function U R different behaviors will arise A special a Barrett WAM in a sim ulation environment b Barrett WAM at IAS TU Darmstadt Fig 1 Robot badminton evaluation task 2019 IEEE RSJ International Conference on Intelligent Robots and Systems IROS Macau China November 4 8 2019 978 1 7281 4003 2 19 31 00 2019 IEEE1101 choice of the utility function which attracted a lot of interest in various fi elds 18 is given by the exponential function U R exp R 1 with a risk sensitivity factor R Depending on the sign of the expected utility based on 1 needs to be either minimized or maximized 19 as explained in the following For a positive 0 the expectation E U R is a convex decreasing function of R therefore it needs to be minimized in order to maximize the reward In this case a certain expected reward with lower variance is favored and the utility function is called risk averse On the other hand when 0 and 0 this quantity is called the entropic risk mea sure 9 We slightly abuse terminology and refer to it as the entropic risk measure in the risk seeking case 0 and risk seeking 0 which corre sponds to the KL bound in 9 Note the appearance of the log partition function again 1 q logEq exp R The optimal value of is found by dual optimization argmin 0 1 q 11 Since only black box access to function R is assumed Eq 10 does not yield the new policy as an explicit function of but rather only provides samples from it A parametric policy is fi tted to the samples ob tained from 10 by moment projection 2 minimize KL k 12 maximize E q log exp R 1 q The gradient of 12 serves as the link to the risk sensitive policy gradient 6 Indeed compare KL E q h log n e 1 R 1 q oi 13 to the risk sensitive gradient 6 The correspondence is established by identifying 1 and noting that the argument in the curly braces in 13 is proportional to the one in 6 up to a scaling factor The key difference between 6 and 13 is the sampling distribution Whereas the risk sensitive policy gradient 6 requires samples from an auxiliary distribution q is used in REPS In theory it means that REPS can perform several gradient update steps according to 13 with the same samples from q while the risk sensitive gradient 6 requires gathering new data after each parameter update However in practice optimizing the ML objective 12 till 1103 convergence is problematic due to the fi nite sample size and the associated overfi tting problems 2 That is why alternatives to the policy update objective 12 are often used such as performing the information projection instead of the moment projection 24 or constraining the policy fi tting step with another KL divergence 25 which can be viewed as a form of maximum a posteriori estimation 26 Thus the policy update of REPS 12 can be identifi ed with the risk sensitive update 6 under the assumption that the information loss bound is small such that q and one step in the direction of the gradient 13 solves 12 Importantly though the temperature parameter 1 gets optimized in REPS and thus changes with iterations whereas when applying 6 it has to be scheduled manually Another interesting distinction between risk sensitive opti mization and REPS stems from the fact that the temperature parameter must be positive in REPS This means 0 the lower the variance of the returns however the mean is also lower in this case higher risks whereas lower risk returns yield lower mean reward When comparing two policies 1and 2correspond ing to risk factors 1 2 policy 1will prefer lower risk assets and yield lower return on average than 2 Return distributions for various values of the risk aversion factor are shown in Fir 3 Simulation was run over 10 random seeds with 1000 samples per trial Results confi rm the theory A more pessimistic objective higher leads to a narrow reward distribution with lower mean implying a smaller but more consistent reward Smaller values of and therefore more risk neutral objectives on the other hand lead to an asset distribution that almost entirely aims to maximize the expected reward not taking the variance into account B Toy Badminton System We consider a simplifi ed scenario of a robot learning to return a shuttlecock in the game of badminton We assume a two dimensional world and a ball following a parabolic fl ight trajectory The goal is to determine the hitting velocity of the racket which results in the ball arriving at a desired target location The hypothesis is that for different values of the risk aversion factor the agent will learn different strategies either aggressive hits but with high variability or safe returns however with smaller expected reward The problem is specifi ed as follows minimize 1 logE exp xdes x1 14 subject tox1 x0 vx 0 vy 0 g s v2 y 0 g2 2y0 g The initial position of the ball x0 y0 is assumed known Due to a perfectly elastic collision the initial velocity of the ball equals the velocity of the racket Therefore we treat the initial ball velocity as the control variable Keeping in mind that this model should resemble the real robot setup considered later we add a bit of noise to the initial ball velocity such that vx 0vy 0 v0 N u v0 with u being our control variable Constant g is the gravitational constant and the equality constraint in 14 is derived from the equations of motion Optimization variable u u 1104 0 31 0 33 E C 0 05 0 06 Var C 0 55 0 58 Var x1 6 0 6 1 6 2 v m s 10 2100102 0 0 2 e m a risk averse 0 0 30 0 32 0 05 0 06 0 60 0 62 6 25 6 35 10 2 100 102 0 2 0 b risk seeking 0 Fig 4 Simulation results on the toy badminton system with a risk sensitive objective optimized for varying risk factor values 0 01 0 1 1 5 10 100 1000 System noises are fi xed vx 0 vx 0 0 6 and the initial position is x0 y0 0 contains the parameters of the higher level policy As usually we employ a Gaussian policy u u N u u u Evaluation results on this simulated problem are shown in Fig 4 Optimization was run over a range of values of with 1000 samples per trial averaged over 10 random seeds Error e is defi ned as e xdes x1 the cost is C e and v is the initial speed Several trends can be observed in Fig 4 First variance in the landing location x1is inversely proportional to risk aversion when risk aversion increases variance in the landing location decreases However such a clear trend cannot be observed in the variance of the cost function Second from the plot of the fi nal position error e we can read that both risk seeking and risk averse policies corresponding to extreme values of fail at returning the ball to the desired target This effect is due to the dual nature of the objective function which trades mean performance agains variability Extreme risk averse policies tend to undershoot the target while extreme risk seeking ones tend to overshoot it The same conclusion can be made based on the plot of velocities v Risk averse pessimistic policies favor smaller initial velocities In contrast risk seeking policies chose larger initial velocities Third variance bars are larger for large negative values of This effect is due to objective 14 becoming very sharp close to a delta function which nega tively affects optimization C Robot Badminton Finally we proceed to apply the risk sensitive policy gradient on a real robotic system consisting of a Barrett WAM supplied with an optical tracking setup and equipped with a badminton racket see Fig 1 The goal is to learn movement primitives of different levels of riskiness on the scale between an aggressive smash and a defensive backhand To represent movements we encode them using proba bilistic movement primitives ProMPs 28 Since a ProMP is given by a distribution over trajectories generalization is a rest b hit c swing d retract Fig 5 Phases of the hitting movement of the Barrett WAM accomplished through probabilistic conditioning and trajec tories can be shaped as desired by including via points We utilize these properties of ProMPs to learn hitting movements encoded by a small set of meta parameters 20 The meta parameters in our case encode a via point defi ned by joint positions and velocities of the robot arm at a desired hitting time qT t hit q T t hit T Using the extended Kalman fi lter EKF we process shuttlecock observations and predict its future state s xT b x T b T at the interception plane Control policy s N MT s maps state s to meta parameters where random Fourier fea tures s are tuned as described in 29 Reward func tion R i x y z xi ball xi racket rtarget is based on two terms it encourages contact between the racket and the shuttlecock and it provides a bonus rtargetif the shuttlecock reaches the desired target afterwards This problem falls into the realm of contextual policy search 2 Therefore we state the optimization objective as follows maximize 1 log Z s Z s e R s d ds 15 where s is the distribution over contexts s and R s is the reward given for the combination of s and Returning a shuttlecock in badminton to a desired location requires a high degree of precision In our experiments we had to relax the requirements due to constraints of the hardware We only optimized for returning the shuttlecock at all and forced the projectile to follow the same trajectory for every iteration and always hit the same point at the interception plane 050100150 20 10 0 iterations J R 0 0 1 1 2 00 511 52 1 1 2 E R 00 511 52 0 10 0 15 0 20 Var R 00 511 52 1 4 1 5 vhit Fig 6 The upper left plot shows the convergence curves for different values of 0 The other three plots compare performance of the fi nal learned policies in terms of expected return E R variance Var R and hitting velocity vhit 1105 Evaluation results are shown in Fig 6 Policies are op timized in simulation using 35 roll outs per iteration over 150 iterations and experiments are repeated over 5 random seeds Only risk sensitive policies are shown but it is noted that risk seeking also yielded converging results Although the convergence plots for positive values of look noisier upper left plot the fi nal performance achieved by the policies trained with non zero risk aversion e g 1 or 2 is higher in terms of the expected reward upper right plot It is hard to make a conclusive judgement about the variance lower left plot due to high variability in the results Surprisingly the hitting velocity vhitis actually higher for more risk averse policies This observation stands in contrast to what we had in the toy badminton model where less aggressive policies favored smaller velocities Unfortunately with our current setup we were not able to achieve the goal of training skills of varying degree of riskiness Nevertheless we wanted to test the limits of achievable performance in the badminton task following a risk neutral objective The interception of the shuttlecock and hitting plane was enlarged to cover an area of approximately 1m2 We carried out an extended learning trial in simulation with 100 roll outs per iteration over 800 iterations The best risk neutral controller could return 95 of the served balls The learned policy could be transferred to the real robot and was able to successfully return a shuttle cock An example hitting movement is shown in Fig 5 VI CONCLUSION The entropic risk measure was considered as the opti mization objective for policy gradient methods By analyzing the exact form of its gradient we found that it is related to the standard policy gradient but inherently incorporates a baseline Furthermore risk sensitive policy update was shown to correspond to a certain limiting case of the policy update in REPS Exploring this connection to information theoretic methods appears to be a fruitful direction for future work Entanglement between exploration variance and inherent system variability was found to be a strong limiting factor Approaches for separating these two sources of uncertainty need to be searched for To reveal strengths and weaknesses of risk sensitive op timization in a real robotic context we applied our policy gradient method to the problem of learning risk sensitive movement primitives in a badminton task In a simplifi ed model we observed that policies optimized for different values of risk aversion demonstrate qualitatively different behaviors Namely risk averse policies hit the shuttlecock with smaller velocity and tended to undershoot whereas risk seeking policies favored higher velocities and typically overshot the target Finally we carried out experiments on the real robot which showed that moderate values of risk aversion can help fi nding better solutions for the original risk neutral problem However our attempt at learning risk sensitive movement primitives on the real robot had limited success due to limitations of the hardware platform and the entanglement of sources of variability REFERENCES 1 R E Bellman Dynamic programming Princeton Univ Press 1957 2 M P Deisenroth G Neumann J Peters et al A survey on policy search for robotics Foundations and TrendsR in Robotics vol 2 no 1 2 pp 1 142 2013 3 R J Williams Simple statistical gradient following algorithms for connectionist reinforcement learning Machine learning vol 8 no 3 4 pp 229 256 1992 4 R S Sutton D A McAllester S P Singh and Y Mansour Policy gradient methods for reinforcement learning with function approxima tion in NIPS 2000 pp 1057 1063 5 S M Kakade A natural policy gradient in Advances in neural inform
温馨提示
- 1. 本站所有资源如无特殊说明,都需要本地电脑安装OFFICE2007和PDF阅读器。图纸软件为CAD,CAXA,PROE,UG,SolidWorks等.压缩文件请下载最新的WinRAR软件解压。
- 2. 本站的文档不包含任何第三方提供的附件图纸等,如果需要附件,请联系上传者。文件的所有权益归上传用户所有。
- 3. 本站RAR压缩包中若带图纸,网页内容里面会有图纸预览,若没有图纸预览就没有图纸。
- 4. 未经权益所有人同意不得将文件中的内容挪作商业或盈利用途。
- 5. 人人文库网仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对用户上传分享的文档内容本身不做任何修改或编辑,并不能对任何下载内容负责。
- 6. 下载文件中如有侵权或不适当内容,请与我们联系,我们立即纠正。
- 7. 本站不保证下载资源的准确性、安全性和完整性, 同时也不承担用户因使用这些下载资源对自己和他人造成任何形式的伤害或损失。
最新文档
- 2026年山东电子职业技术学院单招综合素质考试题库附答案解析
- 2026年四川文化产业职业学院单招职业技能测试必刷测试卷附答案解析
- 2025年初级经济师之初级金融专业通关题库(附答案)
- 2026年内蒙古呼和浩特市单招职业倾向性测试必刷测试卷及答案解析(名师系列)
- 2026年哈尔滨幼儿师范高等专科学校单招职业适应性考试题库及答案解析(名师系列)
- 2026年南昌应用技术师范学院单招职业倾向性考试必刷测试卷及答案解析(名师系列)
- 2026年吕梁师范高等专科学校单招职业技能考试题库及答案解析(夺冠系列)
- 2026年亳州职业技术学院单招综合素质考试必刷测试卷及答案解析(名师系列)
- 2026年河北省衡水市单招职业适应性测试必刷测试卷及答案解析(夺冠系列)
- 2026年宁波财经学院单招职业技能测试题库及答案解析(名师系列)
- 徐工XCT75起重机详细参数说明
- 加速康复外科理念下骨盆骨折诊疗规范的专家共识
- 2025至2030中国槟榔果行业未来发展趋势及投资风险分析报告
- 涉氨制冷企业检查表
- 子宫肌瘤教学查房
- 医务人员职业暴露预防及处理课件
- 普通高中课程方案2025修订解析
- 人工智能赋能教育:探索与实践
- GB/T 2684-2025铸造用砂及混合料试验方法
- 2025年高考语文作文专项第06讲 高考新材料作文(练习)(解析版)
- 超市熟食操作管理制度
评论
0/150
提交评论