IROS2019国际学术会议论文集2408_第1页
IROS2019国际学术会议论文集2408_第2页
IROS2019国际学术会议论文集2408_第3页
IROS2019国际学术会议论文集2408_第4页
IROS2019国际学术会议论文集2408_第5页
已阅读5页,还剩3页未读 继续免费阅读

下载本文档

版权说明:本文档由用户提供并上传,收益归属内容提供方,若内容存在侵权,请进行举报或认领

文档简介

Learning Based Robotic Bin-picking for Potentially Tangled Objects Ryo Matsumura1, Yukiyasu Domae2, Weiwei Wan1,2, and Kensuke Harada1,2 AbstractIn this research, we tackle the challenge of picking only one object from a randomly stacked pile where the objects can potentially be tangled. No solution has been proposed to solve this challenge due to the complexity of picking one and only one object from the bin of tangled objects. Therefore, we propose a method for avoiding the situation where a robot picks multiple objects. In our proposed method, first, grasping pose candidates are computed by using the graspability index. Then, a Convolutional Neural Network (CNN) is trained to predict whether or not the robot can pick one and only one object from the bin. Additionally, since a physics simulator is used to collect data to train the CNN, an automatic picking system can be built. The effectiveness of the proposed method is confirmed through experiments on robot Nextage and compare with previous bin- picking methods. I. INTRODUCTION Bin picking is a common task in the manufacturing processes, where, typically a human worker is in charge of unloading one part from the bin at a time to feed a machine. Automation of bin picking using industrial robots is then crucial, on the one hand, to relieve workers hands from this tedious, monotonous and potentially dangerous task or, on the other hand, to remove the necessity of implementing any parts-feeding machines to arrange the objects to be picked by a robot. However, robotic bin picking presents a tough challenge that in- volve looking, selecting and picking randomly oriented objects. The topic of solving the bin picking task with a robot has been extensively researched such as in 1 8, nevertheless, it remains as a difficult and complex challenge. One of the difficulties of the task is due to the complex physical phenomena of the contact between the objects and the robot fingers. Hence, the success rate of a picking task is usually between 80 and 90 % even for simple-shaped objects 5, 9, 10. Let us consider the picking task from a bin full of randomly mixed and staked parts. There are two par- ticular cases where the contact dynamics between the robot fingers and the objects makes robotic bin-picking difficult. First, while a multi-fingered gripper approaches the target object within the pile, one of the fingers may con- tact a surrounding object. In this case, if the surrounding 1Graduate School of Engineering Science, Osaka University, Japan 2Artificial Intelligence Research Center, National Institute of Advanced Industrial Science and Technology (AIST), Japan Fig. 1.Robotic bin-picking. The robot picked the target object but mistakenly picked up more objects, since they are tangled. object does not move after making contact with a finger, it is impossible for the gripper to grasp the target object. In the second case, depending on the objects shape, randomly stacked objects may be entangled with another object as shown in Fig. 1. In this case, a careful selection of the object is required to guarantee that only one is picked. If a robot picks an object entangled with other objects, the robots gripper may unexpectedly pick more than one object at the same time. In the worst scenario, the overall picking task will fail because the target object is tangled with other parts in such a way that lifting it becomes impossible. Especially, the second case has not been considered in conventional researches on robotic bin-picking. To cope with the above two problems at the same time, we propose a new learning based approach for robotic bin-picking. To pick a complex shaped object from the pile, the robot has to carefully select and pick an object such that the picked object is not entangled with another object. In our proposed method, we first generate candidates of grippers grasping poses by using the graspability index 5 where we can avoid a finger of a gripper contacting a surrounding object while the gripper approaches to the target object. The graspability generates a 2D map of the pile indicating that fingers of a gripper can be inserted in between the objects without making collision. However, just by using the graspability index, we cannot avoid a picked object being entangled with another object. Hence, we introduce the learning based approach by using a Convolutional Neural Net- work (CNN). While conventional learning based methods on robotic bin-picking just predict whether or not a robot successfully pick an object from the pile, we propose a 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) Macau, China, November 4-8, 2019 978-1-7281-4003-2/19/$31.00 2019 IEEE7984 Binary image Pixel-wise AND Reverse Contact Collision Gaussian Graspability map Peak Detected position Not peak Depth image Not peak Not peak Fig. 2.Overview of the graspability calculation. novel approach where a CNN predicts whether or not a picked object is entangled with another object. We show that a robot can pick a complex shaped object from the pile with higher success rate compared to the conventional CNN based robotic bin-picking methods. Since collecting enough number of training data used for CNN through the real world experiment is not easy especially for detecting if a picked object is entangled with another object, we use a physics simulation of picking an object from randomly stacked pile. We show that, even if we collect the training data through physics simulation, the success rate of a picking task by using our proposed method is high enough as we will discuss in the section of experiment. The rest of this paper is organized as follows. After discussing the related works in Section II, we will discuss our proposed method in Section III. In Section IV, we will show several experimental results to confirm the validity of the proposed approach. II. RELATED WORK So far, research on robotic bin-picking has been mainly done on image segmentation 1, pose identification 2 4, and picking methods 5, 6, 8, 9, 1114. Among them, the research on bin-picking method can be fur- ther categorized into three categories, i.e., the picking methods based on the pose identification of piled objects 4, 6, 8, the methods without using the pose iden- tification 5, and the learning-based methods 914. As for the methods based on the pose identification of piled objects 4, 6, 8, we need to carefully set several parameters used in both visual recognition and grasp planning corresponding for each picked object. When a robot has to pick several different objects, it is not easy for setting parameters for all objects to be picked. On the other hand, as for the method without using the pose identification, Domae et al. 5 proposed a method for determining the grasping pose of an object directly from the depth image of the pile based on the graspability index. However, it is impossible by using this method to avoid a robot picking an object entangled with another object. As for the learning based methods, Levine et al. 11 proposed an end-to-end approach by using deep neural network while this method requires extremely large number of training data obtained from real world experiments. Recently, there are some trials on reducing the effort to collect a number of training data by using a method so called GraspGAN 14, cloud database 12, 1517 and physics simulation 10. However, the con- ventional learning based approach on robotic bin-picking just predicts whether or not a robot can successfully pick an object from the pile and have not explicitly considered the case where a picked object is entangled with another object. Hence, if a robot picks a complex shaped object, a picked object may be entangled with another object. On the other hand, our proposed CNN is used to just predict whether a picked object is entangled with another object. The learning approach has also been used for grasping a novel daily object placed isolated on a table 1821 and for warehouse automation 22, 23. Pas et al. 20 developed a method for learning an antipodal grasp of a novel object by using the SVM (Support Vector Machine). Lenz et al. 19 used deep learning to detect the appropriate grasping pose of an object. Zeng et al. 22 proposed a learning based picking method used for warehouse automation. As for the manipulation of potentially-tangled object, some researchers such as 24, 25 researched the robotic manipulation of a rope. On the other hand, this research is the first trial on picking an object from the tangled pile. III. PROPOSED METHOD This section explains our proposed method for picking only one object from the bin with potentially tangled objects. We assume a parallel jaw gripper to pick an object from the pile. In our proposed method, some candidates grasping poses are first detected using the graspability index 5. These candidates are guaranteed to avoid collision of the robots finger with the nearby objects while the hand approaches the target object. Next, we explain the physics simulation used to train our proposed CNN. Our CNN is used to predict whether 7985 or not the robot can pick only one object from the bin with potentially tangled objects. Finally, the grasping pose is selected based on the harmonic mean of the graspability index and the CNN prediction. The grasping pose detected in these procedures ensure that a robot can pick only one object from the tangled pile with a high success probability. This section details the approach followed in this work. A. Graspability Index The graspability 5 is calculated by using two mask images, one describing a contact region that should be filled by a target object for stable grasping, and the other describing a collision region that should not be filled by other objects to avoid collisions during grasping. To calculate the graspability, two fingers of a gripper are represented by the cross section in the 2D horizontal plane. The graspability is computed by convolving the mask images with binarized depth maps, which are thresholded differently in each region according to the minimum height of the 3D points in the region and the length of the gripper. After computing the region of contact without collision of the gripper and object, the graspability map for the gripper model is then obtained by adding a blur with a Gaussian filter. As a result, the portion of the map with higher pixel value is regarded as more stable. These processes are as shown in Fig. 2. The grasping poses obtained are guaranteed that there is no collision between the hand and the obstacles, and that target object exists in the trajectory of closing the gripper. For our method, the top five grasping poses detected with the graspability index method were used as grasp candidates. B. Physics Simulator Our proposed method is based on supervised learning, so in order to predict whether or not the robot can pick only one object from a bin with randomly placed and possibly entangled objects, labeled training data is necessary. More specifically, the goal of our CNN is to predict whether or not an attempt to pick an object will result in one and only one object being picked. We collect training data using a physics simulator. The basic idea for collecting training data is, within the simulation, the robot attempts to pick an object from a randomly gener- ated pile of objects, then the number of objects picked is recorded. Therefore, we needed to build a simulator that could tell the number of objects picked at the end of each picking attempt. However, first, we have to define how to judge the number of objects picked in a single picking motion in the simulator, more than one object picked would mean that the target object was entangled with nearby objects. To define the entanglement in the simu- lator, we follow the reasoning described below. First, if the entanglement is defined as just the number of objects that a parallel jaw gripper successfully picked up, then the tangled objects that may drop during the motion are Fig. 3.Dataset collection with physics simulation. not taken into account. So, instead of considering only the number of objects picked in by the end of a single picking motion, we defined the entanglement as whether or not the objects near the target object moved upward during the picking motion. A threshold determines where the part was lifted or not. However, when several objects were strongly tangled with each other, they were not lifted upwards at all due to the combined weight. So, in this case, the entanglement could not be detected. So finally, we changed the specification of the simulator to automatically lift the object existing at the grasping position, without trying to simulate the grasping with the robot gripper. Based on the above, the specification of the simulator is as follow. First, our simulator generates a bin with a pile of randomly placed objects and searches for a grasping pose based on the graspability index as shown in the top left of Fig. 3. Second, the target object, the part that is in the closing trajectory of the parallel jaw gripper, is automatically lifted. At this time, this object turns green as shown in the top right of Fig. 3. Then, if objects excluding the green one move more than a certain amount in an upward direction, they turn red as shown in the bottom of Fig. 3. Finally, our simulator records the picking as tangled if the number of picked objects during a single picking motion is more than one. Otherwise, if only one object is lifted up, it records the picking as not tangled. In addition, if several objects exist in the closing trajectory, the gripper will grasp them together. So, if there are two or more objects detected in the trajectory of gripper closing, the case is immediately recorded as tangled regardless of whether or not other nearby objects move upward. By repeating this cycle, we can collect a lot of training data labeled with tangled or not tangled during a single picking motion. 7986 Conv.1 Pool.1 Fully conn.1 Conv.2 Conv.4 Conv.5Pool.2 Pool.3 Fully conn.2 Fully conn.3 Not tangle Tangle Conv.3 Fig. 4.Architecture of the convolutional neural network used for classifying whether a candidate grasping pose will result in the target object being picked without being entangled with another object. C. Convolutional Neural Network The CNNs job is to predict whether or not the robot will be able to pick one and only one object from a pile of potentially tangled objects. The input given to the CNN is a depth image of the top view of the bin. The archi- tecture of CNN is inspired by the AlexNet 26 network, which accomplished significant results in ILSVRC (Im- ageNet Large Scale Visual Recognition Challenge) 27. Here, the input to CNN is a rotated depth image so that the orientation of the robot hand is always horizontal. The reason for this is that the learning result was better when inputting only a depth image rotated according to a hand than when using both of them separately. Our CNN is composed of a series of convolutional layers followed by pooling layers as shown in Fig. 4. A max pooling of 2x2 pixel is applied in each pooling layer. The result is then processed by three fully-connected layers that classify whether or not the robot can pick only one object for the given grasping candidate. These classes are denoted by y0(Not tangled) and y1= 1-y0(Tangled) by using the following softmax function: yk= eak n i=1ea i (1) where akdenotes the weight of the input of the last fully connected layer. An activation function is used in the convolution layers and the fully connected layers to avoid the problem of gradient loss. For this reason, we use a ReLU(Rectified Linear Unit) function as the activation function: f(x) = max(x,0)(2) In our proposed method, the candidates of grasping poses detected by using the graspability index were used as the input to the CNN. Finally, the grasping pose is selected based on the harmonic mean of the graspability index and the CNN prediction. IV. EXPERIMENT A. Setup In order to evaluate the proposed method, two dif- ferent experiments were conducted with the following equipment: an industrial robot Nextage from Kawada Robotics Inc., a depth sensor, a parallel jaw gripper and Depth sensor Parallel jaw gripper Randomly stacked pile Fig. 5.Experimental setup: Depth sensor with a fixed top view of the bin, parallel jaw gripper and a bin filled with randomly placed objects. a bin with randomly placed objects. The setup is shown in Fig. 5. First, we consider testing our method on several items, a different CNN was trained for each item. The task of the robot was to pick one object from a pile full of that kind of objects. This approach was followed because we consider that whether or not the robot can pick only one object from the tangled pile strongly depends on the shape of each object. In this first experiment, the proposed method is compared to the picking method that only uses the graspability index 5 and our previous method using a CNN for discriminating success/failure of picking 10. For testing, 15 items of four different types of objects were used, as shown in Fig. 6. In the physics simulator, we use the 3D models of the objects as shown in the right column of Fig. 6. The simulation was used to collect a dataset of 5,100 pairs of depth images and the number of objects lifted. The bin-picking simulation is shown in Fig. 3. The training data is sampled such that the number of cases of picking a single object is the same as the number of cases of picking more than one object. 90% of the data is used to train the CNN and the remaining 10% is used for validation of the trained CNN. 7987 The training dataset was augmented by rotating and inverting the depth images, the number of training data obtained at the end was 36,720. For each of the

温馨提示

  • 1. 本站所有资源如无特殊说明,都需要本地电脑安装OFFICE2007和PDF阅读器。图纸软件为CAD,CAXA,PROE,UG,SolidWorks等.压缩文件请下载最新的WinRAR软件解压。
  • 2. 本站的文档不包含任何第三方提供的附件图纸等,如果需要附件,请联系上传者。文件的所有权益归上传用户所有。
  • 3. 本站RAR压缩包中若带图纸,网页内容里面会有图纸预览,若没有图纸预览就没有图纸。
  • 4. 未经权益所有人同意不得将文件中的内容挪作商业或盈利用途。
  • 5. 人人文库网仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对用户上传分享的文档内容本身不做任何修改或编辑,并不能对任何下载内容负责。
  • 6. 下载文件中如有侵权或不适当内容,请与我们联系,我们立即纠正。
  • 7. 本站不保证下载资源的准确性、安全性和完整性, 同时也不承担用户因使用这些下载资源对自己和他人造成任何形式的伤害或损失。

评论

0/150

提交评论