Abstract
Slip detection enables robotic hands to perform complex manipulation tasks by predicting when a held object is about to be dropped. Here we use a support vector machine classifier to detect slip with a biomimetic optical tactile sensor: the TacTip. Previously, this method has been shown to be effective on various artificial stimuli such as flat or curved surfaces. Here, we investigate whether this method generalises to novel, everyday objects. Five different objects are tested which vary in shape, weight, compliance and texture as well as being common objects that one might encounter day-to-day. Success of up to 90% is achieved which demonstrates the classifier’s ability to generalise to a variety of previously unseen, natural objects.
Supported by the EPSRC Centre for Doctoral Training in Future Autonomous and Robotic Systems (FARSCOPE) and a Leadership Award from the Leverhulme Trust on ‘A biomimetic forebrain for robot touch’ (RL-2016-39).
Access provided by CONRICYT-eBooks. Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
Humans have highly effective slip detection mechanisms that are deployed when grasping an object. Humans continuously adjust the grasp to prevent an object from being dropped [1]. Meissner corpuscles respond to local skin movements that are present when slip occurs which initiates a reflexive response [2]. Meissner corpuscles are densely populated in the fingertip which is why the human hand is such an effective grasping mechanism. Mimicking the function of the Meissner corpuscles and subsequent response in a robotic hand will yield a more sophisticated sense of touch and enable complex object manipulation by reducing the likelihood of an object being dropped [3].
Slip detection has been a poplar area of research since the 1980s with many different sensors and methods used [4, 5]. The aim of this study is to investigate whether a previously developed slip detection method generalises to novel objects. We use the TacTip biomimetic tactile sensor [6]. Previous work with the TacTip has focused mainly on object perception [7, 8] so slip detection capabilities will further demonstrate the sensor’s versatility.
2 Methods
The TacTip is a 3D-printed, biomimetic, optical tactile sensor containing 127 pins on the inside of a soft, compliant hemisphere. A camera is focused on the pins and records their positions when the TacTip contacts an object. The TacTip is mounted on a six degree-of-freedom robotic arm (UR5, Universal Robots).
For this work we use the same classification method described in James et al. [9]. This involves using the velocity of each pin in the TacTip as an input to a support vector machine (SVM). Image capture, processing and classification leads to the data being sampled at 100 FPS.
Being able to detect the velocity of each pin individually is analogous to the Meissner corpuscles which each contain a single nerve fiber capable of detecting slip in its receptive field. This will allow for slip to be detected in objects where contact area varies.
This work, presented by James et al. has already demonstrated a high degree of success at slip detection and preventing objects from being dropped. However, the objects used were regularly shaped, solid objects and a variety of common household objects was not tested. Here, we use the same classifier used to test those objects which is not modified or re-trained in any way.
3 Experiments
Each experiment proceeds as follows. The UR5 robotic arm - to which the TacTip is attached - presses an object against a vertically mounted metal plate until it is securely held with no slipping. The arm then begins to slowly retract (0.2 mms\(^{-1}\)) until the object slips (Fig. 1).
The data from the sensor is continually being passed through the SVM classifier and if slip is detected the arm moves forward to try to prevent the object from being dropped. The experiment is considered a success if the object is visually seen to start falling, the classifier detects the slip, the arm moved forward and the object is secured against the wall without further slippage. Any other outcomes such as slip being detected but the object still being dropped is classed as a failure.
The five objects chosen were; a paper box containing a plastic cube, a metal coffee container, a mustard bottle, a water bottle (approx. 1/3 full) and a plastic banana. Figure 2 shows each object that was tested. The objects were chosen as each has a different shape, weight, compliance and texture as well as being common objects and shapes that one might encounter day-to-day.
4 Results and Discussion
Each object was tested ten times in the manner described above and the results are summarised in Table 1. The results show that four of the five objects are stopped from falling with a high success of up to 90%. The only low scoring object is the banana which was only stopped 30% of the time. For this case the classifier was able to detect the onset of slip but the shape of the banana meant that moving to re-secure it mostly pushed the banana out of the way and onto the floor.
Each object used was novel to the classifier and no parameters of the object were known to the classifier prior to each experiment. This demonstrates the versatility of the classifier to generalise to previously unseen objects.
Having a single tactile sensor moving with a single degree of freedom means that the ability to handle complex objects is severely limited. However, the fact that slip was reliably detected, even on the banana, means that using a gripper with more degrees of freedom is likely to be successful at preventing slipping objects from being dropped.
References
Johansson, R.S., Flanagan, J.R.: Coding and use of tactile signals from the fingertips in object manipulation tasks. Nat. Rev. Neurosci. 10(5), 345 (2009)
Saal, H.P., Bensmaia, S.J.: Touch is a team effort: interplay of submodalities in cutaneous sensibility. Trends Neurosci. 37(12), 689–697 (2014)
Yousef, H., Boukallel, M., Althoefer, K.: Tactile sensing for dexterous in-hand manipulation in robotics-a review. Sens. Actuators A Phys. 167(2), 171–187 (2011)
Howe, R.D., Cutkosky, M.R.: Sensing skin acceleration for slip and texture perception. In: 1989 Proceedings of the IEEE International Conference on Robotics and Automation, pp. 145–150. IEEE (1989)
Veiga, F., Van Hoof, H., Peters, J., Hermans, T.: Stabilizing novel objects by learning to predict tactile slip. In: 2015 Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 5065–5072. IEEE (2015)
Chorley, C., Melhuish, C., Pipe, T., Rossiter, J.: Development of a tactile sensor based on biologically inspired edge encoding. In: 2009 Proceedings of the International Conference on Advanced Robotics, ICAR 2009, pp. 1–6. IEEE (2009)
Lepora, N.F., Aquilina, K., Cramphorn, L.: Exploratory tactile servoing with active touch. IEEE Robot. Autom. Lett. 2(2), 1156–1163 (2017)
Ward-Cherrier, B., Pestell, N., Cramphorn, L., Winstone, B., Giannaccini, M.E., Rossiter, J., Lepora, N.F.: The TacTip family: soft optical tactile sensors with 3D-printed biomimetic morphologies. Soft Robot. (2018)
James, J.W., Pestell, N., Lepora, N.F.: Slip detection with a biomimetic tactile sensor. IEEE Robotics and Automation Letters (2018, in press)
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer International Publishing AG, part of Springer Nature
About this paper
Cite this paper
James, J.W., Lepora, N.F. (2018). Slip Detection on Natural Objects with a Biomimetic Tactile Sensor. In: Vouloutsi , V., et al. Biomimetic and Biohybrid Systems. Living Machines 2018. Lecture Notes in Computer Science(), vol 10928. Springer, Cham. https://doi.org/10.1007/978-3-319-95972-6_24
Download citation
DOI: https://doi.org/10.1007/978-3-319-95972-6_24
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-95971-9
Online ISBN: 978-3-319-95972-6
eBook Packages: Computer ScienceComputer Science (R0)