Functional contour-following via haptic perception and reinforcement learning

buir.contributor.authorTekin, Cem
dc.citation.epage72
dc.citation.issueNumber1
dc.citation.spage61
dc.citation.volumeNumber11
dc.contributor.authorHellman, R. B.
dc.contributor.authorTekin, Cem
dc.contributor.authorSchaar, M. V.
dc.contributor.authorSantos, V. J.
dc.date.accessioned2019-02-21T16:05:53Z
dc.date.available2019-02-21T16:05:53Z
dc.date.issued2018
dc.departmentDepartment of Electrical and Electronics Engineering
dc.description.abstractMany tasks involve the fine manipulation of objects despite limited visual feedback. In such scenarios, tactile and proprioceptive feedback can be leveraged for task completion. We present an approach for real-time haptic perception and decision-making for a haptics-driven, functional contour-following task: The closure of a ziplock bag. This task is challenging for robots because the bag is deformable, transparent, and visually occluded by artificial fingertip sensors that are also compliant. A deep neural net classifier was trained to estimate the state of a zipper within a robot's pinch grasp. A Contextual Multi-Armed Bandit (C-MAB) reinforcement learning algorithm was implemented to maximize cumulative rewards by balancing exploration versus exploitation of the state-action space. The C-MAB learner outperformed a benchmark Q-learner by more efficiently exploring the state-action space while learning a hard-to-code task. The learned C-MAB policy was tested with novel ziplock bag scenarios and contours (wire, rope). Importantly, this work contributes to the development of reinforcement learning approaches that account for limited resources such as hardware life and researcher time. As robots are used to perform complex, physically interactive tasks in unstructured or unmodeled environments, it becomes important to develop methods that enable efficient and effective learning with physical testbeds.
dc.description.sponsorshipThe authors wish to thank Peter Aspinall for assistance with the construction of the robot testbed. This work was supported in part by National Science Foundation Awards #1461547, #1463960, and #1533983, and the Office of Naval Research Award #N00014-16-1-2468.
dc.identifier.doi10.1109/TOH.2017.2753233
dc.identifier.eissn2329-4051
dc.identifier.issn1939-1412
dc.identifier.urihttp://hdl.handle.net/11693/50279
dc.language.isoEnglish
dc.publisherInstitute of Electrical and Electronics Engineers
dc.relation.isversionofhttps://doi.org/10.1109/TOH.2017.2753233
dc.relation.project1461547 - 1463960 - 1533983 - Office of Naval Research, ONR: 00014-16-1-2468
dc.source.titleIEEE Transactions on Haptics
dc.subjectActive touch
dc.subjectContour-following
dc.subjectDecision making
dc.subjectHaptic perception
dc.subjectManipulation
dc.subjectReinforcement learning
dc.titleFunctional contour-following via haptic perception and reinforcement learning
dc.typeArticle

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Functional_Contour_following_via_Haptic_Perception_and_Reinforcement_Learning.pdf
Size:
888.84 KB
Format:
Adobe Portable Document Format
Description:
Full printable version