Learning to Segment Object Affordances on Synthetic Data for Task-oriented Robotic Handovers

Albert Daugbjerg Christensen, Daniel Lehotský, Marius Willemoes Jørgensen, Dimitrios Chrysostomou

Research output: Contribution to book/anthology/report/conference proceedingArticle in proceedingResearchpeer-review

95 Downloads (Pure)

Abstract

The ability to perform successful robot-to-human handovers has the potential to improve robot capabilities in the circumstances involving symbiotic human-robot collaboration. Recent computer vision research has shown that object affordance segmentation can be trained on large hand-labeled datasets and perform well in task-oriented grasping pipelines. However, producing and training in such datasets can be time-consuming and resource-intensive. In this paper, we eliminate the necessity for training in these datasets by proposing a novel approach in which training occurs on a synthetic dataset that accurately translates to real-world robotic manipulation scenarios. The synthetic training dataset contains 30245 RGB images with ground truth affordance masks and bounding boxes with class labels for each rendered object. The object set used for rendering consists of 21 object classes capturing 10 affordance classes. We propose a variant of AffordanceNet enhanced with domain randomization on the generated dataset to perform affordance segmentation without the need of fine-tuning on real-world data. Our approach, outperforms the state-of-the-art method on synthetic data, by 23%, and achieves performance levels similar to other methods trained on massive, hand-labeled RGB datasets and fine-tuned on real images from the experimental setup. We demonstrate the effectiveness of our approach on a collaborative robot setup with an end-to-end robotic handover pipeline using various objects in real-world scenarios. Code, the synthetic training dataset, and supplementary material will be made publicly available.
Original languageEnglish
Title of host publicationThe 33rd British Machine Vision Conference (BMVC)
Number of pages13
PublisherBritish Machine Vision Association
Publication date1 Dec 2022
Commissioning bodyHorizon Europe
Publication statusPublished - 1 Dec 2022
EventThe 33rd British Machine Vision Conference - The Kia Oval - Home of Surrey County Cricket Club, London, United Kingdom
Duration: 21 Nov 202224 Nov 2022
Conference number: 33
https://bmvc2022.org/

Conference

ConferenceThe 33rd British Machine Vision Conference
Number33
LocationThe Kia Oval - Home of Surrey County Cricket Club
Country/TerritoryUnited Kingdom
CityLondon
Period21/11/202224/11/2022
Internet address

Bibliographical note

Funding Information:
This research was partly supported by EU’s SMART EUREKA programme S0218-chARmER, Innovation Fund Denmark (Grant no. 9118-00001B) and the H2020-WIDESPREAD project no. 857061 "Networking for Research and Development of Human Interactive and Sensitive Robotics Taking Advantage of Additive Manufacturing – R2P2".

Publisher Copyright:
© 2022. The copyright of this document resides with its authors. It may be distributed unchanged freely in print or electronic forms.

Keywords

  • bmvc
  • british machine vision conference
  • Robotic Handovers
  • object affordances
  • task-oriented grasp

Fingerprint

Dive into the research topics of 'Learning to Segment Object Affordances on Synthetic Data for Task-oriented Robotic Handovers'. Together they form a unique fingerprint.

Cite this