Team

Christoph Pohl
KIT, H2T

Fabian Reister
KIT, H2T

Fabian Peller-Konrad
KIT, H2T
Prof. Dr.-Ing Tamim Asfour
KIT, H2T
Abstract
Achieving versatile mobile manipulation in human-centered environments relies on efficiently transferring learned tasks and experiences across robots and environments. Our framework, MAkEable, facilitates seamless transfer of capabilities and knowledge across tasks, environments, and robots by integrating an affordance-based task description into the memory-centric cognitive architecture of the ARMAR humanoid robot family. This integration enables the sharing of experiences for transfer learning. By representing actions through affordances, MAkEable provides a unified framework for autonomous manipulation of known and unknown objects in diverse environments. Real-world experiments demonstrate MAkEable’s effectiveness across multiple robots, tasks, and environments, including grasping objects, bimanual manipulation, and skill transfer between humanoid robots.
System overview

MAkEable is embedded into the memory-centric cognitive architecture implemented in ArmarX. Several strategies that implement the five steps of the architecture are connected to the robot’s memory.
Use Cases
Grasping of Known Objects

Grasping of Unknown Objects

Object Placement at Common Places

Bimanual Grasping of Unknown Objects

Learning to Interact with Articulated Objects using Kinesthetic Teaching

Learning and Transferring Motion of Affordance Frames from Human Demonstration

Experiments
Clearing a Table with Known and Unknown Objects
Memory-enabled Transfer of Drawer-Opening Skill
Box Picking through Bimanual Grasping of Unknown Objects
Grasping and Pouring with Non-Humanoid Robot
BibTex
@INPROCEEDINGS {PohlReister2024,
author = {Christoph Pohl and Fabian Reister and Fabian Peller-Konrad and Tamim Asfour},
title = {MAkEable: Memory-centered and Affordance-based Task Execution Framework for Transferable Mobile Manipulation Skills},
booktitle = {IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)},
pages = {0--0},
year = {2024}
}
Acknowledgements
The research leading to these results has received funding from the German Federal Ministry of Education and Research (BMBF) under the competence center ROBDEKON (13N14678), the Carl Zeiss Foundation through the JuBot project, and European Union’s Horizon Europe Framework Programme under grant agreement No 101070596 (euROBIN).