===== SUTURO VaB RoboCup@Home 2023=====
The SUTURO VaB team video showing our results and scientific expertise.
Welcome to the SUTURO VaB team! It is a joint team between the Institute for Artificial Intellience at the University of Bremen(IAI) and the TU Vienna(TUW). By joining forces, our team for RoboCup@Home combines state-of-the-art robotic perception (TUW) with state-of-the-art robot control architectures (IAI). Both parties in the SUTURO VaB team are strongly interested in autonomous mobile manipulation on robot platforms. The robot control architecture developed by the IAI is also integrated within the CRC 1320 Everyday Activity Science and Engineering (EASE 🔗). EASE is an interdisciplinary research center at the University of Bremen that investigates everyday activity science & engineering. Its core purpose is to advance the understanding of how human-scale manipulation tasks can be mastered by robotic agents. Within EASE, general software frameworks for robotic agents are created and further developed. These frameworks (CRAM, KnowRob, Giskard, and RoboKudo) can also be well applied to the RoboCup@Home context, since there is a substantial overlap in robot capabilities required to pass the RoboCup@Home challenges with what capabilities robots within EASE provide. SUTURO VaB participants will benefit from expertise acquired by EASE researchers through multiple years of experience in working with autonomous robots performing everyday activities. TUW delivers a comprehensive suite of perception modules that provide semantic information, such as object identities, locations and grasp configurations. The focus of the research of these perception approaches is the target domain of mobile robot manipulation. We are looking forward to successfully employing and extending our approaches based on the exiciting challenges in the RoboCup@Home domain.The HSR at cleaning up. This video is accelerated to 600% speed. This results with a real-time factor of approximately 0.22 to an effective speed of 132%.
The HSR at storing groceries. This video is accelerated to 600% speed. This results with a real-time factor of approximately 0.22 to an effective speed of 132%.
Ana Khasia khasia[at]uni-bremen.de Informatics M.Sc. |
Malte Dörgeloh mdoergel[at]uni-bremen.de Informatics M.Sc. | Kamal Atteibi atteibi[at]uni-bremen.de Informatics M.Sc. |
Tede von Knorre tede[at]uni-bremen.de Informatics B.Sc. |
---|
Karla Brück kbrueck[at]uni-bremen.de Informatics M.Sc. |
Phillip Kehr pkehr[at]uni-bremen.de Informatics B.Sc. |
---|
Lennart Christian Heinbokel lenhei[at]uni-bremen.de Informatics M.Sc. |
Naser Azizi naser1[at]uni-bremen.de Informatics M.Sc. |
Sorin Arion sorin[at]uni-bremen.de Informatics M.Sc. |
---|---|---|
Alexander Haberl Weibel[at]acin.tuwien.ac.at Informatics M.Sc. |
Christian Eder Weibel[at]acin.tuwien.ac.at Informatics M.Sc. |
Jean-Baptiste Weibel Weibel[at]acin.tuwien.ac.at Informatics M.Sc. |
Peter Hoenig Weibel[at]acin.tuwien.ac.at Informatics M.Sc. |
Felix Krause krause4[at]uni-bremen.de Informatics B.Sc. |
Luca Krohm luc_kro[at]uni-bremen.de Informatics B.Sc. |
Tim Alexander Rienits trienits[at]uni-bremen.de Informatics M.Sc. |
---|
Knowledge:
To fulfill complex tasks a robot needs knowledge and memory of its environment. While the robot acts in its world, it recognizes objects and manipulaties them through pick-and-place tasks. With the use of KnowRob, a belief state provides episodic memory of the robot's experience, recording the robot's memory of each cognitive activity. Via ontologies objects can be classified and put into context, which enables logical reasoning over the environment and intelligent decision making.
Manipulation:
For manipulation tasks in the environment, we use the open source motion planning framework Giskard.
It uses constraint and optimization based task space control to generate trajectories for the whole body of mobile manipulators.
Giskard offers interfaces to plan and execute motion goals and to modify its world model.
A selection of predefined basic motion goals can be arbitrarily combined to describe a motion.
If an environment model is present, such goals can also be defined on the environment, e.g., to open a door.
Simulation of the HSR while opening a human sized door. This gif is accelerated to 200% speed. This results with a real-time factor of approximately 0.46 to an effective speed of 92%.
Planning:
Planning is responsible for the high-level control and failure handling of the autonomous robot system by utilizing generic cognitive strategies. It combines the frameworks of perception, knowledge and manipulation within high-level plans written within the CRAM (Cognitive Robot Abstract Machine) system. CRAM enables the implementation of various recovery strategies for failures, as well as a lightweight simulation tool for prospection, allowing to simulate the potential outcome of the current plans and their respective parameters before executing it on the real robot, hereby increasing the success of the performed action by discarding faulty parameters in advance. CRAM also allows the high-level plans to be written generically in a way, so that he plans are robot-platform independent. The plan execution results can furthermore be recorded in order to be reasoned about in the future, adapting and increasing the success of upcoming plan performance.
Perception:
The perception framework has the task to process the visual data received by the robot's camera sensors and establish the communication between the high-level and visual perception. RoboKudo is an open source robotic perception framework based on the principles of unstructured information management.
The framework allows for the creation of perception systems that employ an ensembles of experts approach and treat perception as a question-answering problem.
Based on the queries issued to the system a perception plan is created consisting of a list of experts to be executed.
The perception experts generate object hypotheses, annotate these hypotheses and test and rank them in order to come up with the best possible interpretation of the data and generate the answer to the query. The methods provided by TUW are integrated as experts into the RoboKudo perception framework allowing the reasoning about the perception results and the communication with the high-level planning CRAM to close the perception-action loop.
{{ :robocupfiles:objects_perception01.png?nolink&200 |}}