1

IROS 2020 – New Horizons for Robot Learning

Abstract

Robot learning combines the challenges of understanding, modeling and applying dynamical systems with task learning from rewards, through human robot interaction or from intrinsic motivation signals. While outstanding results using machine and deep learning have been generated in robot learning in the last years, current challenges in industrial applications are underrepresented. The goal of this workshop is to go beyond discussing potential industrial applications like in related past workshops.

These topics were discussed with Pieter Abbeel, Dileep George, Sergey Levine, Jan Peters, Freek Stulp, Marc Toussaint, Patrick van der Smagt, and Georg von Wichert.

Links

Details to the workshop, the speakers and links to slides can be found on the workshop webpage.

 




Elmar Rückert, „KI-Denker der Zukunft 2019“

Print Media Article in the BILANZ magazine 2019


Links

Link to the PDF.




NeurIPS 2019 – Robot open-Ended Autonomous Learning

NeurIPS 2019 Competition Track

Open-ended learning aims to build learning machines and robots that are able to acquire skills and knowledge in an incremental fashion in a certain environment. This competition addresses autonomous open-ended learning with a focus on simulated robot systems that: (a) acquire a sensorimotor competence that allows them to interact with objects and physical environments; (b) learn in a fully autonomous way, i.e. with no human intervention (e.g., no tasks or reward functions) on the basis of mechanisms such as curiosity, intrinsic motivations, task-free reinforcement learning, self-generated goals, and any other mechanism that might support autonomous learning. The competition challenge will feature two phases: during an initial “”intrinsic phase”” the system will have a certain time to freely explore and learn in an environment containing multiple objects, and then during an “”extrinsic phase”” the quality of the autonomously acquired knowledge will be measured with tasks unknown at design time and during the intrinsic phase.

Links

Details on the competition can be found on the project webpage.

 

Publications

2020

Cartoni, E.; Mannella, F.; Santucci, V. G.; Triesch, J.; Rueckert, E.; Baldassarre, G.

REAL-2019: Robot open-Ended Autonomous Learning competition Journal Article

In: Proceedings of Machine Learning Research, vol. 123, pp. 142-152, 2020, (NeurIPS 2019 Competition and Demonstration Track).

Links | BibTeX

REAL-2019: Robot open-Ended Autonomous Learning competition




H2020 Goal-Robots 11/2016-10/2020

This project aims to develop a new paradigm to build open-ended learning robots called `Goal-based Open ended Autonomous Learning’ (GOAL). GOAL rests upon two key insights. First, to exhibit an autonomous open-ended learning process, robots should be able to self-generate goals, and hence tasks to practice. Second, new learning algorithms can leverage self-generated goals to dramatically accelerate skill learning. The new paradigm will allow robots to acquire a large repertoire of flexible skills in conditions unforeseeable at design time with little human intervention, and then to exploit these skills to efficiently solve new user-defined tasks with no/little additional learning.

Link: http://www.goal-robots.eu




Loomo Innovation am Campus Lübeck

Print Media Article im FMSH Magazin 2019


Links

Link to the PDF.




How to build and use a low-cost sensor glove

This post discusses how to develop a low cost sensor glove with tactile feedback using flex sensors and small vibration motors. MATLAB and JAVA code is linked.

Note that this project is not longer maintained. Use the GitHub project instead.

Publications

2016

Weber, Paul; Rueckert, Elmar; Calandra, Roberto; Peters, Jan; Beckerle, Philipp

A Low-cost Sensor Glove with Vibrotactile Feedback and Multiple Finger Joint and Hand Motion Sensing for Human-Robot Interaction Proceedings Article

In: Proceedings of the IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN), 2016.

Links | BibTeX

A Low-cost Sensor Glove with Vibrotactile Feedback and Multiple Finger Joint and Hand Motion Sensing for Human-Robot Interaction

2015

Rueckert, Elmar; Lioutikov, Rudolf; Calandra, Roberto; Schmidt, Marius; Beckerle, Philipp; Peters, Jan

Low-cost Sensor Glove with Force Feedback for Learning from Demonstrations using Probabilistic Trajectory Representations Proceedings Article

In: ICRA 2015 Workshop on Tactile and force sensing for autonomous compliant intelligent robots, 2015.

Links | BibTeX

Low-cost Sensor Glove with Force Feedback for Learning from Demonstrations using Probabilistic Trajectory Representations

Details to the Hardware

  • Arduino Mega 2560 Board
  • Check which USB device is used (e.g., by running dmesg). On most of our machines it is /dev/ttyACM0
  • Enable read/write permissions if necessary, e.g., run sudo chmod o+rw /dev/ttyACM0
  • Serial protocoll based communication: Flex sensor readings are streamed and Vibration motor PWM values can be set between 0 and 255
  • Firmware can be found here (follow the instructions in the README.txt to compile and upload the firmware)
  • Features frame rates of up to 350Hz
  • Five flex sensors provide continuous readings within the range [0, 1024]

Simple Matlab Serial Interface (max 100Hz)

  • Download the Matlab demo code from here
  • Tell Matlab which serial ports to use: copy the java.opts file to your Matlab bin folder, e.g., to /usr/local/MATLAB/R2012a/bin/glnxa64/
  • Run FastComTest.m

Fast Mex-file based Matlab Interface – max 350Hz

  • Install libserial-dev
  • Compile the mex function with: mex SensorGloveInterface.cpp -lserial
  • Run EventBasedSensorGloveDemo.m

 




AI and Learning in Robotics

Robotics AI requires autonomous learning capabilities

The challenges in understanding human motor control, in brain-machine interfaces and anthropomorphic robotics are currently converging. Modern anthropomorphic robots with their compliant actuators and various types of sensors (e.g., depth and vision cameras, tactile fingertips, full-body skin, proprioception) have reached the perceptuomotor complexity faced in human motor control and learning. While outstanding robotic and prosthetic devices exist, current brain machine interfaces (BMIs) and robot learning methods have not yet reached the required autonomy and performance needed to enter daily life.

The groups vision is that four major challenges have to be addressed to develop truly autonomous learning systems. These are, (1) the decomposability of complex motor skills into basic primitives organized in complex architectures, (2) the ability to learn from partial observable noisy observations of inhomogeneous high-dimensional sensor data, (3) the learning of abstract features, generalizable models and transferable policies from human demonstrations, sparse rewards and through active learning, and (4), accurate predictions of self-motions, object dynamics and of humans movements for assisting and cooperating autonomous systems.




Neural and Probabilistic Robotics

Neural and Probabilistic Robotics

Neural models have incredible learning and modeling capabilities which was demonstrated in complex robot learning tasks (e.g., Martin Riedmiller’s or Sergey Levine’s work). While these results are promising we lack a theoretical understanding of the learning capabilities of such networks and it is unclear how learned features and models can be reused or exploited in other tasks.

The ai-lab investigates deep neural network implementations that are theoretical grounded in the framework of probabilistic inference and develops deep transfer learning strategies for stochastic neural networks. We evaluate our models in challenging robotics applications where the networks have to scale to high-dimensional control signals and need to generate reactive feedback command in real-time.

Our developments will enable complex online adaptation and skill learning behavior in autonomous systems and will help to gain a better understanding of the meaning and function of the learned features in large neural networks with millions of parameters.




NIPS 2016 – Neurorobotics: A Chance for New Ideas, Algorithms and Approaches

Abstract

Modern robots are complex machines with many compliant actuators and various types of sensors including depth and vision cameras, tactile electrodes and dozens of proprioceptive sensors. The obvious challenges are to process these high dimensional input patterns, memorize low dimensional representations of them and to generate the desired motor commands to interact in dynamically changing environments. Similar challenges exist in brain machine interfaces (BMIs) where complex prostheses with perceptional feedback are controlled, or in motor neuroscience where in addition cognitive features need to be considered. Despite this broad research overlap the developments happened mainly in parallel and were not ported or exploited in the related domains. The main bottleneck for collaborative studies has been a lack of interaction between the core robotics, the machine learning and the neuroscience communities.

Link

Link to the workshop description at the nips webpage. 

Agenda and Speaker

Day One, Neurorobotics WS, Fri Dec 9th 2016
  14.20-14.30 Introduction by Elmar Rueckert and Martin Riedmiller  
Session One: Reinforcement Learning, Imitation, and Active Learning
1 14.30-15.00 Juergen Schmidhuber (Scientific Director of the Swiss AI Lab IDSIA)  
  15.00-15.30 Posters and Coffee  
2 15.30-16.00 Sergey Levine (University of California, Berkeley)  
3 16.00-16.30 Pieter Abbeel (University of California, Berkeley)  
4 16.30-17.00 Johanni Brea (École polytechnique fédérale de Lausanne, EPFL)  
  17.00-17.20 Posters and Coffee  
5 17.20-17.45 Paul Schrater (University of Minnesota)  
6 17.45-18.10 Frank Hutter (University Freiburg)  
7 18.10-18.35 Raia Hadsell (Google DeepMind)  
  18.35-19.00 Panel Discussion, Session One
Day Two, Neurorobotics WS, Sat Dec 10th 2016
Session One: Reinforcement Learning, Imitation, and Active Learning
  08.30-08.35 Introduction by Elmar Rueckert and Martin Riedmiller  
8 08.35-09.05 Robert Legenstein (Graz University of Technology)  
9 09.05-09.35 Sylvain Calinon(Idiap Research Institute, EPFL Lausanne)  
10 09.35-10.05 Chelsea Finn (University of California, Berkeley)  
11 10.05-10.35 Peter Stone (University of Texas at Austin)  
  10.35-11.00 Posters and Coffee  
12 11.00-11.30 Paul Verschure (Catalan Institute of Advanced Research)  
Session Two: Model Representations and Features
13 11.30-12.00 Tobi Delbrück (University of Zurich and ETH Zurich)  
14 12.00-12.30 Moritz Grosse-Wentrup (Max Planck Institute Tuebingen)  
15 12.30-13.00 Kristian Kersting (Technische Universität Dortmund)  
  13.00-14.00 Lunch break  
Session Three: Feedback and Control
16 14.00-14.30 Emo Todorov (University of Washington)  
17 14.30-15.00 Richard Sutton (University of Alberta)  
  15.00-15.30 Posters and Coffee  
18 15.30-16.00 Bert Kappen (Radboud University)  
19 16.00-16.30 Jean-Pascal Pfister (University of Zurich and ETH Zurich)  
  16.30-17.00 Posters and Coffee  
20 17.00-17.30 Jan Babic (Josef Stefan Institute Ljubijana)  
21 17.30-18.00 Martin Giese (University Clinic Tübingen)  
  18.00-18.30 Panel Discussion, Session Two and Session Three

Accapted Wokshop Papers

  • Kyuhwa Lee, Ruslan Aydarkhanov, Luca Randazzo and José Millán. Neural Decoding of Continuous Gait Imagery from Brain Signals. (ID 2)
  • Aviv Tamar, Garrett Thomas, Tianhao Zhang, Sergey Levine and Pieter Abbeel. Episodic MPC Improvement with the Hindsight Plan. (ID 11)
  • Jim Mainprice, Arunkumar Bryavan, Daniel Kappler, Dieter Fox, Stefan Schaal and Nathan Ratliff. Functional manifold projections in Deep-LEARCH. (ID 12)
  • Nutan Chen, Maximilian Karl and Patrick van der Smagt. Dynamic Movement Primitives in Latent Space of Time-Dependent Variational Autoencoders. (ID 1)
    → Alexander Gabriel, Riad Akrour and Gerhard Neumann. Empowered Skills. (ID 7)
  • Dieter Buechler, Roberto Calandra and Jan Peters. Modeling Variability of Musculoskeletal Systems with Heteroscedastic Gaussian Processes. (ID 10)
  • David Sharma, Daniel Tanneberg, Moritz Grosse-Wentrup, Jan Peters and Elmar Rueckert. Adapting Brain Signals with Reinforcement Learning Strategies for Brain Computer Interfaces. (ID 16)
  • Dmytro Velychko, Benjamin Knopp and Dominik Endres. The Variational Coupled Gaussian Process Dynamical Model. (ID 5)
  • Felix End, Riad Akrour and Gerhard Neumann. Layered Direct Policy Search for Learning Hierarchical Skills. (ID 6)
  • Erwan Renaudo, Benoît Girard, Raja Chatila and Mehdi Khamassi. Bio-inspired habit learning in a robotic architecture. (ID 9)

Organizer


Elmar Rueckert is a postdoctoral scholar at the Intelligent Autonomous Systems lab headed by Jan Peters. He has a strong expertise in learning spiking neural networks, probabilistic planning and robot control. Before joining IAS in 2014, he has been with the Institute for Theoretical Computer Science at Graz University of Technology, where he received his Ph.D. under the supervision of Wolfgang Mass. His Thesis, “On Biologically inspired motor skill learning in robotics through probabilistic inference” concentrated on probabilistic inference for motor skill learning and on learning biologically inspired movement representations.

Martin Riedmiller joined Google DeepMind in 2015 as research scientist. He received a Diploma in Computer Science in 1992 and a PhD on Self-learning Neural Controllers in 1996 from University of Karlsruhe. He has been a professor at TU Dortmund (2002), University of Osnabrück (2003-2009), and University of Freiburg (2009-2015) where he headed the Machine Learning Lab. His general research interest is applying machine learning techniques to interesting real world problems. His RoboCup team Brainstormes won five international competitions in the 2D Simulation and MiddleSize leagues.