Future robots are expected to perform a multitude of complex tasks with high variability, in close collaboration or even physical contact with humans, and in industrial as well as in non-industrial settings. Both human-robot inter...
ver más
¿Tienes un proyecto y buscas un partner? Gracias a nuestro motor inteligente podemos recomendarte los mejores socios y ponerte en contacto con ellos. Te lo explicamos en este video
Proyectos interesantes
ROSETTA
RObot control for Skilled ExecuTion of Tasks in natural inte...
10M€
Cerrado
Predict-Plan-Control
Integrating robotic control and planning with human activity...
183K€
Cerrado
RobotSL
Robot skill learning imitation exploitation and control
213K€
Cerrado
REAL-RL
Model-based Reinforcement Learning for Versatile Robots in t...
2M€
Cerrado
GeRT
Generalising Robot Manipulation Task
4M€
Cerrado
CONVERGE
Convergent Human Learning for Robot Skill Generation
100K€
Cerrado
Información proyecto ROBOTGENSKILL
Duración del proyecto: 70 meses
Fecha Inicio: 2018-05-02
Fecha Fin: 2024-03-31
Fecha límite de participación
Sin fecha límite de participación.
Descripción del proyecto
Future robots are expected to perform a multitude of complex tasks with high variability, in close collaboration or even physical contact with humans, and in industrial as well as in non-industrial settings. Both human-robot interaction and task variability are major challenges. A lot of progress is needed so that: (1) robots recognize the intention of the human and react with human-like motions; (2) robot end-users, such as operators on the factory floor or people at home, are able to deploy robots for new tasks or new situations in an intuitive way, for example by just demonstrating the task to the robot.
The fundamental challenge addressed in this proposal is: how can a robot generalize a skill that has been demonstrated in a particular situation and apply it to new situations? This project focuses on skills involving rigid objects manipulated by a robot or a human and follows a model-based approach consisting of: (1) conversion of the demonstrated data to an innovative invariant representation of motion and interaction forces; (2) generalization of this representation to a new situation by solving an optimal control problem in which similarity with the invariant representation is maintained while complying with the constraints imposed by the new context. Additional knowledge about the task can be added in the constraints.
Major breakthroughs are that the required number of demonstrations and hence the training effort decrease drastically, similarity with the demonstration is maintained in view of preserving the human-like nature, and task knowledge is easily included.
The methodology is applied to program robot skills involving motion in free space (e.g. human-robot hand over tasks) as well as advanced manipulation skills involving contact (e.g. assembly, cleaning), aiming at impact in industrial and non-industrial settings.
Application of the invariant motion representation in the neighbouring field of biomechanics will further leverage impact.