Calinon, S. and Billard, A. (2007)
Learning of Gestures by Imitation in a Humanoid Robot
Dautenhahn, K. and Nehaniv, C.L. (eds.). Imitation and Social Learning in Robots, Humans and Animals: Behavioural, Social and Communicative Dimensions, pp. 153-177. Cambridge University Press.

Abstract

In this chapter, we explore the issue of encoding, recognizing, generalizing and reproducing arbitrary gestures. We address one major and generic issue, namely how to discover the essence of a gesture, i.e. how to find a representation of the data that encapsulates only the key aspects of the gesture, and discards the intrinsic variability across people motions.

The model is tested and validated in a humanoid robot, using kinematics data of human motion. In order for the robot to learn new skills by imitation, it must be endowed with the ability to generalize over multiple demonstrations. To achieve this, the robot must encode multivariate time-dependent data in an efficient way. Principal Component Analysis and Hidden Markov Models are used to reduce the dimensionality of the dataset and to extract the primitives of the motion.

The model takes inspiration in a recent trend of research that aims at defining a formal mathematical framework for imitation learning. In particular, it stresses the fact that the observed elements of a demonstration, and the organization of these elements should be stochastically described to have a robust robotic application. It bears similarities with theoretical models of animal imitation, and offers at the same time a probabilistic description of the data, more suitable for a real-world application.

Bibtex reference

@incollection{Calinon07chapter,
  author="Calinon, S. and Billard, A.",
  title="Learning of Gestures by Imitation in a Humanoid Robot",
  booktitle="Imitation and Social Learning in Robots, Humans and Animals: Behavioural, Social and Communicative Dimensions",
  edition="{K}. {D}autenhahn and  {C}.{L}. {N}ehaniv",
  publisher="Cambridge University Press",
  year="2007",
  pages="153--177"
}

Video

Observation of the user's gestures through stereoscopic vision and inertial motion sensors attached to the upper body of the user.

Learning and reproduction of a set of various gestures (waving goodbye, knocking on a door, dringking from a glass and writing of alphabet letters on a vertical board). For the last set of gestures, we see on the vidoe that the robot is able to generalize the motion by writing the alphabet letters on a different plane.

Go back to the list of publications