Deep unsupervised network for multimodal perception, representation and classification - Sorbonne Université
Article Dans Une Revue Robotics and Autonomous Systems Année : 2015

Deep unsupervised network for multimodal perception, representation and classification

Résumé

In this paper, we tackle the problem of multimodal learning for autonomous robots.Autonomous robots interacting with humans in an evolving environment need the ability to acquire knowledge from their multiple perceptual channels in an unsupervised way.Most of the approaches in the literature exploit engineered methods to process each perceptual modality. In contrast, robots should be able to acquire their own features from the raw sensors, leveraging the information elicited by interaction with their environment: learning from their sensorimotor experience would result in a more efficient strategy in a life-long perspective.To this end, we propose an architecture based on deep networks, which is used by the humanoid robot iCub to learn a task from multiple perceptual modalities (proprioception, vision, audition).By structuring high-dimensional, multimodal information into a set of distinct sub-manifolds in a fully unsupervised way, it performs a substantial dimensionality reduction by providing both a symbolic representation of data and a fine discrimination between two similar stimuli. Moreover, the proposed network is able to exploit multimodal correlations to improve the representation of each modality alone.
Fichier principal
Vignette du fichier
hal_version.pdf (8.12 Mo) Télécharger le fichier
Origine Fichiers produits par l'(les) auteur(s)
Loading...

Dates et versions

hal-01083521 , version 1 (17-11-2014)

Identifiants

Citer

Alain Droniou, Serena Ivaldi, Olivier Sigaud. Deep unsupervised network for multimodal perception, representation and classification. Robotics and Autonomous Systems, 2015, 71, pp.83-98. ⟨10.1016/j.robot.2014.11.005⟩. ⟨hal-01083521⟩
573 Consultations
2173 Téléchargements

Altmetric

Partager

More