Embodied Cognition for Autonomous Interactive Robots
Version of Record online: 14 AUG 2012
Copyright © 2012 Cognitive Science Society, Inc.
Topics in Cognitive Science
Volume 4, Issue 4, pages 759–772, October 2012
How to Cite
Hoffman, G. (2012), Embodied Cognition for Autonomous Interactive Robots. Topics in Cognitive Science, 4: 759–772. doi: 10.1111/j.1756-8765.2012.01218.x
- Issue online: 11 OCT 2012
- Version of Record online: 14 AUG 2012
- Received 28 December 2010; received in revised form 25 March 2012; accepted 18 April 2012
- Embodied cognition;
- Artificial intelligence;
- Perceptual symbols;
- Human–robot interaction;
- Cognitive computational models;
In the past, notions of embodiment have been applied to robotics mainly in the realm of very simple robots, and supporting low-level mechanisms such as dynamics and navigation. In contrast, most human-like, interactive, and socially adept robotic systems turn away from embodiment and use amodal, symbolic, and modular approaches to cognition and interaction. At the same time, recent research in Embodied Cognition (EC) is spanning an increasing number of complex cognitive processes, including language, nonverbal communication, learning, and social behavior.
This article suggests adopting a modern EC approach for autonomous robots interacting with humans. In particular, we present three core principles from EC that may be applicable to such robots: (a) modal perceptual representation, (b) action/perception and action/cognition integration, and (c) a simulation-based model of top-down perceptual biasing. We describe a computational framework based on these principles, and its implementation on two physical robots. This could provide a new paradigm for embodied human–robot interaction based on recent psychological and neurological findings.