Cue Integration With Categories: Weighting Acoustic Cues in Speech Using Unsupervised Learning and Distributional Statistics

Authors

Errata

This article is corrected by:

  1. Errata: Erratum for: Cue Integration With Categories: Weighting Acoustic Cues in Speech Using Unsupervised Learning and Distributional Statistics, by Joseph C. Toscano and Bob McMurray in Cognitive Science, 34(3) Volume 36, Issue 7, 1337–1338, Article first published online: 3 August 2012

should be sent to Joseph Toscano, Department of Psychology, E11 SSH, University of Iowa, Iowa City, IA 52242. E-mail: joseph-toscano@uiowa.edu

Abstract

During speech perception, listeners make judgments about the phonological category of sounds by taking advantage of multiple acoustic cues for each phonological contrast. Perceptual experiments have shown that listeners weight these cues differently. How do listeners weight and combine acoustic cues to arrive at an overall estimate of the category for a speech sound? Here, we present several simulations using a mixture of Gaussians models that learn cue weights and combine cues on the basis of their distributional statistics. We show that a cue-weighting metric in which cues receive weight as a function of their reliability at distinguishing phonological categories provides a good fit to the perceptual data obtained from human listeners, but only when these weights emerge through the dynamics of learning. These results suggest that cue weights can be readily extracted from the speech signal through unsupervised learning processes.

Ancillary