8th Speech in Noise Workshop, 7-8 January 2016, Groningen

Pure linguistic interference during comprehension of competing speech signals

Bohan Dai(a)
Max Planck Institute for Psycholinguistics, 6500 AH Nijmegen, NL

Anne Kösem(b)
Donders Institute for Brain, Cognition and Behaviour, Radboud University, 6500 HB Nijmegen, NL

James M. McQueen(b)
Donders Institute for Brain, Cognition and Behaviour, Radboud University, 6500 HB Nijmegen, NL; Max Planck Institute for Psycholinguistics, 6500 AH Nijmegen, NL

Peter Hagoort
Max Planck Institute for Psycholinguistics, 6500 AH Nijmegen, NL; Donders Institute for Brain, Cognition and Behaviour, Radboud University, 6500 HB Nijmegen, NL

(a) Presenting
(b) Attending

Human listeners have more difficulty in understanding speech in a multi-talker environment than in the presence of non-intelligible noise. The costs of speech-in-speech masking have been attributed to informational masking, i.e. to the competing processing of the target and the distractor speech’s information. Yet, it remains unclear what kind of information is competing, as intelligible speech and unintelligible speech-like signals (e.g. reversed, noise-vocoded, and foreign speech) differ both in linguistic content and in acoustic information. Thus, intelligible speech could be a stronger distractor than unintelligible speech because it presents closer acoustic information to the target speech, or because it carries competing linguistic information. In this study, we intended to isolate the linguistic component of speech-in-speech masking and we tested its influence on the comprehension of target speech.

To do so, participants performed a dichotic listening task in which the interfering stimuli consisted of noise-vocoded sentences that could become intelligible through training. The experiment included three steps: first, the participants were instructed to report the clear target speech from a mixture of one clear speech channel and one unintelligible noise-vocoded speech channel; second, they were trained on the interfering noise-vocoded sentences so that they became intelligible; third, they performed the dichotic listening task again. Crucially, before and after training, the distractor speech had the same acoustic features but not the same linguistic information. We thus predicted that the distracting noise-vocoded signal would interfere more with target speech comprehension after training than before training. To control for practice/fatigue effects, we used additional noise-vocoded sentences that participants were not trained on, as interfering signals in the dichotic listening tasks. We expected that performance on these trials would not change after training, or would change less than that on trials with trained noise-vocoded sentences.

The first results are consistent with our predictions. The trained noise-vocoded speech interfered more with target speech intelligibility after training than before training, but only for low SNR (-6 dB). Crucially, the interference was significantly stronger for trained noise-vocoded sentences than for untrained noise-vocoded speech, ruling out a fatigue effect. In line with past reports, the present results show that intelligible distractors interfere more with the processing of target speech. These findings further suggest that speech-in-speech interference originates, to a certain extent, from the parallel processing of competing linguistic content.

Last modified 2016-05-12 14:22:09