Spectral weights for localization and speech-in-speech recognition with spatial separation of talkers on the horizontal plane

用于定位和语音中语音识别的频谱权重,其中说话者在水平面上具有空间分离

阅读:1

Abstract

Some previous research has suggested that sound source localization may not rely on the same cues that support the segregation of speech produced by talkers separated in space. The present experiments evaluated spectral weights for the spatial cues underlying these two tasks by filtering stimuli into 1-octave-wide bands and dispersing them on the horizontal plane. Target stimuli were 100-ms bursts of speech-shaped noise or words produced by 24 male and female talkers, and maskers (when present) were sequences of words. For localization in quiet, weights differed depending on the midpoint and band dispersion range, but they were similar for speech and noise stimuli. For bands dispersed between -15° and +15°, weights peaked at 500 and 1000 Hz. Introducing a speech masker changed the magnitude of weights for localization, but not the relative weight by frequency. For speech-in-speech recognition, sequences of masker words produced predominantly informational masking, such that participants had to rely on spatial cues to segregate the target. As for localization, recognition appeared to rely predominantly on spatial cues in the 500- and 1000-Hz bands. Trial-by-trial data suggest that correct word recognition relied on differences in perceived location of target and masker speech for some but not for all participants.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。