Using automatic speech recognition for the prediction of impaired speech identification (Speech In Noise Workshop, Ghent, Belgique, 10/01/2019-11/01/2019)

2020 
Age-related hearing loss (ARHL) is a very prevalent hearing disorder in adults that negatively impacts on the ability to understand speech, especially in noisy environments. The most common rehabilitation strategy is to fit hearing aids (HAs). Their benefit is generally assessed by measuring speech-identification performance with and without HAs. However, such so-called “speech audiometry” can be fairly lengthy, and its results are likely to be influenced by the patient's level of fatigue, cognitive state and familiarity with the speech material used for the assessment. In order to overcome these issues, the feasibility of using objective measures based on automatic speech recognition (ASR) to predict human speech-identification performances was recently investigated (Fontan et al., 2017; Fontan et al., in preparation; Kollmeier et al., 2016). Here, we present the results of a series of experiments, that combined ASR and an ARHL simulation to predict human performances for various tasks ranging from phoneme discrimination to sentences identification. More specifically, signal processing techniques (Nejime & Moore, 1997) were used to process the speech tokens to mimic some of the perceptual consequences of ARHL on speech perception (i.e., elevated thresholds, reduced frequency selectivity and loudness recruitment), and the processed speech tokens were then fed to an ASR system for analysis. To provide “proof-of-concept”, our first experiments focussed on the prediction of unaided speech perception in quiet, while subsequent experiments investigated the applicability of the ASR system to aided and unaided speech perception in noise.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    0
    References
    0
    Citations
    NaN
    KQI
    []