ISSN: 2455-1759
Archives of Otolaryngology and Rhinology
Review Article       Open Access      Peer-Reviewed

Impact of perceptual-learning strategies and background noise on disordered speech intelligibility

Paul M Evitts1*, Connie K Porcaro2 and Tom Gollery3

1Communication Sciences and Disorders, Penn State University-Harrisburg, USA
2Department of Communication Sciences and Disorders, Valdosta State University, 1500 N. Patterson St. Valdosta, GA 31698, Georgia
3Department of Education, Southeastern University, 1000 Longfellow Blvd, Lakeland FL 33801, UK
*Corresponding authors: Paul M Evitts, Ph.D., CCC-SLP, Program Chair, Communication Sciences and Disorders, Penn State University-Harrisburg, USA, E-mail: pevitts@psu.edu
Received: 06 March, 2024 | Accepted: 29 March, 2024 | Published: 30 March, 2024
Keywords: Speech perception; Dysphonia; Speech intelligibility; Background noise; Voice disorders

Cite this as

Evitts PM, Porcaro CK, Gollery T. (2024) Impact of perceptual-learning strategies and background noise on disordered speech intelligibility. Arch Otolaryngol Rhinol 10(1): 004-015. DOI: 10.17352/2455-1759.000153

Copyright

© 2024 Evitts PM, et al. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.

Objective: There is a plethora of research showing reduced speech intelligibility for a variety of voice disorders (i.e., dysphonia, alaryngeal). Therapeutic approaches to improve intelligibility typically involve targeting the speaker (e.g., clear speech, reduced rate) with minimal attention to the listener. Therefore, there were three purposes of this study: 1) to determine the impact of background noise on the speech intelligibility of disordered speakers; 2) to determine the impact of providing listeners with perceptual-learning strategies on the speech intelligibility of speakers with a voice disorder; and 3) to determine if subjective ratings of voice quality can predict speech intelligibility.

Methods: Sentences were recorded from 12 speakers (2 typical, 3 alaryngeal, 7 dysphonic). Sentences were divided into one of three groups of signal-to-noise ratios (SNR: quiet, +5 dB SNR, and 0 dB SNR) and individually presented to 129 healthy listeners divided into one of three groups (i.e., control, acknowledgment of disorder, cognitive-perceptual strategies). Orthographic transcription was used to assess speech intelligibility. In addition, three expert listeners provided subjective voice quality ratings of all speakers.

Results: Listeners had significantly more intelligibility errors with increased background noise (p <.001) and providing strategies to listeners did not result in a statistical improvement level F(6, 486) = 1.53, p = .17, η2p = 0.02. Regression analysis showed that the subjective voice quality overall severity was able to predict speech intelligibility in the noisy condition (0 dB SNR), accounting for 37% of the variance, R2 = .365, F(1,10) = 5.759, p = .037.

Conclusion: Results suggest that increased background noise has a deleterious effect on the speech intelligibility of those with a voice disorder but that providing listeners with strategies in hopes of improving speaker intelligibility was not successful. Results did provide support, however, for the use of subjective voice quality ratings as a potential index of speech intelligibility.

Introduction

It has been said that “simply to exist as a normal human being requires interaction with other people” [1]. It then stands to reason that people with a communication disorder, including voice and speech disorders, may potentially be at risk for disrupted social interactions due to the nature of their impairment. There are three components inherent to this communicative interaction: the speaker, the listener, and the listening environment. Decades of research have shown that each plays a pivotal role in the successful verbal exchange of ideas. One group particularly vulnerable to impaired communication are those speakers with either a speech or voice disorder [2]. The two main outcome measures typically utilized in the research to determine a speaker’s ability to transmit ideas successfully to the listener are speech intelligibility and listener comprehension. Speech intelligibility has been defined as the ‘amount of speech understood from the signal alone’ [3,4] while listener comprehension (or speech comprehension as described by can be an extension of speech intelligibility as it measures a listener’s ability to either answer questions [5] or surmise the gist of the speaker’s message [6]. Although some research has shown a strong relationship between the two [7], most research shows a weak-moderate correlation [5,8,9] suggesting that listeners are not reliant on the acoustic signal alone to arrive at the speakers’ intended message and instead, rely on signal-independent information for comprehension purposes.

One recent model of speech perception that illustrates the role of the speaker and the listener and highlights the various factors that may impact both speech intelligibility and listener comprehension was developed by Evitts for alaryngeal speakers [10]. The model, originally based on dysarthric speech [11], depicts those factors that may have a positive or negative impact on the acoustic signal produced by the speaker but also incorporates the additional component of listener processing (Figure 1) as well as the new addition of environmental noise. The original model was based on the recognition of the importance of the listener in the communication exchange with speakers with dysarthria [11].

Additionally, reduced intelligibility was primarily attributed to the speech signal, hence the bulk of this body of research focused on dysarthria. A much smaller body of research has shown, however, that a disordered voice quality also plays a role in intelligibility [9,12,13], listener comprehension [9,14], and amount of listener effort [15]. Regardless, much of the research seeking to improve disordered speech intelligibility targeted speaker modifications (e.g., increased effort, reduced rate). Since speaker modifications may be inherently limited due to disease and existing disability, recent research has begun to recognize the importance of the listener. Such listener-targeted treatment utilizes a perceptual-learning approach which is well founded in the psychological literature [16]. Such treatments are based on the belief that while listeners implicitly experience perceptual learning when presented with minor acoustic degradations, more direct and explicit directions are needed to successfully process more severely disordered speech [17].

As shown in the model of speech intelligibility by Evitts (2019), any treatments aimed at the listener clearly are founded on listener processing. Multiple factors1 may play a role in this step, including listener attitudes as well as the visual information provided by the speaker [18-20]. In fact, targeting listener attitudes has been shown to have a strong and positive correlation with speech intelligibility for speakers with dysarthria [21]. Research on the use of self-disclosure to improve listener attitudes has also shown promise in the stuttering and alaryngeal speech literature [22,23]. Visual information may include such factors as facial paralysis, the presence of a stoma, the use of an electrolarynx, increased muscular effort, or other factors inherent to speech production. Recently, the COVID-19 pandemic and the subsequent use of facial masks have also shown the variable effects of the altered acoustic and visual signal on speech intelligibility across different types of speech styles (i.e., casual, clear, positive-emotional) [24]. Any of these factors discussed thus far may either have a positive or negative impact on a listeners’ ability to perceive and process the incoming signal. Cognitive workload can then be considered an index of listener processing and listeners have been shown to exert more cognitive workload and experience more errors in intelligibility when presented with dysphonic voices [9].

-----------------------

1Other factors inherent to the listener also clearly have an impact, including listeners with a hearing impairment, aphasia, and other neurologic or cognitive disorders. For the purposes of this paper, listener processing will focus on typical, healthy listeners.

-----------------------

To improve the speech intelligibility of speakers with dysarthria, several studies have targeted the listener with mixed results. Strategies such as providing topic knowledge to the listener and identifying that listeners use information on syllabic strength and lexical boundary analysis have been promising [25-27]. A recent review on dysarthria remediation, however, highlighted familiarization as perhaps the most efficacious approach to improving speech intelligibility for dysarthric speakers [17]. It should be noted, however, that most of the studies provided in the review only included speech samples from one speaker, thus significantly reducing the ability to generalize results.

Further limiting the ability to generalize the results of this body of research to a broad spectrum of disordered speakers is the fact that the speech samples used were from speakers with dysarthria. Since dysarthria involves disordered speech, there is limited information that has targeted such listener-treatment approaches for speakers with voice disorders. Recently, Porcaro, et al. (2019) provided listeners with specific perceptual-learning strategies (i.e., acknowledgment of disorder, cognitive-perceptual strategies) in hopes of improving the speech intelligibility of 12 speakers with varying degrees of dysphonia. Listeners were divided into three groups: control; acknowledgment of the disorder; and cognitive-perceptual processes. The acknowledgment tactic (self-disclosure) has been used as a strategy to increase communication or change psychosocial perceptions for people who stutter or had a laryngectomy and now use alaryngeal speech [22,23]. Listeners in the cognitive-perceptual processes group were provided with information (strategies) on how to potentially overcome reduced intelligibility due to the degraded acoustic stimuli [28]. While the dysphonic voices were significantly less intelligible than the healthy controls, the use of perceptual-learning strategies did not result in improved intelligibility [13]. A review of the speakers with dysphonia used in the study though showed that nine of the dysphonic speakers were >90% intelligible and the remaining three speakers were 84-85% intelligible. Additionally, only three of the dysphonic speakers had a subjective overall severity rating as moderate-severe as measured on the Consensus Auditory-Perceptual Evaluation of Voice (CAPE-V) [29]. Thus, conclusions on the potential efficacy of perceptual-learning strategies to improve the speech intelligibility of dysphonic speakers, particularly those with moderate-severe dysphonia, are premature.

Aside from those speakers with a speech or voice disorder, there are other instances when typical healthy speakers can present with reduced speech intelligibility, particularly those speakers with a foreign accent. For example, when English listeners are presented with highly and moderately intelligible Korean-accented English with varying degrees of background noise, results suggest that the foreign-accented speech requires more cognitive effort which ultimately may affect both intelligibility and comprehension [30]. Other research using English-as-a-Second-Language (ESL) speakers also highlights the varying importance of segmental, prosodic, and temporal features on pronunciation as well as the role of segmental accuracy for nativelike pronunciations [31].

An additional factor that has also begun to receive more recent attention in the disordered speech and voice literature is the deleterious effect of background noise on speech intelligibility for disordered speakers. Based on Figure 1, this may directly impact listener processing, which in turn, may influence speech intelligibility and listener comprehension. The inclusion of background noise as a predicating factor on speech intelligibility is well-founded due to its increased ecological validity. That is, communication does not occur in a vacuum and intuitively, both speaker and listener may be performing their roles accordingly but the message is disrupted due to the presence of increased background noise. Research has repeatedly shown that increased signal-to-noise ratio (SNR) negatively impacts speech intelligibility [32,33].

While this is certainly true for typical, healthy speakers, a nascent body of research is showing that disordered speakers are particularly vulnerable to reduced speech intelligibility, especially in the presence of background noise [2]. Such results have been shown for speakers with Parkinson’s Disease [34,35], alaryngeal speakers [2,36], and speakers with dysphonia [37]. In fact, decreased speech intelligibility across the studies was found at similar noise levels, specifically +5 dB SNR and 0 dB SNR. Furthermore, Ishikawa, et al. (2017) found that although speakers with dysphonia were significantly less intelligible than typical speakers with background noise, there were no differences between the two speaker groups in quiet. Speakers used in that study were mild-moderately dysphonic, as evidenced by a mean overall severity rating of 31 mm on the CAPE-V. It should also be noted that two of the speakers, although categorized as dysphonic, had an overall severity rating of < 10 mm. Previous work suggests that speakers with an overall severity rating on the CAPE-V of moderately-deviant may have reduced intelligilbity [9]. Finally, Yoho and Borrie (2018) investigated the effect of background noise on dysarthric speech and found no multiplicative effect of the presence of noise on intelligibility. Contrary to Ishikawa, et al. (2017), this suggests that the presence of background noise has a similar effect on both disordered and typical, healthy speech [38].

In sum, speakers with a speech or voice disorder frequently experience reduced speech intelligibility which may have a negative impact on their ability to successfully communicate their ideas and place them at a disadvantage from numerous perspectives. While there is a plethora of research detailing this reduced speech intelligibility, there is limited research that targets the listener as a means of improving speech intelligibility, particularly for speakers with a disordered voice (i.e., dysphonia, alaryngeal). Furthermore, much of the existing research is marked by reduced ecological validity in that the impact of background noise has not been fully explored. Therefore, the purpose of this study is to investigate the impact of perceptual-learning strategies provided to the listener on the speech intelligibility of disordered speakers in the presence of background noise. One previous attempt at this approach proved unsuccessful [13]. The majority of speakers in that study, however, were highly intelligible (i.e., > 90%). Thus, continued investigation is warranted. The overall hypothesis is that speech intelligibility will decrease as background noise increases across all speakers but providing listeners with perceptual-learning strategies will result in higher intelligibility scores. Specific research questions are as follows:

What is the impact of providing listeners with perceptual-learning strategies (i.e., acknowledgment, cognitive-perceptual) on the speech intelligibility of speakers with a voice disorder in the presence of background noise?

Can subjective ratings of voice quality predict speech intelligibility in a quiet setting and with increased background noise?

The second research question is included to recognize the importance of clinical subjective voice

evaluation and a continued pursuit to determine if such perceptual judgments can be used to predict impairments in speech intelligibility for both dysphonic and alaryngeal speakers [9,13]. Currently, the gold standard of assessing speech intelligibility involves listener transcription and then calculating the percent of words correctly identified [4]. Clinically, this process can be cumbersome and time-consuming. Using the CAPE-V [29], Evitts, et al. (2016) found that the rating category of overall severity was able to successfully predict speech intelligibility while Porcaro, et al. (2019) found that breathiness predicted intelligibility. More work is needed to elucidate the potential use of subjective voice evaluations to predict deficits in speech intelligibility.

Methods

This study was approved by the Institutional Review Boards of Towson University, Johns Hopkins School of Medicine, and Florida Atlantic University.

Participants

Participants included two groups: speakers and listeners. The speaker selection was based on a corpus of recordings and previously published studies from the speech perception lab of the first author (PE). Those disordered speakers from previous studies with the lowest rates of speech intelligibility were included [9,13,19]. While it is recognized that alaryngeal speech is markedly different than laryngeal speech [39], both alaryngeal and dysphonic speakers were included as the primary purpose of the study was to determine if providing listeners with cognitive-perceptual processes could improve disordered speech intelligibility, regardless of nature of the voice disorder. Ultimately, 12 speakers were included in the final study, including seven speakers with dysphonia, three alaryngeal, and two healthy controls. The seven speakers with dysphonia were all female with a mean age of 25.3 years who received a diagnosis of phonotrauma as a result of vocal nodules, vocal polyps, or secondary muscle tension dysphonia by a board-certified laryngologist at Johns Hopkins School of Medicine, Department of Otolaryngology. Alaryngeal speakers consisted of three males with a mean age of 70.3 years representing each mode of alaryngeal speech (i.e., tracheoesophageal [TE], esophageal [ES], and electrolaryngeal [EL]). Finally, two healthy female speakers with perceptually normal voice quality served as controls (mean age = 34 years). Inclusion criteria for all speakers were no history of cognitive, hearing, speech, or language impairments that affected speech or voice production (other than those directly associated with laryngeal cancer or the presence of dysphonia), and English as their first language.

Listeners were recruited through two different university programs. Inclusion criteria included: English as their primary language; no reported history of a learning disability; no reported history of a language or cognitive disorder; no history of a traumatic brain injury or any other injury that affected hearing or cognition. All listeners passed a hearing screening at 500 Hz, 1000 Hz, 2000 Hz, and 4000 Hz. Ultimately, 129 listeners (115 female, 14 male, mean age = 21.9 years) were included in the study.

Stimuli recording and preparation

All speakers were recorded in a quiet room while seated. A headset microphone (AKG C-420 III, AKG Acoustics, Vienna, Austria) was placed two inches from the corner of their mouth and audio was recorded at a 48 kHz sampling rate. Audio files were analyzed using an acoustic analysis program (Computerized Speech Laboratory 4500, KayPentax, Montvale, NJ) and saved as individual wave files. To reduce the potential of familiarity, each speaker produced a different set of phonemically balanced sentences from the Hearing in Noise Test (HINT) [40]. Similar to Ishikawa, et al. (2017), edited files were downsampled to 22500 Hz, and intensity was stabilized at 71.2-71.5 dB SPL. Background noise (+5 SNR, 0 SNR) was added to two-thirds of the files using MATLAB thus creating three sets of audio stimuli (quiet, +5 SNR, 0 SNR) across the 12 speakers. Background noise was extracted from a recording of cafeteria noise (Auditec, St. Louis, MO). Using previous methods [9,13,19], three master lists of audio stimuli were created with each list containing the 161 audio stimuli in a randomized order with a 10-second pause between each stimulus. This was done to reduce the potential of familiarity as listeners would not be presented with the same sentences. In sum, each list contained 161 sentences randomized for speaker and noise level. Each list contained 53 quiet stimuli, 54 with + 5 dB SNR background noise added, and 54 stimuli with 0 dB SNR background noise added. The 161 total stimuli also included 10% additional stimuli added in order to determine inter- and intra-rater reliability.

Listening procedure

The procedure for the listening task was also based on previous work [9,13,19]. Briefly, listeners were individually presented with 161 audio stimuli at two separate university locations. All listeners passed a hearing screening at 500, 1000, 2000, and 4000 Hz dB SPL. Listeners were seated in a soundproof booth and were first provided instructions on the task, followed by two examples and an opportunity to adjust the volume to a desired level. Listeners were also randomly assigned to one of three perceptual-learning strategy groups: control, acknowledgment strategy, and cognitive-perceptual strategy group. Listeners in the control group were provided the following instructions: “You will hear a series of sentences produced by different speakers. Please write down exactly what you hear on the form in front of you”. Based on methods by Blood and Blood (1982), listeners in the acknowledgment group were provided the following instructions: “You are going to hear a series of sentences produced by different speakers. Some of the speakers you hear will have a voice disorder. The medical term for a voice disorder is dysphonia. Dysphonia occurs when a person’s vocal folds or vocal cords do not move vibrate or move like they should. For this task, please write down exactly what you hear on the form in front of you”. Finally, listeners in the cognitive-perceptual group were provided the instructions which were based on Klasner and Yorkston (2005): “You are going to hear a series of sentences produced by different speakers. The speech and voices you hear may be difficult to understand. Please use the following strategies to help out with the task of understanding them”. Listeners were then provided with a paper with descriptions of three cognitive-perceptual strategies and were able to follow along as the examiner read each strategy.

  1. Segmental strategy: If the sounds are difficult to understand, try using the other sounds within the word to figure out what the word is.
  2. Cognitive strategy: Some of the voices that you’re going to hear may sound distorted. For those voices, try to pay close attention to the words being said.
  3. Linguistic strategy: If you’re having a hard time understanding some of the words, try using the other words around it to figure out what the words may be.

After reading the paper, the listener was instructed to “write down exactly what you hear on the form in front you using the strategies we discussed to help you”. Listeners were then presented the signals binaurally through two speakers (Bose Companion 2 Series II, Bose Corporation, Framingham, MA).

Scoring

Speaker intelligibility was determined by counting the number of correctly transcribed words in each sentence and dividing by the total number of words possible. Scores across listeners were then calculated. Since different words convey different levels of meaning, articles, and content words were also counted.

Voice quality ratings

As discussed earlier, subjective voice quality ratings are considered the gold standard of voice evaluation. In hopes of using such ratings as a more expedient and efficient measure of speech intelligibility, all speakers were evaluated by three expert, licensed, and certified speech-language pathologists, each with at least 10 years of specialized voice experience. Expert raters were provided with a randomized set of folders, each containing 10 HINT sentences produced by each speaker. Expert raters were asked to complete a CAPE-V [29] on each speaker. Expert rater responses were measured in millimeters using a digital caliper (Avenger Products, Henderson, NV).

Results

Results are divided into the following sections: preliminary analyses, descriptive results, and primary analysis.

Preliminary analyses

In order to appropriately determine the final data set, the following preliminary analyses were performed: comparison of intelligibility lists; assessment of intra- and inter-listener agreement, assessment of intra- and inter-expert rater agreement, and finally, assessment of intra-scorer measurement.

Comparison across intelligibility lists: To determine if the intelligibility lists had different rates, a one-way ANOVA was calculated. The one-way ANOVA showed no significant differences across the lists, F(2,404) = 0.0113, p = 0.989.

Intra- and inter-listener agreement: To determine the reliability between and across listeners, Cronbach’s α for a random 10% of the listeners was calculated. Values for intra-listener and inter-listener agreement ranged from 0.875 to 1.00 showing strong support for the use of the data [41,42].

Learning and fatigue effects: To determine if there were a significant number of errors in the first, middle, or third portions of the listening experiment, a one-way ANOVA was calculated for a random 10% of the listeners. The ANOVA showed there was not a significant difference in the number of errors across listening portions, F(2, 23)= 0.521, p = 0.601.

Intra- and inter-expert rater agreement: To determine the reliability of the expert listeners, Cronbach’s α was also calculated across the three listeners. All those values > 0.6 were considered acceptable [41,42]. α values were as follows: overall severity – 0.89; Roughness – 0.54; Breathiness – 0.9; Strain – 0.6; Pitch – 0.42; and Loudness – 0.48. Due to the low-reliability ratings for Roughness, Pitch, and Loudness, data from those ratings would not be used in the final analysis.

Descriptive results

Descriptive statistical techniques were used to assess mean intelligibility by speaker type and background noise level (Table 1) as well as mean intelligibility by speaker type, noise level, and strategy (Tables 2).

Article and content errors

It was also of interest to determine if there was a significant difference in article vs. content word errors across listeners (Figure 2). A one-way ANOVA showed a significant difference among modes of voice, F(1,3)=49.10, p < .001. Bonferroni Post hoc analysis showed that the number of article errors with typical voice was significantly different than alaryngeal (p < .001), and that dysphonic voice was significantly different than alaryngeal voice (p < .001). A one-way ANOVA for content words also showed a significant difference among modes of voice, F(1, 3)=302.96, p < .001. Bonferroni post analysis showed that typical was significantly different than dysphonic voice (p < .001), typical was significantly different than alaryngeal voice (p < .001), and dysphonic voice was significantly different than alaryngeal voice (p < .001) (Figure 2).

Primary analysis

Prior to the primary analysis, it was of interest to ensure that the typical speakers used in the study were, indeed, typical. To determine this, a Repeated Measures ANOVA statistical technique with one within-subjects factor was conducted to determine whether significant differences in intelligibility exist among the three speaker types of Alaryngeal, Dysphonic, and Typical featured in the study. The main effect for the within-subjects factor was statistically significant, F(2, 492) = 1010.39; p < .001, indicating there were significant differences between the values for intelligibility rates for the speaker types of Alaryngeal, Dysphonic, and Typical. The effect for speaker intelligibility across typical speakers and disordered speakers (Alaryngeal and Dysphonic) was also statistically significant, F(1, 246) = 1488.38, p < .001, indicating there were significant differences between the intelligibility rates of typical and disordered speakers (alaryngeal + dysphonic speakers combined). The mean intelligibility rate difference of 0.26 (SD = 0.01) favoring that of typical speakers was statistically significant (t (246) = 38.58, p < .001.

The first two research questions focused on the number of speech intelligibility errors by the amount of background noise and by strategy. Figure 3 provides descriptive information on the speech intelligibility of each speaker type by level of background noise.

A multivariate analysis of variance (MANOVA) statistical technique was conducted to assess if there were statistically significant differences in the linear combination of speaker type (Alaryngeal; Dysphonic; and Typical) for the levels of strategy. As a result, the main effect for strategy was not manifested at a statistically significant, level F(6, 486) = 1.53, p = .17, η2p = 0.02, indicating the linear combination of the speaker types of Alaryngeal, Dysphonic, and Typical was similar for each level of the variable strategy. The effect of strategy upon intelligibility by typical speakers and disordered speakers (Alaryngeal and Dysphonic combined) was similarly non-statistically significant, F(4, 488) = 1.19, p = .31, η2p = 0.01, indicating that the linear combination of intelligibility rates for typical speakers and disordered speakers was similar for each level of strategy.

Intelligibility of speaker type for levels of strategy

A Mixed Model ANOVA with one within-subjects factor and one between-subjects factor was conducted to determine whether significant differences exist among intelligibility rates between the levels of Strategy. The main effect for strategy was not statistically significant, F(2, 244) = 1.20, p = .30, indicating the levels of strategy were all similar for the intelligibility rates of the three speaker types represented in the study. The main effect for the within-subjects factor was statistically significant, F(2, 488) = 1003.90, p < .001, indicating there were significant differences between the values of intelligibility rates across the three speaker types. The interaction effect between the within-subjects factor and strategy was not statistically significant, F(4, 488) = 0.56, p = .66, indicating that the relationships between intelligibility rates of speaker types were similar between the levels of strategy (Table 3).

A follow-up Mixed Model ANOVA with one within-subjects factor (intelligibility rates of typical and disordered speakers) and one between-subjects factor was conducted to determine whether significant differences exist among intelligibility rates between the levels of Strategy. The main effect for Strategy was not statistically significant, F(2, 244) = 1.64, p = .20, indicating the levels of Strategy were all similar for intelligibility rates for typical and disordered speakers. The main effect for the within-subjects factor was statistically significant, F(1, 244) = 1474.22, p < .001, indicating there were significant differences between the values of intelligibility rates for typical and disordered speakers. The interaction effect between the within-subjects factor and Strategy was not significant, F(2, 244) = 0.32, p = .73, indicating that the relationship between intelligibility rates of typical and disordered speakers was similar between the levels of Strategy. Tukey HSD comparisons were conducted to assess the differences in the estimated marginal means for each combination of between-subject and within-subject effects (Table 4).

Effect of strategy upon speaker intelligibility rates by background noise level

A multivariate analysis of variance (MANOVA) statistical technique was conducted to assess if there were statistically significant differences in the linear combination of speaker type among the levels of Strategy. Using strategy as the variable, results for alaryngeal speakers did not show a significant difference between speech intelligibility levels and background noise levels (Alaryngeal Quiet, Alaryngeal +5 dB SNR, Alaryngeal 0 dB SNR), F(6, 1128) = 1.61, p = .14, ηp2 = .01. The main effect for strategy was also not statistically significant for dysphonic speakers, F(6, 116) = 1.56, p = .17, η2p = 0.07, indicating that the linear combination of the mean scores for Dysphonic Quiet, Dysphonic + 5 dB SNR, and Dysphonic 0 dB SNR was similar for each level of strategy. Finally, a multivariate analysis of variance (MANOVA) was conducted to assess if there were statistically significant differences in the linear combination of the mean scores for Typical Quiet, Typical +5 dB SNR, and Typical 0 dB SNR between the levels of strategy. The main effect for strategy was not statistically significant, F(6, 494) = 0.89, p = .50, η2p = 0.01, indicating that the linear combination of the mean scores for Typical Quiet, Typical +5 dB SNR, and Typical 0 dB SNR was similar for each level of strategy.

Predictive ability of voice quality ratings and speech intelligibility

The second research question focused on the ability of subjective voice quality ratings provided by expert listeners to predict the speech intelligibility of disordered speakers. As discussed earlier, this was included in a continued effort to explore the use of subjective voice quality ratings as a means of predicting speech intelligibility. If possible, this would be clinically useful as existing assessments such as the CAPE-V [29] could also serve to provide information on speech intelligibility in lieu of more time-consuming methods. Descriptive information on the speech intelligibility of speakers and results of the CAPE-V by expert listeners is provided in Table 5. Please note that +5 dB SNR was not used in the statistical analysis in an effort to increase the statistical power and to better reflect more severe listening conditions. Due to the low reliability of the rating roughness, pitch, and loudness, only overall severity, breathiness, and strain were included in the regression analysis.

Regression analysis showed that none of the voice qualities from the CAPE-V were able to predict speech intelligibility in the quiet condition. There were two different models that were predictive of speech intelligibility using the CAPE-V ratings as the independent variables and percent intelligibility in the noisy condition (0 dB SNR) as the dependent variable. The first model, which included the rating overall severity, was able to predict 37% of the variance associated with percent intelligibility: R2 = .365, F(1, 10) = 5.759, p = .037. The beta coefficient for this model was -.605 indicating that for every unit increase in overall severity, there would be a .605 decrease in the percent intelligibility. The second model included the ratings of overall severity and strain from the CAPE-V ratings and was able to predict 66% of the variance associated with percent intelligibility in the noisy condition: R2 = .663, F(2,9) = 8.849, p = .007. Beta coefficients in this model were -1.431 for overall severity and .990 for strain indicating that for every unit increase in overall severity and strain, there would be a 1.431 decrease and a .990 increase in speech intelligibility.

Discussion

The overarching purpose of this study was to investigate if providing listeners with cognitive-perceptual strategies would increase the speech intelligibility of disordered speakers in situations that reflect everyday communication (i.e., increased SNR/background noise). As discussed earlier, although numerous gains have been made to help disordered speakers improve their intelligibility, it is also advantageous to target reducing the effort on the part of the listener to recognize the speech [43]. With the current study, overall results showed that increased background noise did, for the most part, reduce speech intelligibility but that providing listeners with cognitive-perceptual strategies in hopes of overcoming the degraded acoustic stimuli did not help improve intelligibility scores.

Background noise and speech intelligibility

The finding that increased background noise negatively impacted speech intelligibility is not new. This is true for typical, healthy speakers [44] as well as speakers with dysarthria [38], dysphonia [37], and laryngectomy [2]. Across speaker modes, alaryngeal speakers experienced a mean 29% reduction, followed by a 13% reduction for dysphonic speakers, and a 3% mean reduction in speech intelligibility from the quiet listening condition to the noisy or 0 dB SNR. The finding that disordered speech was more vulnerable to background noise was also consistent with previous research on speakers with a disordered voice [2,45]. It should be noted that the speakers used in the Ishikawa, et al (2017) included two speakers with normal voice quality ratings (< 10 mm on the CAPE-V overall severity) and the remaining four speakers were rated as mild-moderately dysphonic (35 mm - 53 mm on the CAPE-V overall severity). Furthermore, three of the speakers’ ratings of breathiness on the CAPE-V were < 4 mm. Additionally, alaryngeal speakers used in the Eadie, et al. (2021) exhibited only ‘mild speech imprecisions’ and five speakers had ‘intact speech’. Previous research suggests that dysphonic speakers may not experience intelligibility deficits until subjective CAPE-V ratings of breathiness are moderate-severe [13] or ratings of overall severity are moderate-severe [9].

Results on the impact of background noise on the speech intelligibility of speakers with dysarthria are less conclusive. For example, Yoho and Borrie (2018) found that the combination of background noise and dysarthric speech did not have a multiplicative effect on speech intelligibility. That study only used one speaker with mild-moderate ataxia dysarthria. Conversely, other studies have suggested that dysarthric speech is more vulnerable to background noise due to the presence of a multiplicative effect of speech degradation and background noise [46,47]. Additional research is warranted to clarify this effect but, intuitively, it makes sense that disordered speakers are more susceptible to reduced speech intelligibility given adverse listening conditions.

Generally speaking, listeners utilize both bottom-up processing where phonemic and acoustic elements are analyzed as well as top-down processing where listeners rely on their own experience/knowledge to fill in the gaps of a message [43]. However, as an acoustic stimulus becomes more degraded, it forces the listener to employ more top-down processing which results in increased effort or increased cognitive workload [9,43]. Using reaction times to measure cognitive workload, Evitts, et al. (2016) found that as listeners employed more cognitive workload, intelligibility scores decreased. Listener comprehension, on the other hand, was not affected. This may speak to the relationship between intelligibility and comprehension. As discussed in Figure 1, most research shows a weak-moderate relationship between the two [5,8,9]. It may be that intelligibility tasks are inherently more difficult and require more cognitive workload than tasks aimed at comprehension where the listener only needs to arrive at the gist of the speakers’ message, not each and every individual phonetic unit.

Listener strategies and speech intelligibility

Aside from the issue of background noise, the primary aim of the study was to determine if providing listeners with cognitive-perceptual strategies improved speech intelligibility. Unfortunately, this hypothesis proved untrue. Even when the disordered speakers were combined in the current study (dysphonia + alaryngeal), providing listeners with information on the degraded auditory stimuli did not improve speech intelligibility. While this is consistent with previous results using similar methods and similar speakers [13], it is in stark contrast to numerous studies showing that listeners do, indeed, use such strategies and that intelligibility can increase when trained on such strategies. For example, initial research aimed at determining the source of listener transcription errors showed that lexical boundary errors accounted for many of those errors by typical listeners when presented with hypokinetic dysarthric speech [25]. Subsequent research showed that listeners also have different error patterns depending on the type of dysarthric speech [48]. Based on this listener-perceptual approach, Klasner and Yorkston (2005) investigated the effect of providing listeners with different strategies (segmental, suprasegmental, linguistic, cognitive) on the intelligibility of dysarthric speech. Results showed that listeners employed different strategies based on the type of dysarthric speech and that listeners used both top-down and bottom-up strategies to improve intelligibility [28].

The tactic of acknowledgment was also included in this study. This strategy included acknowledging the presence of the voice disorder for the listener prior to the listening task in hopes of reducing the cognitive demands of the listener by explaining the source of the different-sounding voice. The listener, in turn, could then theoretically focus on the intelligibility task. The approach showed promise in the psychosocial literature as a means of improving communication between speakers with a disorder and typical, healthy listeners [22,23]. Clearly, the impact of acknowledging a disorder may be limited to psychosocial perceptions related to personality and willingness to communicate, among others [22,23].

Perhaps the most promising listener-oriented approach thus far in recognizing the need to shift the burden of behavioral change from the speaker to the listener is the issue of familiarity [17]. Two comprehensive reviews of interventions targeted at listeners to improve speech intelligibility for dysarthric speakers both point to this strategy as having the most efficacy [16,17]. It is argued that familiarization with a disordered signal “induces an attentional shift toward more phonetically informative acoustic cues” [16]. This notion is supported by recent findings of increased cognitive workload when listeners are presented with dysphonic and alaryngeal speech [9,49]. In fact, research has shown that simply exposing listeners to disordered speech (passive familiarization) results in intelligibility gains, listeners trained with specific feedback on transcription tasks (explicit familiarization) made substantially higher gains in intelligibility with dysarthric speech [16]. Thus, continued efforts targeting the listener to improve speech intelligibility for disordered speakers, regardless of the nature of the disorder (voice or speech), should focus on this promising area.

Other than familiarization, there are other factors related to the listener that need to be addressed when discussing speech perception. Many of these are specifically explored by a consensus of interdisciplinary experts and resulted in a Framework for Understanding Effortful Listening (FUEL) [50]. The framework recognizes the principle that the presence of background noise results in depleted cognitive resources for the listener, thus affecting intelligibility and comprehension. Pertinent to studies targeting the listener and the perception of disordered speech, the issue of conation is also presented. Conation is a centuries-old neuropsychological concept that refers to the listener's ability to focus one’s attention on a task and, although there is overlap with motivation, is considered a separate and distinct factor [51] and should be considered in future research in the disordered speech perception literature.

In addition to conation, other factors inherent to the listener also need to be considered in future research. For example, although related to conation, motivation of the listener plays an integral role in speech perception tasks. In fact, when listeners have reduced motivation to process a degraded acoustic signal or in the presence of background noise, the result may still be no change in effort [52]. Furthermore, when listeners are given a severely degraded acoustic signal, they may implicitly determine that they will not be successful and will thus, reduce effort and cognitive resources [52]. Aside from conation and motivation, working memory and overall cognitive ability may be additional listener factors to consider in this area of research [50,52]. Given that older adults may have reduced working memory [53] and potential hearing loss, their ability to not only perform a transcription task but also to perceive individual phonetic units may be diminished, especially when presented with a degraded acoustic signal.

Voice quality ratings and speech intelligibility

This study also sought to further elucidate the role of subjective voice quality ratings in predicting speech intelligibility. This reflects a continued effort in the disordered speech literature to replace the time-consuming process of orthographic transcription to assess speech intelligilbity [54,55]. Results from this study indicate that the subjective rating of overall severity on the CAPE-V was able to account for 37% of the variance of speech intelligibility and that the ratings overall severity + strain were able to account for 66% of the variance associated with speech intelligibility with 0 dB SNR as background noise. This is promising given previous results using similar methods. Specifically, regression analysis by Evitts, et al. (2016) found that overall severity accounted for 32% of the variance and overall severity + strain accounted for 36% of the variance. However, similar regression analysis also found that only the voice quality rating breathiness was predictive of speech intelligibility, accounting for 41% of speech intelligibility [13]. Both of these results were in a quiet setting with no background noise. Regression analyses in the current study found that no voice quality ratings were able to predict speech intelligibility in a quiet setting. It should be noted that in the current study, expert listener ratings for rough, pitch, and loudness were found to be statistically unreliable and were thus, not included in the final analysis. The issue of reduced listener reliability of voice quality, even with expert listeners, certainly poses a problem as this area of research develops but clearly, there are issues with perceptual ratings of voice quality that need to be resolved [56,57].

The use of crowdsourcing may be helpful in addressing some of the issues with subjective voice quality ratings. Crowdsourcing through companies such as Amazon Mechanical Turk, offers researchers the ability to access a much larger pool of listeners than previous logistics would permit. Traditionally, researchers would need to recruit listeners and present those individual listeners with the intelligibility task in person and in a soundproof booth. All of which was very time-consuming. Crowdsourcing allows access to a potentially unlimited pool of listeners and thus, much stronger results and generalizability [58]. Since crowdsourcing platforms also generally require financial payment to participants, the issue of conation and motivation may also be better addressed through such recruiting methods.

Limitations

There were specific limitations to this study which prevented generalizing the results to all disordered speakers. First, the sample of speakers was relatively limited. Specifically, only three alaryngeal speakers were included, one from each mode of alaryngeal speech. Also, even though attempts were made to include dysphonic speakers with lower baseline intelligibility, this did not turn out to be the case. Speech intelligibility data in quiet for dysphonic speakers ranged from 85-100%. Granted, the study did include a larger sample of dysphonic speakers (n = 7) than other studies but, given the large degree of variability across dysphonic and alaryngeal speakers, a larger sample size is warranted prior to generalizing the results. In addition, other factors that may impact speech intelligibility were addressed (e.g., prosody, and articulation). The second limitation involved the reliability ratings of the expert listeners. Three of the voice quality ratings on the CAPE-V were not included in the final regression analysis. Although results were consistent with previous attempts (i.e., specific voice quality ratings were able to predict speech intelligibility), it is again, difficult to generalize the results based on the reduced reliability. Third, one could argue that the current study had reduced ecological validity. Specifically, listeners were presented with audio-only stimuli and not audiovisual. Mode of presentation (audio-only vs. audiovisual) has been shown to influence the amount of cognitive resources required to complete a listening task [59]. Since most communication includes a visual component, future research should include audiovisual stimuli. Additionally, the current study did not incorporate a dual-task paradigm when assessing speech intelligibility. Communication often occurs in situations where listeners have divided attention. Including a dual-task paradigm in future studies would afford insight into how disordered speakers are perceived in real-world situations. Fourth, the mean age of the listeners in the current study was 21.9 years. Previous research has shown that age impacts the degree of listener effort in varying degrees of background noise [60]. Future research should include a larger sample size of listeners to better reflect different age groups in order to better generalize the results.

Conclusion

The main purpose of the study was to determine the effects of background noise on disordered speech and to determine if providing listeners with strategies would improve speech intelligibility. This is in recognition of the fact that disordered speakers may have inherent limitations to the amount of improvement possible and that strategies targeting the listener are warranted. Overall, results confirmed previous findings that disordered speakers are more vulnerable to reduced speech intelligibility in the presence of background noise but that providing listeners with perceptual strategies in hopes of improving speech intelligibility was unsuccessful. In addition, results showed that the subjective voice quality rating overall severity provided by expert listeners was able to predict 37% of the variance associated with speech intelligibility. Clinically, continued investigation of voice quality ratings as an index of speech intelligibility can be a more efficient means of assessment.

Portions of the results were previously presented at the 2019 ASHA Convention, in Orlando, FL. The authors would like to thank Callan Cloonan, Elizabeth Coletti, Christie Getejanc, Anna Horn, Jillian Scott, Allison Stolz, and Sarah Yinger for their assistance with data collection and data entry. The authors would also like to thank speech-language pathologists Rina Abrams, Christina Dastolfo-Hromack, and Ashley Davis for serving as expert raters. Finally, the authors would like to thank Dr. Nirmal Srinivasan for his invaluable assistance with stimulus preparation.

  1. Gawande A. Atul Gawande Quotes. 2022. https://www.brainyquote.com/quotes/atul_gawande_527239
  2. Eadie TL, Durr H, Sauder C, Nagle K, Kapsner-Smith M, Spencer KA. Effect of Noise on Speech Intelligibility and Perceived Listening Effort in Head and Neck Cancer. Am J Speech Lang Pathol. 2021 Jun 18;30(3S):1329-1342. doi: 10.1044/2020_AJSLP-20-00149. Epub 2021 Feb 25. PMID: 33630664; PMCID: PMC8702834.
  3. Keintz CK, Bunton K, Hoit JD. Influence of visual information on the intelligibility of dysarthric speech. Am J Speech Lang Pathol. 2007 Aug;16(3):222-34. doi: 10.1044/1058-0360(2007/027). PMID: 17666548.
  4. Fontan L, Tardieu J, Gaillard P, Woisard V, Ruiz R. Relationship Between Speech Intelligibility and Speech Comprehension in Babble Noise. J Speech Lang Hear Res. 2015 Jun;58(3):977-86. doi: 10.1044/2015_JSLHR-H-13-0335. PMID: 25809922.
  5. Hustad KC, Beukelman DR. Listener comprehension of severely dysarthric speech: effects of linguistic cues and stimulus cohesion. J Speech Lang Hear Res. 2002 Jun;45(3):545-58. doi: 10.1044/1092-4388(2002/043). PMID: 12069006.
  6. Higginbotham DJ, Drazek AL, Kowarsky K, Scally C, Segal E. Discourse comprehension of synthetic speech delivered at normal and slow presentation rates. AAC: Augmentative and Alternative Communication. 1994; 10(3):191-202. doi:10.1080/07434619412331276900
  7. Beukelman DR, Yorkston KM. The relationship between information transfer and speech intelligibility of dysarthric speakers. J Commun Disord. 1979 May;12(3):189-96. doi: 10.1016/0021-9924(79)90040-6. PMID: 438358.
  8. Hustad KC. The relationship between listener comprehension and intelligibility scores for speakers with dysarthria. J Speech Lang Hear Res. 2008 Jun;51(3):562-73. doi: 10.1044/1092-4388(2008/040). PMID: 18506035; PMCID: PMC3016201.
  9. Evitts PM, Starmer H, Teets K, Montgomery C, Calhoun L, Schulze A, MacKenzie J, Adams L. The Impact of Dysphonic Voices on Healthy Listeners: Listener Reaction Times, Speech Intelligibility, and Listener Comprehension. Am J Speech Lang Pathol. 2016 Nov 1;25(4):561-575. doi: 10.1044/2016_AJSLP-14-0183. PMID: 27784031.
  10. Evitts PM. The impact of postlaryngectomy audiovisual changes on verbal communication. In: Doyle P, ed. Clinical Care and Rehabilitation in Head and Neck Cancer. Springer Publishing. 2019; 463-481.
  11. Yorkston K, Beukelman D, Strand E, Bell K. Management of motor speech disorders in children and adults. 2nd ed. Pro-Ed. 1999.
  12. Bender BK, Cannito MP, Murry T, Woodson GE. Speech intelligibility in severe adductor spasmodic dysphonia. J Speech Lang Hear Res. 2004 Feb;47(1):21-32. doi: 10.1044/1092-4388(2004/003). PMID: 15072525.
  13. Porcaro CK, Evitts PM, King N, Hood C, Campbell E, White L, Veraguas J. Effect of Dysphonia and Cognitive-Perceptual Listener Strategies on Speech Intelligibility. J Voice. 2020 Sep;34(5):806.e7-806.e18. doi: 10.1016/j.jvoice.2019.03.013. Epub 2019 Apr 25. PMID: 31031103.
  14. Lyberg-Åhlander V, Haake M, Brännström J, Schötz S, Sahlén B. Does the speaker's voice quality influence children's performance on a language comprehension test? Int J Speech Lang Pathol. 2015 Feb;17(1):63-73. doi: 10.3109/17549507.2014.898098. Epub 2014 Apr 13. PMID: 24725074.
  15. Nagle KF, Eadie TL. Perceived listener effort as an outcome measure for disordered speech. J Commun Disord. 2018 May-Jun;73:34-49. doi: 10.1016/j.jcomdis.2018.03.003. Epub 2018 Mar 13. PMID: 29567465.
  16. Borrie SA, McAuliffe MJ, Liss JM. Perceptual learning of dysarthric speech: a review of experimental studies. J Speech Lang Hear Res. 2012 Feb;55(1):290-305. doi: 10.1044/1092-4388(2011/10-0349). Epub 2011 Dec 22. PMID: 22199185; PMCID: PMC3738172.
  17. Borrie SA, Lansford KL. A Perceptual Learning Approach for Dysarthria Remediation: An Updated Review. J Speech Lang Hear Res. 2021 Aug 9;64(8):3060-3073. doi: 10.1044/2021_JSLHR-21-00012. Epub 2021 Jul 21. PMID: 34289312; PMCID: PMC8740677.
  18. Evitts PM, Van Dine A, Holler A. Effects of audio-visual information and mode of speech on listener perceptions of alaryngeal speakers. Int J Speech Lang Pathol. 2009;11(6):450-60. doi: 10.3109/17549500903003078. PMID: 21271922.
  19. Evitts PM, Portugal L, Van Dine A, Holler A. Effects of audio-visual information on the intelligibility of alaryngeal speech. J Commun Disord. 2010 Mar-Apr;43(2):92-104. doi: 10.1016/j.jcomdis.2009.10.002. Epub 2009 Nov 17. PMID: 20005524.
  20. Evitts P, Gallop R. Objective eye-gaze behaviour during face-to-face communication with proficient alaryngeal speakers: a preliminary study. Int J Lang Commun Disord. 2011 Sep-Oct;46(5):535-49. doi: 10.1111/j.1460-6984.2011.00005.x. Epub 2011 Mar 7. PMID: 21899671.
  21. Hustad KC, Gearhart KJ. Listener attitudes toward individuals with cerebral palsy who use speech supplementation strategies. Am J Speech Lang Pathol. 2004 May;13(2):168-81. doi: 10.1044/1058-0360(2004/017). PMID: 15198635.
  22. Blood GW, Blood IM. A tactic for facilitating social interaction with laryngectomees. J Speech Hear Disord. 1982 Nov;47(4):416-9. doi: 10.1044/jshd.4704.416. PMID: 7186586.
  23. Byrd CT, McGill M, Gkalitsiou Z, Cappellini C. The Effects of Self-Disclosure on Male and Female Perceptions of Individuals Who Stutter. Am J Speech Lang Pathol. 2017 Feb 1;26(1):69-80. doi: 10.1044/2016_AJSLP-15-0164. PMID: 28056467.
  24. Cohn M, Pycha A, Zellou G. Intelligibility of face-masked speech depends on speaking style: Comparing casual, clear, and emotional speech. Cognition. 2021 May;210:104570. doi: 10.1016/j.cognition.2020.104570. Epub 2021 Jan 12. PMID: 33450446.
  25. Liss JM, Spitzer S, Caviness JN, Adler C, Edwards B. Syllabic strength and lexical boundary decisions in the perception of hypokinetic dysarthric speech. J Acoust Soc Am. 1998 Oct;104(4):2457-66. doi: 10.1121/1.423753. PMID: 10491707.
  26. Liss JM, Spitzer SM, Caviness JN, Adler C. The effects of familiarization on intelligibility and lexical segmentation in hypokinetic and ataxic dysarthria. J Acoust Soc Am. 2002 Dec;112(6):3022-30. doi: 10.1121/1.1515793. PMID: 12509024; PMCID: PMC4063207.
  27. Utianski RL, Lansford KL, Liss JM, Azuma T. The Effects of Topic Knowledge on Intelligibility and Lexical Segmentation in Hypokinetic and Ataxic Dysarthria. J Med Speech Lang Pathol. 2011 Dec 1;19(4):25-36. PMID: 24569812; PMCID: PMC3738182.
  28. Klasner E, Yorkston K. Speech intelligibility in ALS and HD Dysarthria: the everyday listener's perspective. Journal of Medical Speech-Language Pathology. 2005; 13:127-139.
  29. Kempster GB, Gerratt BR, Verdolini Abbott K, Barkmeier-Kraemer J, Hillman RE. Consensus auditory-perceptual evaluation of voice: development of a standardized clinical protocol. Am J Speech Lang Pathol. 2009 May;18(2):124-32. doi: 10.1044/1058-0360(2008/08-0017). Epub 2008 Oct 16. PMID: 18930908.
  30. Wilson EO, Spaulding TJ. Effects of noise and speech intelligibility on listener comprehension and processing time of Korean-accented English. J Speech Lang Hear Res. 2010 Dec;53(6):1543-54. doi: 10.1044/1092-4388(2010/09-0100). Epub 2010 Aug 10. PMID: 20699339.
  31. Saito K. What Characterizes Comprehensible and Native-like Pronunciation Among English-as-a-Second-Language Speakers? Meta-Analyses of Phonological, Rater, and Instructional Factors. TESOL Quarterly. 2021; 55(3):866-900. doi:https://doi.org/10.1002/tesq.3027
  32. Van Engen KJ, Bradlow AR. Sentence recognition in native- and foreign-language multi-talker background noise. J Acoust Soc Am. 2007 Jan;121(1):519-26. doi: 10.1121/1.2400666. PMID: 17297805; PMCID: PMC1850527.
  33. Sperry JL, Wiley TL, Chial MR. Word recognition performance in various background competitors. J Am Acad Audiol. 1997 Apr;8(2):71-80. PMID: 9101453.
  34. Chiu YF, Forrest K. The Impact of Lexical Characteristics and Noise on Intelligibility of Parkinsonian Speech. J Speech Lang Hear Res. 2018 Apr 17;61(4):837-846. doi: 10.1044/2017_JSLHR-S-17-0205. PMID: 29587306.
  35. Chiu YF, Neel A. Predicting Intelligibility Deficits in Parkinson's Disease With Perceptual Speech Ratings. J Speech Lang Hear Res. 2020 Feb 26;63(2):433-443. doi: 10.1044/2019_JSLHR-19-00134. Epub 2020 Feb 26. PMID: 32097080.
  36. Eadie TL, Otero DS, Bolt S, Kapsner-Smith M, Sullivan JR. The Effect of Noise on Relationships Between Speech Intelligibility and Self-Reported Communication Measures in Tracheoesophageal Speakers. Am J Speech Lang Pathol. 2016 Aug 1;25(3):393-407. doi: 10.1044/2016_AJSLP-15-0081. PMID: 27379754; PMCID: PMC5270639.
  37. Ishikawa K, Boyce S, Kelchner L, Powell MG, Schieve H, de Alarcon A, Khosla S. The Effect of Background Noise on Intelligibility of Dysphonic Speech. J Speech Lang Hear Res. 2017 Jul 12;60(7):1919-1929. doi: 10.1044/2017_JSLHR-S-16-0012. PMID: 28679008; PMCID: PMC6194928.
  38. Yoho SE, Borrie SA. Combining degradations: The effect of background noise on intelligibility of disordered speech. J Acoust Soc Am. 2018 Jan;143(1):281. doi: 10.1121/1.5021254. PMID: 29390797; PMCID: PMC5775095.
  39. Robbins J, Fisher HB, Blom EC, Singer MI. A comparative acoustic study of normal, esophageal, and tracheoesophageal speech production. J Speech Hear Disord. 1984 May;49(2):202-10. doi: 10.1044/jshd.4902.202. PMID: 6716991.
  40. Nilsson M, Soli SD, Sullivan JA. Development of the Hearing in Noise Test for the measurement of speech reception thresholds in quiet and in noise. J Acoust Soc Am. 1994 Feb;95(2):1085-99. doi: 10.1121/1.408469. PMID: 8132902.
  41. Pallant J. SPSS survival manual: A step by step guide to data analysis using SPSS for windows (version 10). Buckingham Open University Press. 2001.
  42. Nunnally J, Bernstein I. The Assessment of Reliability. Psychometric Theory. 1994; 3:248-292.
  43. Prodi N, Visentin C, Farnetani A. Intelligibility, listening difficulty and listening efficiency in auralized classrooms. J Acoust Soc Am. 2010 Jul;128(1):172-81. doi: 10.1121/1.3436563. PMID: 20649212.
  44. Payton KL, Uchanski RM, Braida LD. Intelligibility of conversational and clear speech in noise and reverberation for listeners with normal and impaired hearing. J Acoust Soc Am. 1994 Mar;95(3):1581-92. doi: 10.1121/1.408545. PMID: 8176061.
  45. Ishikawa K, Boyce S, Kelchner L, Powell MG, Schieve H, de Alarcon A, Khosla S. The Effect of Background Noise on Intelligibility of Dysphonic Speech. J Speech Lang Hear Res. 2017 Jul 12;60(7):1919-1929. doi: 10.1044/2017_JSLHR-S-16-0012. PMID: 28679008; PMCID: PMC6194928.
  46. Lee Y, Sung JE, Sim H. Effects of listeners' working memory and noise on speech intelligibility in dysarthria. Clin Linguist Phon. 2014 Oct;28(10):785-95. doi: 10.3109/02699206.2014.904443. Epub 2014 Apr 8. PMID: 24712561.
  47. Dykstra A, Adams S, Jog M. The effect of background noise on the speech intensity of individuals with hypophonia associated with Parkinson's disease. Journal of Medical Speech-Language Pathology. 2012; 20:19-31.
  48. Liss JM, Spitzer SM, Caviness JN, Adler C, Edwards BW. Lexical boundary error analysis in hypokinetic and ataxic dysarthria. J Acoust Soc Am. 2000 Jun;107(6):3415-24. doi: 10.1121/1.429412. PMID: 10875386.
  49. Evitts PM, Searl J. Reaction times of normal listeners to laryngeal, alaryngeal, and synthetic speech. J Speech Lang Hear Res. 2006 Dec;49(6):1380-90. doi: 10.1044/1092-4388(2006/099). PMID: 17197503.
  50. Pichora-Fuller MK, Kramer SE, Eckert MA, Edwards B, Hornsby BW, Humes LE, Lemke U, Lunner T, Matthen M, Mackersie CL, Naylor G, Phillips NA, Richter M, Rudner M, Sommers MS, Tremblay KL, Wingfield A. Hearing Impairment and Cognitive Energy: The Framework for Understanding Effortful Listening (FUEL). Ear Hear. 2016 Jul-Aug;37 Suppl 1:5S-27S. doi: 10.1097/AUD.0000000000000312. PMID: 27355771.
  51. Reitan RM, Wolfson D. Conation: a neglected aspect of neuropsychological functioning. Arch Clin Neuropsychol. 2000 Jul;15(5):443-53. PMID: 14590220.
  52. Peelle JE. Listening Effort: How the Cognitive Consequences of Acoustic Challenge Are Reflected in Brain and Behavior. Ear Hear. 2018 Mar/Apr;39(2):204-214. doi: 10.1097/AUD.0000000000000494. PMID: 28938250; PMCID: PMC5821557.
  53. Pliatsikas C, Veríssimo J, Babcock L, Pullman MY, Glei DA, Weinstein M, Goldman N, Ullman MT. Working memory in older adults declines with age, but is modulated by sex and education. Q J Exp Psychol (Hove). 2019 Jun;72(6):1308-1327. doi: 10.1177/1747021818791994. Epub 2018 Aug 23. PMID: 30012055.
  54. Chiu YF, Neel A. Predicting Intelligibility Deficits in Parkinson's Disease With Perceptual Speech Ratings. J Speech Lang Hear Res. 2020 Feb 26;63(2):433-443. doi: 10.1044/2019_JSLHR-19-00134. Epub 2020 Feb 26. PMID: 32097080.
  55. Preminger JE, Van Tasell DJ. Quantifying the relation between speech quality and speech intelligibility. J Speech Hear Res. 1995 Jun;38(3):714-25. doi: 10.1044/jshr.3803.714. PMID: 7674662.
  56. Kreiman J, Gerratt BR, Precoda K. Listener experience and perception of voice quality. J Speech Hear Res. 1990 Mar;33(1):103-15. doi: 10.1044/jshr.3301.103. PMID: 2314068.
  57. Kreiman J, Gerratt BR, Kempster GB, Erman A, Berke GS. Perceptual evaluation of voice quality: review, tutorial, and a framework for future research. J Speech Hear Res. 1993 Feb;36(1):21-40. doi: 10.1044/jshr.3601.21. PMID: 8450660.
  58. van Brenk F, Stipancic K, Kain A, Tjaden K. Intelligibility Across a Reading Passage: The Effect of Dysarthria and Cued Speaking Styles. Am J Speech Lang Pathol. 2022 Jan 18;31(1):390-408. doi: 10.1044/2021_AJSLP-21-00151. Epub 2022 Jan 4. PMID: 34982941; PMCID: PMC9135029.
  59. Gosselin PA, Gagné JP. Older adults expend more listening effort than young adults recognizing audiovisual speech in noise. Int J Audiol. 2011 Nov;50(11):786-92. doi: 10.3109/14992027.2011.599870. Epub 2011 Sep 15. PMID: 21916790.
  60. Tangkhpanya F, Le Carrour M, Doucet F, Gagné JP. The Effort Required to Comprehend a Short Documentary in Noise: A Comparison of Younger and Older Francophones. Am J Audiol. 2019 Oct 16;28(3S):756-761. doi: 10.1044/2019_AJA-HEAL18-18-0170. Epub 2019 Oct 16. PMID: 32271122.
 

Help ?