40 results on '"Deroche, Mickael L. D."'
Search Results
2. Cochlear Implant Compression Optimization for Musical Sound Quality in MED-EL Users
- Author
-
Gilbert, Melanie L., Deroche, Mickael L. D., Jiradejvong, Patpong, Chan Barrett, Karen, and Limb, Charles J.
- Published
- 2021
- Full Text
- View/download PDF
3. Cochlear Implant Compression Optimization for Musical Sound Quality in MED-EL Users
- Author
-
Gilbert, Melanie L., Deroche, Mickael L. D., Jiradejvong, Patpong, Chan Barrett, Karen, and Limb, Charles J.
- Published
- 2022
- Full Text
- View/download PDF
4. Processing of Acoustic Cues in Lexical-Tone Identification by Pediatric Cochlear-Implant Recipients
- Author
-
Peng, Shu-Chen, Lu, Hui-Ping, Lu, Nelson, Lin, Yung-Song, Deroche, Mickael L. D., and Chatterjee, Monita
- Abstract
Purpose: The objective was to investigate acoustic cue processing in lexical-tone recognition by pediatric cochlear-implant (CI) recipients who are native Mandarin speakers. Method: Lexical-tone recognition was assessed in pediatric CI recipients and listeners with normal hearing (NH) in 2 tasks. In Task 1, participants identified naturally uttered words that were contrastive in lexical tones. For Task 2, a disyllabic word ("yanjing") was manipulated orthogonally, varying in fundamental-frequency (F0) contours and duration patterns. Participants identified each token with the second syllable "jing" pronounced with Tone 1 (a high level tone) as "eyes" or with Tone 4 (a high falling tone) as "eyeglasses." Results: CI participants' recognition accuracy was significantly lower than NH listeners' in Task 1. In Task 2, CI participants' reliance on F0 contours was significantly less than that of NH listeners; their reliance on duration patterns, however, was significantly higher than that of NH listeners. Both CI and NH listeners' performance in Task 1 was significantly correlated with their reliance on F0 contours in Task 2. Conclusion: For pediatric CI recipients, lexical-tone recognition using naturally uttered words is primarily related to their reliance on F0 contours, although duration patterns may be used as an additional cue.
- Published
- 2017
- Full Text
- View/download PDF
5. Adaptation to pitch-altered feedback is independent of one’s own voice pitch sensitivity
- Author
-
Alemi, Razieh, Lehmann, Alexandre, and Deroche, Mickael L. D.
- Published
- 2020
- Full Text
- View/download PDF
6. A tonal-language benefit for pitch in normally-hearing and cochlear-implanted children
- Author
-
Deroche, Mickael L. D., Lu, Hui-Ping, Kulkarni, Aditya M., Caldwell, Meredith, Barrett, Karen C., Peng, Shu-Chen, Limb, Charles J., Lin, Yung-Song, and Chatterjee, Monita
- Published
- 2019
- Full Text
- View/download PDF
7. Perception of Child-Directed Versus Adult-Directed Emotional Speech in Pediatric Cochlear Implant Users
- Author
-
Barrett, Karen Chan, Chatterjee, Monita, Caldwell, Meredith T., Deroche, Mickael L. D., Jiradejvong, Patpong, Kulkarni, Aditya M., and Limb, Charles J.
- Published
- 2020
- Full Text
- View/download PDF
8. Neurophysiological Differences in Emotional Processing by Cochlear Implant Users, Extending Beyond the Realm of Speech
- Author
-
Deroche, Mickael L. D., Felezeu, Mihaela, Paquette, Sébastien, Zeitouni, Anthony, and Lehmann, Alexandre
- Published
- 2019
- Full Text
- View/download PDF
9. Dynamic networks differentiate the language ability of children with cochlear implants.
- Author
-
Koirala, Nabin, Deroche, Mickael L. D., Wolfe, Jace, Neumann, Sara, Bien, Alexander G., Doan, Derek, Goldbeck, Michael, Muthuraman, Muthuraman, and Gracco, Vincent L.
- Subjects
COCHLEAR implants ,LANGUAGE ability ,CHILDREN'S language ,POOR children ,SUPPORT vector machines - Abstract
Background: Cochlear implantation (CI) in prelingually deafened children has been shown to be an effective intervention for developing language and reading skill. However, there is a substantial proportion of the children receiving CI who struggle with language and reading. The current study-one of the first to implement electrical source imaging in CI population was designed to identify the neural underpinnings in two groups of CI children with good and poor language and reading skill. Methods: Data using high density electroencephalography (EEG) under a resting state condition was obtained from 75 children, 50 with CIs having good (HL) or poor language skills (LL) and 25 normal hearing (NH) children. We identified coherent sources using dynamic imaging of coherent sources (DICS) and their effective connectivity computing time-frequency causality estimation based on temporal partial directed coherence (TPDC) in the two CI groups compared to a cohort of age and gender matched NH children. Findings: Sources with higher coherence amplitude were observed in three frequency bands (alpha, beta and gamma) for the CI groups when compared to normal hearing children. The two groups of CI children with good (HL) and poor (LL) language ability exhibited not only different cortical and subcortical source profiles but also distinct effective connectivity between them. Additionally, a support vector machine (SVM) algorithm using these sources and their connectivity patterns for each CI group across the three frequency bands was able to predict the language and reading scores with high accuracy. Interpretation: Increased coherence in the CI groups suggest overall that the oscillatory activity in some brain areas become more strongly coupled compared to the NH group. Moreover, the different sources and their connectivity patterns and their association to language and reading skill in both groups, suggest a compensatory adaptation that either facilitated or impeded language and reading development. The neural differences in the two groups of CI children may reflect potential biomarkers for predicting outcome success in CI children. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
10. Timing variability of sensorimotor integration during vocalization in individuals who stutter
- Author
-
Sares, Anastasia G., Deroche, Mickael L. D., Shiller, Douglas M., and Gracco, Vincent L.
- Published
- 2018
- Full Text
- View/download PDF
11. Reverberation limits the release from informational masking obtained in the harmonic and binaural domains
- Author
-
Deroche, Mickael L. D., Culling, John F., Lavandier, Mathieu, and Gracco, Vincent L.
- Published
- 2017
- Full Text
- View/download PDF
12. Grouping by Time and Pitch Facilitates Free but Not Cued Recall for Word Lists in Normally-Hearing Listeners.
- Author
-
Sares, Anastasia G., Gilbert, Annie C., Zhang, Yue, Iordanov, Maria, Lehmann, Alexandre, and Deroche, Mickael L. D.
- Subjects
MEMORY ,PHYSIOLOGICAL aspects of speech ,TIME ,MUSICAL pitch ,RESEARCH funding ,DESCRIPTIVE statistics ,MUSICAL perception ,PROMPTS (Psychology) - Abstract
Auditory memory is an important everyday skill evaluated more and more frequently in clinical settings as there is recently a greater recognition of the cost of hearing loss to cognitive systems. Testing often involves reading a list of unrelated items aloud; but prosodic variations in pitch and timing across the list can affect the number of items remembered. Here, we ran a series of online studies on normally-hearing participants to provide normative data (with a larger and more diverse population than the typical student sample) on a novel protocol characterizing the effects of suprasegmental properties in speech, namely investigating pitch patterns, fast and slow pacing, and interactions between pitch and time grouping. In addition to free recall, and in line with our desire to work eventually with individuals exhibiting more limited cognitive capacity, we included a cued recall task to help participants recover specifically the words forgotten during the free recall part. We replicated key findings from previous research, demonstrating the benefits of slower pacing and of grouping on free recall. However, only slower pacing led to better performance on cued recall, indicating that grouping effects may decay surprisingly fast (over a matter of one minute) compared to the effect of slowed pacing. These results provide a benchmark for future comparisons of short-term recall performance in hearing-impaired listeners and users of cochlear implants. Graphical Abstract This is a visual representation of the abstract. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
13. Visual biases in evaluation of speakers' and singers' voice type by cis and trans listeners.
- Author
-
Knight, Jay Marchand, Sares, Anastasia G., and Deroche, Mickael L. D.
- Subjects
TRANSGENDER people ,SINGERS ,SOPRANOS (Singers) ,HUMAN voice ,JUDGES ,IMPLICIT learning - Abstract
Introduction: A singer's or speaker's Fach (voice type) should be appraised based on acoustic cues characterizing their voice. Instead, in practice, it is often influenced by the individual's physical appearance. This is especially distressful for transgender people who may be excluded from formal singing because of perceived mismatch between their voice and appearance. To eventually break down these visual biases, we need a better understanding of the conditions under which they occur. Specifically, we hypothesized that trans listeners (not actors) would be better able to resist such biases, relative to cis listeners, precisely because they would be more aware of appearance-voice dissociations. Methods: In an online study, 85 cisgender and 81 transgender participants were presented with 18 different actors singing or speaking short sentences. These actors covered six voice categories from high/bright (traditionally feminine) to low/dark (traditionally masculine) voices: namely soprano, mezzo-soprano (referred to henceforth as mezzo), contralto (referred to henceforth as alto), tenor, baritone, and bass. Every participant provided voice type ratings for (1) Audio-only (A) stimuli to get an unbiased estimate of a given actor's voice type, (2) Video-only (V) stimuli to get an estimate of the strength of the bias itself, and (3) combined Audio-Visual (AV) stimuli to see how much visual cues would affect the evaluation of the audio. Results: Results demonstrated that visual biases are not subtle and hold across the entire scale, shifting voice appraisal by about a third of the distance between adjacent voice types (for example, a third of the bass-to-baritone distance). This shift was 30% smaller for trans than for cis listeners, confirming our main hypothesis. This pattern was largely similar whether actors sung or spoke, though singing overall led to more feminine/high/bright ratings. Conclusion: This study is one of the first demonstrations that transgender listeners are in fact better judges of a singer's or speaker's voice type because they are better able to separate the actors' voice from their appearance, a finding that opens exciting avenues to fight more generally against implicit (or sometimes explicit) biases in voice appraisal. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
14. Luminance effects on pupil dilation in speech-in-noise recognition.
- Author
-
Zhang, Yue, Malaval, Florian, Lehmann, Alexandre, and Deroche, Mickael L. D.
- Subjects
PUPILLARY reflex ,PUPILLOMETRY ,SIGNAL-to-noise ratio ,CLINICAL medicine ,HEARING disorders ,DAYLIGHT - Abstract
There is an increasing interest in the field of audiology and speech communication to measure the effort that it takes to listen in noisy environments, with obvious implications for populations suffering from hearing loss. Pupillometry offers one avenue to make progress in this enterprise but important methodological questions remain to be addressed before such tools can serve practical applications. Typically, cocktail-party situations may occur in less-than-ideal lighting conditions, e.g. a pub or a restaurant, and it is unclear how robust pupil dynamics are to luminance changes. In this study, we first used a well-known paradigm where sentences were presented at different signal-to-noise ratios (SNR), all conducive of good intelligibility. This enabled us to replicate findings, e.g. a larger and later peak pupil dilation (PPD) at adverse SNR, or when the sentences were misunderstood, and to investigate the dependency of the PPD on sentence duration. A second experiment reiterated two of the SNR levels, 0 and +14 dB, but measured at 0, 75, and 220 lux. The results showed that the impact of luminance on the SNR effect was non-monotonic (sub-optimal in darkness or in bright light), and as such, there is no trivial way to derive pupillary metrics that are robust to differences in background light, posing considerable constraints for applications of pupillometry in daily life. Our findings raise an under-examined but crucial issue when designing and understanding listening effort studies using pupillometry, and offer important insights to future clinical application of pupillometry across sites. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
15. Specificity of Affective Responses in Misophonia Depends on Trigger Identification.
- Author
-
Savard, Marie-Anick, Sares, Anastasia G., Coffey, Emily B. J., and Deroche, Mickael L. D.
- Subjects
AUDITORY masking ,AFFECT (Psychology) ,PLEASANTNESS & unpleasantness (Psychology) ,JUDGMENT (Psychology) ,SIGNAL-to-noise ratio ,PSYCHOACOUSTICS ,INTEREST rates - Abstract
Individuals with misophonia, a disorder involving extreme sound sensitivity, report significant anger, disgust, and anxiety in response to select but usually common sounds. While estimates of prevalence within certain populations such as college students have approached 20%, it is currently unknown what percentage of people experience misophonic responses to such "trigger" sounds. Furthermore, there is little understanding of the fundamental processes involved. In this study, we aimed to characterize the distribution of misophonic symptoms in a general population, as well as clarify whether the aversive emotional responses to trigger sounds are partly caused by acoustic salience of the sound itself, or by recognition of the sound. Using multi-talker babble as masking noise to decrease participants' ability to identify sounds, we assessed how identification of common trigger sounds related to subjective emotional responses in 300 adults who participated in an online study. Participants were asked to listen to and identify neutral, unpleasant and trigger sounds embedded in different levels of the masking noise (signal-to-noise ratios: −30, −20, −10, 0, +10 dB), and then to evaluate their subjective judgment of the sounds (pleasantness) and emotional reactions to them (anxiety, anger, and disgust). Using participants' scores on a scale quantifying misophonia sensitivity, we selected the top and bottom 20% scorers from the distribution to form a Most-Misophonic subgroup (N = 66) and Least-Misophonic subgroup (N = 68). Both groups were better at identifying triggers than unpleasant sounds, which themselves were identified better than neutral sounds. Both groups also recognized the aversiveness of the unpleasant and trigger sounds, yet for the Most-Misophonic group, there was a greater increase in subjective ratings of negative emotions once the sounds became identifiable, especially for trigger sounds. These results highlight the heightened salience of trigger sounds, but furthermore suggest that learning and higher-order evaluation of sounds play an important role in misophonia. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
16. Voice emotion recognition by Mandarin‐speaking pediatric cochlear implant users in Taiwan.
- Author
-
Lin, Yung‐Song, Wu, Che‐Ming, Limb, Charles J., Lu, Hui‐Ping, Feng, I. Jung, Peng, Shu‐Chen, Deroche, Mickael L. D., and Chatterjee, Monita
- Subjects
EMOTION recognition ,COCHLEAR implants ,ABSOLUTE pitch ,AUDITORY perception ,SCHOOL children ,HUMAN voice ,VOCODER - Abstract
Objectives: To explore the effects of obligatory lexical tone learning on speech emotion recognition and the cross‐culture differences between United States and Taiwan for speech emotion understanding in children with cochlear implant. Methods: This cohort study enrolled 60 cochlear‐implanted (cCI) Mandarin‐speaking, school‐aged children who underwent cochlear implantation before 5 years of age and 53 normal‐hearing children (cNH) in Taiwan. The emotion recognition and the sensitivity of fundamental frequency (F0) changes for those school‐aged cNH and cCI (6–17 years old) were examined in a tertiary referred center. Results: The mean emotion recognition score of the cNH group was significantly better than the cCI. Female speakers' vocal emotions are more easily to be recognized than male speakers' emotion. There was a significant effect of age at test on voice recognition performance. The average score of cCI with full‐spectrum speech was close to the average score of cNH with eight‐channel narrowband vocoder speech. The average performance of voice emotion recognition across speakers for cCI could be predicted by their sensitivity to changes in F0. Conclusions: Better pitch discrimination ability comes with better voice emotion recognition for Mandarin‐speaking cCI. Besides the F0 cues, cCI are likely to adapt their voice emotion recognition by relying more on secondary cues such as intensity and duration. Although cross‐culture differences exist for the acoustic features of voice emotion, Mandarin‐speaking cCI and their English‐speaking cCI peer expressed a positive effect for age at test on emotion recognition, suggesting the learning effect and brain plasticity. Therefore, further device/processor development to improve presentation of pitch information and more rehabilitative efforts are needed to improve the transmission and perception of voice emotion in Mandarin. Level of evidence: 3. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
17. Neural Correlates of Vocal Pitch Compensation in Individuals Who Stutter.
- Author
-
Sares, Anastasia G., Deroche, Mickael L. D., Ohashi, Hiroki, Shiller, Douglas M., and Gracco, Vincent L.
- Subjects
INDEPENDENT component analysis ,FUNCTIONAL magnetic resonance imaging ,SENSORIMOTOR integration ,WAGES ,AUDITORY hallucinations - Abstract
Stuttering is a disorder that impacts the smooth flow of speech production and is associated with a deficit in sensorimotor integration. In a previous experiment, individuals who stutter were able to vocally compensate for pitch shifts in their auditory feedback, but they exhibited more variability in the timing of their corrective responses. In the current study, we focused on the neural correlates of the task using functional MRI. Participants produced a vowel sound in the scanner while hearing their own voice in real time through headphones. On some trials, the audio was shifted up or down in pitch, eliciting a corrective vocal response. Contrasting pitch-shifted vs. unshifted trials revealed bilateral superior temporal activation over all the participants. However, the groups differed in the activation of middle temporal gyrus and superior frontal gyrus [Brodmann area 10 (BA 10)], with individuals who stutter displaying deactivation while controls displayed activation. In addition to the standard univariate general linear modeling approach, we employed a data-driven technique (independent component analysis, or ICA) to separate task activity into functional networks. Among the networks most correlated with the experimental time course, there was a combined auditory-motor network in controls, but the two networks remained separable for individuals who stuttered. The decoupling of these networks may account for temporal variability in pitch compensation reported in our previous work, and supports the idea that neural network coherence is disturbed in the stuttering brain. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
18. Adults who stutter and metronome synchronization: evidence for a nonspeech timing deficit.
- Author
-
Sares, Anastasia G., Deroche, Mickael L. D., Shiller, Douglas M., and Gracco, Vincent L.
- Subjects
- *
MUSICAL perception , *STUTTERING , *ADULTS , *SYNCHRONIZATION , *EVIDENCE - Abstract
Speech timing deficits have been proposed as a causal factor in the disorder of stuttering. The question of whether individuals who stutter have deficits in nonspeech timing is one that has been revisited often, with conflicting results. Here, we uncover subtle differences in a manual metronome synchronization task that included tempo changes with adults who stutter and fluent speakers. We used sensitive circular statistics to examine both asynchrony and consistency in motor production. While both groups displayed a classic negative mean asynchrony (tapping before the beat), individuals who stutter anticipated the beat even more than their fluent peers, and their consistency was particularly affected at slow tempi. Surprisingly, individuals who stutter did not have problems with interval correction at tempo changes. We also examined the influence of music experience on synchronization behavior in both groups. While music perception and training were related to synchronization behavior in fluent participants, these correlations were not present for the stuttering group; however, one measure of stuttering severity (self‐rated severity) was negatively correlated with music training. Overall, we found subtle differences in paced auditory−motor synchronization in individuals who stutter, consistent with a timing problem extending to nonspeech. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
19. Processing of Acoustic Information in Lexical Tone Production and Perception by Pediatric Cochlear Implant Recipients.
- Author
-
Deroche, Mickael L. D., Lu, Hui-Ping, Lin, Yung-Song, Chatterjee, Monita, and Peng, Shu-Chen
- Subjects
ABSOLUTE pitch ,TONE (Phonetics) ,COCHLEAR implants ,INFORMATION processing ,NATIVE language ,EYEGLASSES - Abstract
Purpose: This study examined the utilization of multiple types of acoustic information in lexical tone production and perception by pediatric cochlear implant (CI) recipients who are native speakers of Mandarin Chinese. Methods: Lexical tones were recorded from CI recipients and their peers with normal hearing (NH). Each participant was asked to produce a disyllabic word, yan jing , with which the first syllable was pronounced as Tone 3 (a low dipping tone) while the second syllable was pronounced as Tone 1 (a high level tone, meaning "eyes") or as Tone 4 (a high falling tone, meaning "eyeglasses"). In addition, a parametric manipulation in fundamental frequency (F0) and duration of Tones 1 and 4 used in a lexical tone recognition task in Peng et al. (2017) was adopted to evaluate the perceptual reliance on each dimension. Results: Mixed-effect analyses of duration, intensity, and F0 cues revealed that NH children focused exclusively on marking distinct F0 contours, while CI participants shortened Tone 4 or prolonged Tone 1 to enhance their contrast. In line with these production strategies, NH children relied primarily on F0 cues to identify the two tones, whereas CI children showed greater reliance on duration cues. Moreover, CI participants who placed greater perceptual weight on duration cues also tended to exhibit smaller changes in their F0 production. Conclusion: Pediatric CI recipients appear to contrast the secondary acoustic dimension (duration) in addition to F0 contours for both lexical tone production and perception. These findings suggest that perception and production strategies of lexical tones are well coupled in this pediatric CI population. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
20. Segregation of voices with single or double fundamental frequencies.
- Author
-
Deroche, Mickael L. D. and Gracco, Vincent L.
- Subjects
- *
HUMAN voice , *AUDITORY selective attention , *AUDITORY perception , *LISTENING , *HEARING , *SEMANTICS - Abstract
In cocktail-party situations, listeners can use the fundamental frequency (F0) of a voice to segregate it from competitors, but other cues in speech could help, such as co-modulation of envelopes across frequency or more complex cues related to the semantic/syntactic content of the utterances. For simplicity, this (non-pitch) form of grouping is referred to as "articulatory." By creating a new type of speech with two steady F0s, it was examined how these two forms of segregation compete: articulatory grouping would bind the partials of a double-F0 source together, whereas harmonic segregation would tend to split them in two subsets. In experiment 1, maskers were two same-male sentences. Speech reception thresholds were high in this task (vicinity of 0 dB), and harmonic segregation behaved as though double-F0 stimuli were two independent sources. This was not the case in experiment 2, where maskers were speech-shaped complexes (buzzes). First, double-F0 targets were immune to the masking of a single-F0 buzz matching one of the two target F0s. Second, double-F0 buzzes were particularly effective at masking a single-F0 target matching one of the two buzz F0s. As a conclusion, the strength of F0-segregation appears to depend on whether the masker is speech or not. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
21. Modulation of Speech Motor Learning with Transcranial Direct Current Stimulation of the Inferior Parietal Lobe.
- Author
-
Deroche, Mickael L. D., Nguyen, Don L., and Gracco, Vincent L.
- Subjects
PARIETAL lobe ,MOTOR learning - Abstract
The inferior parietal lobe (IPL) is a region of the cortex believed to participate in speech motor learning. In this study, we investigated whether transcranial direct current stimulation (tDCS) of the IPL could influence the extent to which healthy adults (1) adapted to a sensory alteration of their own auditory feedback, and (2) changed their perceptual representation. Seventy subjects completed three tasks: a baseline perceptual task that located the phonetic boundary between the vowels/e/ and /a/; a sensorimotor adaptation task in which subjects produced the word "head" under conditions of altered or unaltered feedback; and a post-adaptation perceptual task identical to the first. Subjects were allocated to four groups which differed in current polarity and feedback manipulation. Subjects who received anodal tDCS to their IPL (i.e., presumably increasing cortical excitability) lowered their first formant frequency (F1) by 10% in opposition to the upward shift in F1 in their auditory feedback. Subjects who received the same stimulation with unaltered feedback did not change their production. Subjects who received cathodal tDCS to their IPL (i.e., presumably decreasing cortical excitability) showed a 5% adaptation to the F1 alteration similar to subjects who received sham tDCS. A subset of subjects returned a few days later to reiterate the same protocol but without tDCS, enabling assessment of any facilitatory effects of the previous tDCS. All subjects exhibited a 5% adaptation effect. In addition, across all subjects and for the two recording sessions, the phonetic boundary was shifted toward the vowel /e/ being repeated, consistently with the selective adaptation effect, but a correlation between perception and production suggested that anodal tDCS had enhanced this perceptual shift. In conclusion, we successfully demonstrated that anodal tDCS could (1) enhance the motor adaptation to a sensory alteration, and (2) potentially affect the perceptual representation of those sounds, but we failed to demonstrate the reverse effect with the cathodal configuration. Overall, tDCS of the left IPL can be used to enhance speech performance but only under conditions in which new or adaptive learning is required. [ABSTRACT FROM AUTHOR]
- Published
- 2017
- Full Text
- View/download PDF
22. Similar abilities of musicians and non-musicians to segregate voices by fundamental frequency.
- Author
-
Deroche, Mickael L. D., Limb, Charles J., Chatterjee, Monita, and Gracco, Vincent L.
- Subjects
- *
SPEECH perception , *HEARING , *MUSICIANS , *AUDIO frequency , *PHYSIOLOGY ,SOCIAL aspects - Abstract
Musicians can sometimes achieve better speech recognition in noisy backgrounds than nonmusicians, a phenomenon referred to as the “musician advantage effect.” In addition, musicians are known to possess a finer sense of pitch than non-musicians. The present study examined the hypothesis that the latter fact could explain the former. Four experiments measured speech reception threshold for a target voice against speech or non-speech maskers. Although differences in fundamental frequency (ΔF0s) were shown to be beneficial even when presented to opposite ears (experiment 1), the authors’ attempt to maximize their use by directing the listener’s attention to the target F0 led to unexpected impairments (experiment 2) and the authors’ attempt to hinder their use by generating uncertainty about the competing Δ0s led to practically negligible effects (experiments 3 and 4). The benefits drawn from ΔF0s showed surprisingly little malleability for a cue that can be used in the complete absence of energetic masking. In half of the experiments, musicians obtained better thresholds than non-musicians, particularly in speech-on-speech conditions, but they did not reliably obtain larger DF0 benefits. Thus, the data do not support the hypothesis that the musician advantage effect is based on greater ability to exploit ΔF0s. [ABSTRACT FROM AUTHOR]
- Published
- 2017
- Full Text
- View/download PDF
23. Processing of Acoustic Cues in Lexical-Tone Identification by Pediatric Cochlear-Implant Recipients.
- Author
-
Shu-Chen Peng, Hui-Ping Lu, Lu, Nelson, Yung-Song Lin, Deroche, Mickael L. D., and Chatterjee, Monita
- Subjects
ACOUSTIC phonetics ,COCHLEAR implants ,TONE (Phonetics) ,SPEECH perception in children ,MANDARIN dialects ,PHONOLOGICAL awareness ,DEAF children -- Language ,CHINESE language ,PATIENTS ,LEXICOLOGY ,PROMPTS (Psychology) ,CHINESE people ,PEDIATRICS ,RESEARCH funding ,SOUND - Abstract
Purpose: The objective was to investigate acoustic cue processing in lexical-tone recognition by pediatric cochlear-implant (CI) recipients who are native Mandarin speakers. Method: Lexical-tone recognition was assessed in pediatric CI recipients and listeners with normal hearing (NH) in 2 tasks. In Task 1, participants identified naturally uttered words that were contrastive in lexical tones. For Task 2, a disyllabic word (yanjing) was manipulated orthogonally, varying in fundamental-frequency (F0) contours and duration patterns. Participants identified each token with the second syllable jing pronounced with Tone 1 (a high level tone) as eyes or with Tone 4 (a high falling tone) as eyeglasses. Results: CI participants' recognition accuracy was significantly lower than NH listeners' in Task 1. In Task 2, CI participants' reliance on F0 contours was significantly less than that of NH listeners; their reliance on duration patterns, however, was significantly higher than that of NH listeners. Both CI and NH listeners' performance in Task 1 was significantly correlated with their reliance on F0 contours in Task 2. Conclusion: For pediatric CI recipients, lexical-tone recognition using naturally uttered words is primarily related to their reliance on F0 contours, although duration patterns may be used as an additional cue. [ABSTRACT FROM AUTHOR]
- Published
- 2017
- Full Text
- View/download PDF
24. Deficits in the Sensitivity to Pitch Sweeps by School-Aged Children Wearing Cochlear Implants.
- Author
-
Deroche, Mickael L. D., Kulkarni, Aditya M., Christensen, Julie A., Limb, Charles J., Chatterjee, Monita, Xin Luo, and Smith, Faye
- Subjects
COCHLEAR implants ,DEAF children ,INTELLIGIBILITY of speech - Abstract
Sensitivity to static changes in pitch has been shown to be poorer in school-aged children wearing cochlear implants (CIs) than children with normal hearing (NH), but it is unclear whether this is also the case for dynamic changes in pitch. Yet, dynamically changing pitch has considerable ecological relevance in terms of natural speech, particularly aspects such as intonation, emotion, or lexical tone information. Twenty one children with NH and 23 children wearing a CI participated in this study, along with 18 NH adults and 6 CI adults for comparison. Listeners with CIs used their clinically assigned settings with envelope-based coding strategies. Percent correct was measured in one- or three-interval two-alternative forced choice tasks, for the direction or discrimination of harmonic complexes based on a linearly rising or falling fundamental frequency. Sweep rates were adjusted per subject, in a logarithmic scale, so as to cover the full extent of the psychometric function. Data for up- and down-sweeps were fitted separately, using a maximum-likelihood technique. Fits were similar for up- and down-sweeps in the discrimination task, but diverged in the direction task because psychometric functions for down-sweeps were very shallow. Hits and false alarms were then converted into d' and beta values, from which a threshold was extracted at a d' of 0.77. Thresholds were very consistent between the two tasks and considerably higher (worse) for CI listeners than for their NH peers. Thresholds were also higher for children than adults. Factors such as age at implantation, age at profound hearing loss, and duration of CI experience did not play any major role in this sensitivity. Thresholds of dynamic pitch sensitivity (in either task) also correlated with thresholds for static pitch sensitivity and with performance in tasks related to speech prosody. [ABSTRACT FROM AUTHOR]
- Published
- 2016
- Full Text
- View/download PDF
25. Phase effects in masking by harmonic complexes: Detection of bands of speech-shaped noise.
- Author
-
Deroche, Mickael L. D., Culling, John F., and Chatterjee, Monita
- Subjects
- *
NOISE , *SPEECH research , *SPEECH perception , *HUMAN voice , *ACOUSTICS research - Abstract
When phase relationships between partials of a complex masker produce highly modulated temporal envelopes on the basilar membrane, listeners may detect speech information from temporal dips in the within-channel masker envelopes. This source of masking release (MR) is however located in regions of unresolved masker partials and it is unclear how much of the speech information in these regions is really needed for intelligibility. Also, other sources of MR such as glimpsing in between resolved masker partials may provide sufficient information from regions that disregard phase relationships. This study simplified the problem of speech recognition to a masked detection task. Target bands of speech-shaped noise were restricted to frequency regions containing either only resolved or only unresolved masker partials, as a function of masker phase relationships (sine or random), masker fundamental frequency (F0) (50, 100, or 200 Hz), and masker spectral profile (flat-spectrum or speech-shaped). Although masker phase effects could be observed in unresolved regions at F0s of 50 and 100 Hz, it was only at 50-Hz F0 that detection thresholds were ever lower in unresolved than in resolved regions, suggesting little role of envelope modulations for harmonic complexes with F0s in the human voice range and at moderate level. [ABSTRACT FROM AUTHOR]
- Published
- 2014
- Full Text
- View/download PDF
26. Roles of the target and masker fundamental frequencies in voice segregation.
- Author
-
Deroche, Mickael L. D., Culling, John F., Chatterjee, Monita, and Limb, Charles J.
- Subjects
- *
HUMAN voice , *FREQUENCY curves , *AUDITORY masking , *SPEECH , *MONOTONIC functions - Abstract
Intelligibility of a target voice improves when its fundamental frequency (F0) differs from that of a masking voice, but it remains unclear how this masking release (MR) depends on the two relative F0s. Three experiments measured speech reception thresholds (SRTs) for a target voice against different maskers. Experiment 1 evaluated the influence of target F0 itself. SRTs against white noise were elevated by at least 2dB for a monotonized target voice compared with the unprocessed voice, but SRTs differed little for F0s between 50 and 150 Hz. In experiments 2 and 3, a MR occurred when there was a steady difference in F0 between the target voice and a stationary speech-shaped harmonic complex or a babble. However, this MR was considerably larger when the F0 of the masker was 11 semitones above the target F0 than when it was 11 semitones below. In contrast, for a fixed masker F0, the MR was similar whether the target F0 was above or below. The dependency of these MRs on the masker F0 suggests that a spectral mechanism such as glimpsing in between resolved masker partials may account for an important part of this phenomenon. [ABSTRACT FROM AUTHOR]
- Published
- 2014
- Full Text
- View/download PDF
27. Deficits in the pitch sensitivity of cochlear-implanted children speaking English or Mandarin.
- Author
-
Deroche, Mickael L. D., Hui-Ping Lu, Limb, Charles J., Yung-Song Lin, and Chatterjee, Monita
- Subjects
ENGLISH language ,AMPLITUDE modulation ,ARTIFICIAL implants ,HEARING aids ,COCHLEA surgery - Abstract
Sensitivity to complex pitch is notoriously poor in adults with cochlear implants (CIs), but it is unclear whether this is true for children with CIs. Many are implanted today at a very young age, and factors related to brain plasticity (age at implantation, duration of CI experience, and speaking a tonal language) might have strong influences on pitch sensitivity. School-aged children participated, speaking English or Mandarin, having normal hearing (NH) or wearing a CI, using their clinically assigned settings with envelope-based coding strategies. Percent correct was measured in three-interval three-alternative forced choice tasks, for the discrimination of fundamental frequency (F0) of broadband harmonic complexes, and for the discrimination of sinusoidal amplitude modulation rate (AMR) of broadband noise, with reference frequencies at 100 and 200 Hz to focus on voice pitch processing. Data were fitted using a maximum-likelihood technique. CI children displayed higher thresholds and shallower slopes than NH children in F0 discrimination, regardless of linguistic background. Thresholds and slopes were more similar between NH and CI children in AMR discrimination. Once the effect of chronological age was extracted from the variance, the aforementioned factors related to brain plasticity did not contribute significantly to the CI children's sensitivity to pitch. Unless different strategies attempt to encode fine structure information, potential benefits of plasticity may be missed. [ABSTRACT FROM AUTHOR]
- Published
- 2014
- Full Text
- View/download PDF
28. Speech recognition against harmonic and inharmonic complexes: Spectral dips and periodicity.
- Author
-
Deroche, Mickael L. D., Culling, John F., Chatterjee, Monita, and Limb, Charles J.
- Subjects
- *
AUTOMATIC speech recognition , *COMPUTER input-output equipment , *RADIATION , *SPECTRAL analysis (Phonetics) , *SOUND - Abstract
Speech recognition in a complex masker usually benefits from masker harmonicity, but there are several factors at work. The present study focused on two of them, glimpsing spectrally in between masker partials and periodicity within individual frequency channels. Using both a theoretical and an experimental approach, it is demonstrated that when inharmonic complexes are generated by jittering partials from their harmonic positions, there are better opportunities for spectral glimpsing in inharmonic than in harmonic maskers, and this difference is enhanced as fundamental frequency (F0) increases. As a result, measurements of masking level difference between the two maskers can be reduced, particularly at higher F0s. Using inharmonic maskers that offer similar glimpsing opportunity to harmonic maskers, it was found that the masking level difference between the two maskers varied little with F0, was influenced by periodicity of the first four partials, and could occur in low-, mid-, or high-frequency regions. Overall, the present results suggested that both spectral glimpsing and periodicity contribute to speech recognition under masking by harmonic complexes, and these effects seem independent from one another. [ABSTRACT FROM AUTHOR]
- Published
- 2014
- Full Text
- View/download PDF
29. Voice segregation by difference in fundamental frequency: Effect of masker type.
- Author
-
Deroche, Mickael L. D. and Culling, John F.
- Subjects
- *
AUDITORY masking , *AUDITORY perception , *VOICE frequency , *SPEECH audiometry , *INTELLIGIBILITY of speech , *SPEECH perception - Abstract
Speech reception thresholds were measured for a voice against two different maskers: Either two concurrent voices with the same fundamental frequency (F0) or a harmonic complex with the same long-term excitation pattern and broadband temporal envelope as the masking sentences (speech-modulated buzz). All sources had steady F0s. A difference in F0 of 2 or 8 semitones provided a 5-dB benefit for buzz maskers, whereas it provided a 3- and 8-dB benefit, respectively, for masking sentences. Whether intelligibility of a voice increases abruptly with small ΔF0s or gradually toward larger ΔF0s seems to depend on the nature of the masker. [ABSTRACT FROM AUTHOR]
- Published
- 2013
- Full Text
- View/download PDF
30. Sensitivity of school-aged children to pitch-related cues.
- Author
-
Deroche, Mickael L. D., Zion, Danielle J., Schurman, Jaclyn R., and Chatterjee, Monita
- Subjects
- *
STUDENTS , *SOUND , *HEARING , *AMPLITUDE modulation , *TONALITY - Abstract
Two experiments investigated the ability of 17 school-aged children to process purely temporal and spectro-temporal cues that signal changes in pitch. Percentage correct was measured for the discrimination of sinusoidal amplitude modulation rate (AMR) of broadband noise in experiment 1 and for the discrimination of fundamental frequency (F0) of broadband sine-phase harmonic complexes in experiment 2. The reference AMR was 100 Hz as was the reference F0. A child-friendly interface helped listeners to remain attentive to the task. Data were fitted using a maximum-likelihood technique that extracted threshold, slope, and lapse rate. All thresholds were subsequently standardized to a common d′ value equal to 0.77. There were relatively large individual differences across listeners: eight had relatively adult-like thresholds in both tasks and nine had higher thresholds. However, these individual differences did not vary systematically with age, over the span of 6-16 yr. Thresholds were correlated across the two tasks and were about nine times finer for F0 discrimination than for AMR discrimination as has been previously observed in adults. [ABSTRACT FROM AUTHOR]
- Published
- 2012
- Full Text
- View/download PDF
31. Voice segregation by difference in fundamental frequency: Evidence for harmonic cancellation.
- Author
-
Deroche, Mickael L. D. and Culling, John F.
- Abstract
Two experiments investigated listeners' ability to use a difference of two semitones in fundamental frequency (F0) to segregate a target voice from harmonic complex tones, with speech-like spectral profiles. Masker partials were in random phase (experiment 1) or in sine phase (experiment 2) and stimuli were presented over headphones. Target's and masker's harmonicity were each distorted by F0 modulation and reverberation. The F0 of each source was manipulated (monotonized or modulated by 2 semitones at 5 Hz) factorially. In addition, all sources were presented from the same location in a virtual room with controlled reverberation, assigned factorially to each source. In both experiments, speech reception thresholds increased by about 2 dB when the F0 of the masker was modulated and increased by about 6 dB when, in addition to F0 modulation, the masker was reverberant. Masker partial phases did not influence the results. The results suggest that F0-segregation relies upon the masker's harmonicity, which is disrupted by rapid modulation. This effect is compounded by reverberation. In addition, F0-segregation was found to be independent of the depth of masker envelope modulations. [ABSTRACT FROM AUTHOR]
- Published
- 2011
- Full Text
- View/download PDF
32. Neural correlates of two different types of extinction learning in the amygdala central nucleus.
- Author
-
Iordanova, Mihaela D., Deroche, Mickael L. D., Esber, Guillem R., and Schoenbaum, Geoffrey
- Published
- 2016
- Full Text
- View/download PDF
33. Visual biases in evaluation of speakers' and singers' voice type by cis and trans listeners.
- Author
-
Marchand Knight J, Sares AG, and Deroche MLD
- Abstract
Introduction: A singer's or speaker's Fach (voice type) should be appraised based on acoustic cues characterizing their voice. Instead, in practice, it is often influenced by the individual's physical appearance. This is especially distressful for transgender people who may be excluded from formal singing because of perceived mismatch between their voice and appearance. To eventually break down these visual biases, we need a better understanding of the conditions under which they occur. Specifically, we hypothesized that trans listeners (not actors) would be better able to resist such biases, relative to cis listeners, precisely because they would be more aware of appearance-voice dissociations., Methods: In an online study, 85 cisgender and 81 transgender participants were presented with 18 different actors singing or speaking short sentences. These actors covered six voice categories from high/bright (traditionally feminine) to low/dark (traditionally masculine) voices: namely soprano, mezzo-soprano (referred to henceforth as mezzo), contralto (referred to henceforth as alto), tenor, baritone, and bass. Every participant provided voice type ratings for (1) Audio-only (A) stimuli to get an unbiased estimate of a given actor's voice type, (2) Video-only (V) stimuli to get an estimate of the strength of the bias itself, and (3) combined Audio-Visual (AV) stimuli to see how much visual cues would affect the evaluation of the audio., Results: Results demonstrated that visual biases are not subtle and hold across the entire scale, shifting voice appraisal by about a third of the distance between adjacent voice types (for example, a third of the bass-to-baritone distance). This shift was 30% smaller for trans than for cis listeners, confirming our main hypothesis. This pattern was largely similar whether actors sung or spoke, though singing overall led to more feminine/high/bright ratings., Conclusion: This study is one of the first demonstrations that transgender listeners are in fact better judges of a singer's or speaker's voice type because they are better able to separate the actors' voice from their appearance, a finding that opens exciting avenues to fight more generally against implicit (or sometimes explicit) biases in voice appraisal., Competing Interests: The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest., (Copyright © 2023 Marchand Knight, Sares and Deroche.)
- Published
- 2023
- Full Text
- View/download PDF
34. Changes in Spoken and Sung Productions Following Adaptation to Pitch-shifted Auditory Feedback.
- Author
-
Alemi R, Lehmann A, and Deroche MLD
- Subjects
- Humans, Feedback, Speech physiology, Feedback, Sensory physiology, Pitch Perception physiology, Singing, Voice physiology
- Abstract
Objective: Using voice to speak or to sing is made possible by remarkably complex sensorimotor processes. Like any other sensorimotor system, the speech motor controller guides its actions with maximum performance at minimum cost, using available sources of information, among which, auditory feedback plays a major role. Manipulation of this feedback forces the speech monitoring system to refine its expectations for further actions. The present study hypothesizes that the duration of this refinement and the weight applied on different feedbacks loops would depend on the intended sounds to be produced, namely reading aloud versus singing., Material and Methods: We asked participants to sing "Happy Birthday" and read a paragraph of Harry Potter before and after experiencing pitch-shifted feedback. A detailed fundamental frequency (F0) analysis was conducted for each note in the song and each segment in the paragraph (at the level of a sentence, a word, or a vowel) to determine whether some aspects of F0 production changed in response to the pitch perturbations experienced during the adaptation paradigm., Results: Our results showed that changes in the degree of F0-drift across the song or the paragraph was the metric that was the most consistent with a carry-over effect of adaptation, and in this regard, reading new material was more influenced by recent remapping than singing., Conclusion: The motor commands used by (normally-hearing) speakers are malleable via altered-feedback paradigms, perhaps more so when reading aloud than when singing. But these effects are not revealed through simple indicators such as an overall change in mean F0 or F0 range, but rather through subtle metrics, such as a drift of the voice pitch across the recordings., Competing Interests: COMPETING INTERESTS The authors declare no competing interests. Funding came in part from a cochlear implant manufacturer, but this study does not report on any data collected with cochlear implant users., (Crown Copyright © 2021. Published by Elsevier Inc. All rights reserved.)
- Published
- 2023
- Full Text
- View/download PDF
35. Auditory evoked response to an oddball paradigm in children wearing cochlear implants.
- Author
-
Deroche MLD, Wolfe J, Neumann S, Manning J, Towler W, Alemi R, Bien AG, Koirala N, Hanna L, Henry L, and Gracco VL
- Subjects
- Child, Humans, Acoustic Stimulation, Evoked Potentials, Auditory physiology, Auditory Perception physiology, Electroencephalography, Cochlear Implants, Cochlear Implantation
- Abstract
Objective: Although children with cochlear implants (CI) achieve remarkable success with their device, considerable variability remains in individual outcomes. Here, we explored whether auditory evoked potentials recorded during an oddball paradigm could provide useful markers of auditory processing in this pediatric population., Methods: High-density electroencephalography (EEG) was recorded in 75 children listening to standard and odd noise stimuli: 25 had normal hearing (NH) and 50 wore a CI, divided between high language (HL) and low language (LL) abilities. Three metrics were extracted: the first negative and second positive components of the standard waveform (N1-P2 complex) close to the vertex, the mismatch negativity (MMN) around Fz and the late positive component (P3) around Pz of the difference waveform., Results: While children with CIs generally exhibited a well-formed N1-P2 complex, those with language delays typically lacked reliable MMN and P3 components. But many children with CIs with age-appropriate skills showed MMN and P3 responses similar to those of NH children. Moreover, larger and earlier P3 (but not MMN) was linked to better literacy skills., Conclusions: Auditory evoked responses differentiated children with CIs based on their good or poor skills with language and literacy., Significance: This short paradigm could eventually serve as a clinical tool for tracking the developmental outcomes of implanted children., Competing Interests: Declaration of conflicting interests The authors have declared no competing interest. MD has received research funding from industrial partners Oticon and Med-El but for unrelated projects. JW is a member of the Audiology Advisory Boards of Advanced Bionics and Cochlear, the manufacturers of the cochlear implants used by the participants in this study, but no funding from them was received for this study., (Copyright © 2023 International Federation of Clinical Neurophysiology. Published by Elsevier B.V. All rights reserved.)
- Published
- 2023
- Full Text
- View/download PDF
36. Effect of Frequency Response Manipulations on Musical Sound Quality for Cochlear Implant Users.
- Author
-
Mo J, Jiam NT, Deroche MLD, Jiradejvong P, and Limb CJ
- Subjects
- Adult, Auditory Perception physiology, Humans, Sound, Cochlear Implantation, Cochlear Implants, Music
- Abstract
Cochlear implant (CI) users commonly report degraded musical sound quality. To improve CI-mediated music perception and enjoyment, we must understand factors that affect sound quality. In the present study, we utilize frequency response manipulation (FRM), a process that adjusts the energies of frequency bands within an audio signal, to determine its impact on CI-user sound quality assessments of musical stimuli. Thirty-three adult CI users completed an online study and listened to FRM-altered clips derived from the top songs in Billboard magazine. Participants assessed sound quality using the MUltiple Stimulus with Hidden Reference and Anchor for CI users (CI-MUSHRA) rating scale. FRM affected sound quality ratings (SQR). Specifically, increasing the gain for low and mid-range frequencies led to higher quality ratings than reducing them. In contrast, manipulating the gain for high frequencies (those above 2 kHz) had no impact. Participants with musical training were more sensitive to FRM than non-musically trained participants and demonstrated preference for gain increases over reductions. These findings suggest that, even among CI users, past musical training provides listeners with subtleties in musical appraisal, even though their hearing is now mediated electrically and bears little resemblance to their musical experience prior to implantation. Increased gain below 2 kHz may lead to higher sound quality than for equivalent reductions, perhaps because it offers greater access to lyrics in songs or because it provides more salient beat sensations.
- Published
- 2022
- Full Text
- View/download PDF
37. The intelligibility of speech in a harmonic masker varying in fundamental frequency contour, broadband temporal envelope, and spatial location.
- Author
-
Leclère T, Lavandier M, and Deroche MLD
- Subjects
- Acoustic Stimulation, Adolescent, Adult, Cues, Humans, Speech Reception Threshold Test, Time Factors, Young Adult, Noise adverse effects, Perceptual Masking, Pitch Discrimination, Sound Localization, Speech Intelligibility, Speech Perception
- Abstract
Differences in fundamental frequency (F0), modulations in the masker envelope, and differences in spatial location between a speech target and a masker can improve speech intelligibility in cocktail-party situations. These cues have been thoroughly investigated independently and associated with unmasking mechanisms: F0 segregation, temporal dip listening and spatial unmasking, respectively. Two experiments were conducted to examine whether F0 segregation interacts with spatial unmasking (experiment 1) or temporal modulations in the masker envelope (experiment 2) by measuring speech reception thresholds for a monotonized or an intonated voice against eight types of harmonic complex masker. In experiment 1, the masker varied in F0 contour (monotonized or intonated), mean F0 (0 or 3 semitones above that of the target) and spatial location (co-located or separated from the target). In experiment 2, the masker varied in F0 contour, mean F0 and broadband temporal envelope (stationary or 1-voice modulated). The benefits associated with spatial separation and F0 differences added up linearly in almost all conditions, whereas modulations in the masker envelope improved speech intelligibility only in the presence of intonated maskers. In addition, in both experiments F0 segregation benefited considerably from natural variations in the F0 pattern of the target voice, but was largely disrupted by those of the masker., (Copyright © 2017. Published by Elsevier B.V.)
- Published
- 2017
- Full Text
- View/download PDF
38. Processing of Acoustic Cues in Lexical-Tone Identification by Pediatric Cochlear-Implant Recipients.
- Author
-
Peng SC, Lu HP, Lu N, Lin YS, Deroche MLD, and Chatterjee M
- Subjects
- Acoustic Stimulation methods, Adolescent, Child, Female, Humans, Male, Sound Spectrography, Young Adult, Cochlear Implantation, Deafness rehabilitation, Pattern Recognition, Physiological, Speech Perception
- Abstract
Purpose: The objective was to investigate acoustic cue processing in lexical-tone recognition by pediatric cochlear-implant (CI) recipients who are native Mandarin speakers., Method: Lexical-tone recognition was assessed in pediatric CI recipients and listeners with normal hearing (NH) in 2 tasks. In Task 1, participants identified naturally uttered words that were contrastive in lexical tones. For Task 2, a disyllabic word (yanjing) was manipulated orthogonally, varying in fundamental-frequency (F0) contours and duration patterns. Participants identified each token with the second syllable jing pronounced with Tone 1 (a high level tone) as eyes or with Tone 4 (a high falling tone) as eyeglasses., Results: CI participants' recognition accuracy was significantly lower than NH listeners' in Task 1. In Task 2, CI participants' reliance on F0 contours was significantly less than that of NH listeners; their reliance on duration patterns, however, was significantly higher than that of NH listeners. Both CI and NH listeners' performance in Task 1 was significantly correlated with their reliance on F0 contours in Task 2., Conclusion: For pediatric CI recipients, lexical-tone recognition using naturally uttered words is primarily related to their reliance on F0 contours, although duration patterns may be used as an additional cue.
- Published
- 2017
- Full Text
- View/download PDF
39. Phase effects in masking by harmonic complexes: speech recognition.
- Author
-
Deroche ML, Culling JF, and Chatterjee M
- Subjects
- Acoustic Stimulation, Adult, Auditory Perception, Auditory Threshold, Basilar Membrane physiology, Female, Hearing, Humans, Male, Normal Distribution, Signal Processing, Computer-Assisted, Software, Speech, Time Factors, Voice, Young Adult, Perceptual Masking physiology, Speech Perception physiology
- Abstract
Harmonic complexes that generate highly modulated temporal envelopes on the basilar membrane (BM) mask a tone less effectively than complexes that generate relatively flat temporal envelopes, because the non-linear active gain of the BM selectively amplifies a low-level tone in the dips of a modulated masker envelope. The present study examines a similar effect in speech recognition. Speech reception thresholds (SRTs) were measured for a voice masked by harmonic complexes with partials in sine phase (SP) or in random phase (RP). The masker's fundamental frequency (F0) was 50, 100 or 200 Hz. SRTs were considerably lower for SP than for RP maskers at 50-Hz F0, but the two converged at 100-Hz F0, while at 200-Hz F0, SRTs were a little higher for SP than RP maskers. The results were similar whether the target voice was male or female and whether the masker's spectral profile was flat or speech-shaped. Although listening in the masker dips has been shown to play a large role for artificial stimuli such as Schroeder-phase complexes at high levels, it contributes weakly to speech recognition in the presence of harmonic maskers with different crest factors at more moderate sound levels (65 dB SPL)., (Copyright © 2013 Elsevier B.V. All rights reserved.)
- Published
- 2013
- Full Text
- View/download PDF
40. Narrow noise band detection in a complex masker: masking level difference due to harmonicity.
- Author
-
Deroche ML and Culling JF
- Subjects
- Acoustic Stimulation, Adult, Analysis of Variance, Audiometry, Auditory Threshold, Humans, Psychoacoustics, Sound Spectrography, Time Factors, Young Adult, Auditory Perception, Noise adverse effects, Perceptual Masking, Signal Detection, Psychological
- Abstract
Three experiments investigated listeners' ability to detect a narrow band of noise, centered on one partial of a random-phase complex tone, as a function of inharmonicity. Inharmonicity was generated by randomly mistuning the partial frequencies from a 100-Hz fundamental frequency (F0). In experiment 1, masked detection thresholds were lower when the masker was harmonic than when it was inharmonic for target bands in the range 0.5-2.5 kHz. The presence of this masking level difference due to harmonicity (HMLD) in regions of resolved partials and the reduction of the HMLD with increasing center frequency did not support the idea that HMLD was primarily caused by the envelope modulations produced by the beating of unresolved partials within an auditory filter. In experiment 2, masker mistunings ranging beyond 12% of the F0 disrupted the HMLD while smaller mistunings gave thresholds similar to a harmonic masker. In experiment 3, all partials contributed to some extent to the HMLD, but the harmonicity of partials neighboring the target had a greater influence than distant partials. The observed HMLDs can best be accounted for by a mechanism of harmonic cancellation., (Copyright © 2011 Elsevier B.V. All rights reserved.)
- Published
- 2011
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.