HCS 7367 Speech Perception
|
|
- Adrian Kelly
- 5 years ago
- Views:
Transcription
1 Babies 'cry in mother's tongue' HCS 7367 Speech Perception Dr. Peter Assmann Fall 212 Babies' cries imitate their mother tongue as early as three days old German researchers say babies begin to pick up the nuances of their parents' accents while still in the womb. Long-term spectrum of speech Speech communication in adverse listening conditions Connected speech Males Absolute threshold Females Long-term spectrum of speech Vowels Males Females 2) Absolute threshold Masking and interference Energetic masking reduced audibility of signal components due to overlap in spectral energy within the same auditory channel. Informational masking reduced audibility of signal components due to non-energetic factors such as target-masker similarity. Forward vs. backward speech maskers Familiar vs. foreign language 1
2 Resistance to distortion Effects of noise Articulation score: % items correct on spoken lists of syllables, words or sentences Signal-to-noise ratio (SNR): when speech and noise have the same average rms level (SNR= db), articulation scores are above 5% for listeners with normal hearing Signal-to-noise ratio (SNR) SNR = 2 log 1 [ rms(speech) / rms(noise) ] Specified in decibels When speech and noise have the same average rms level (SNR= db), articulation scores are above 5% for listeners with normal hearing. Signal-to-noise ratio (SNR) SNR = 2 log 1 [ rms(speech) / rms(noise) ] Why is speech intelligible when the masker is presented at the same level as the speech? Articulation Index How much does audibility contribute to difficulty understanding speech in noise? Articulation Index (AI) estimates the contribution of audibility (and other factors) to speech intelligibility Articulation Index 1. Divides the speech and masker spectrum into a small number of frequency bands 2. Estimates the audibility of speech in each band, weighted by its relative importance for intelligibility 3. Derives overall intelligibility by summing the contributions of each band. 2
3 Speech recognition in noise White Noise Spectral properties of the noise: white, pink, speech-shaped, competing speech, speech babble Temporal properties of the noise: steady vs. modulated or interrupted Amplitude Waveform Amplitude (db) Spectrum Effects of noise on speech recognition Non-uniform noise Pink Noise Speech-shaped Noise Amplitude (db) 2 1 Amplitude (db) Source: Miller, Heise and Lichten, J. Exp. Psychol Non-uniform noise Multi-talker babble Speech babble (mixture of 1 sentences from 1 talker) Amplitude (db) Effect of increasing the number of competing voices 1 sentence 2 sentences 4 sentences 8 sentences 16 sentences 3
4 Effects of noise on vowel spectra Vowel / / in quiet and in noise Broadband noise tends to fill up the valleys between the formant peaks. Spectral contrast (peak-to-valley ratio) is reduced by the addition of noise. Because of the sloping long-term spectrum of speech, the upper formants (F3, F4, F5) are more susceptible to masking and distortion by the noise. Amplitude (db) Excitation (db) In quiet Pink noise, +6 db SNR Effects of noise on formant peaks Effects of filtering Effects of filtering on speech High-pass and low-pass filtering 1 Low-pass filtering to remove frequencies above 18 Hz reduces intelligibility from near perfect to around 67%. High-pass filtering to remove components below 18 Hz also produces about 67%. Identification accuracy (%) HP LP Frequency (Hz) 4
5 Bandpass filtering Bandpass filtering with one-third octave filters centered Hz produces better than 95% accuracy for highpredictability sentences (Warren et al., 1995; Stickney & Assmann, 21). Speech communication has an extraordinary resilience to distortion. 1. Intelligibility remains high even when large portions of the spectrum are eliminated by filtering. Stickney and Assmann (JASA 21) Other frequency distortions Notch filtering to remove frequencies between 8 and 3 Hz leads to consonant identification scores better than 9% (Lippman, 1996) Conclusion: speech cues are widely distributed Perception of filtered speech Everyday English sentences filtered using narrow bandpass filters remain highly intelligible (>9% words correct) one-third octave bandwidth, 15 Hz center frequency, 1 db/octave slopes Warren et al. (Percept Psychophys 1995; JASA 2) Perception of filtered speech Speech communication has an extraordinary resiliance to distortion. 2. Large segments of the waveform can be deleted or replaced by silence. Interruption rate = 5 Hz 1 second Stickney and Assmann (JASA 21) 5
6 Speech communication has an extraordinary resiliance to distortion. 3. Noise can be added to the speech signal at equal intensity (Signal-to-noise ratio = db). Speech communication has an extraordinary resiliance to distortion. 3. Noise can be added to the speech signal at equal intensity (Signal-to-noise ratio = db). + Speech Speech-shaped noise Speech communication has an extraordinary resiliance to distortion. 4. When the noise is from a competing voice, target and masker are similar and must be segregated How do listeners achieve this? Statistical redundancy of speech/language Combined strategies of top-down + bottom-up processing Grouping and segregation of auditory objects Tracking speech properties over time Glimpsing speech fragments during noise-free intervals Redundancy in speech and language Coker and Umeda (1974) define redundancy as: any characteristic of the language that forces spoken messages to have, on average, more basic elements per message, or more cues per basic element, than the barest minimum [necessary for conveying the linguistic message]. Redundancy in error correction Redundancy can be used effectively; or it can be squandered on uneven repetition of certain data, leaving other crucial items very vulnerable to noise.... But more likely, if a redundancy is a property of a language and has to be learned, then it has a purpose. Coker and Umeda (1974, p. 349) 6
7 Redundancy contributes speech perception in several ways 1. by limiting perceptual confusion due to errors in speech production; 2. by helping to bridge gaps in the signal created by interfering noise, reverberation, and distortions of the communication channel; and 3. by compensating for momentary lapses in attention and misperceptions on the part of the listener. Effects of context Contextual cues lead to improved speech understanding in noise. Acoustic-phonetic context Prosodic context Semantic context Syntax Miller, Heise & Lichten, 1951 Miller, Heise & Lichten, PERCE NT ITEMS CORRECT Recognition of interrupted speech in quiet Interrupted speech In this condition the speech is turned on and off at regular intervals using an electronic switch. Miller and Licklider JASA 195 7
8 Interrupted speech Miller and Licklider JASA 195 Word Identification Accuracy (%) Interrupted speech Frequency of ofinterruptions (s) (s) Interrupted speech In quiet, speech can be interrupted (turned on and off) periodically without substantial loss of intelligibility (Miller & Licklider, 195). Miller and Licklider found the worst intelligibility for interruption rates < 2 Hz, where large speech fragments (words, phrases) are missing. Interrupted speech Masking of speech by interrupted noise Miller and Licklider found improved performance for interruption rates between 1 and 1 Hz. Why? For very high interruption rates (>1 khz) the signal sounded continuous, and performance was near perfect. Miller and Licklider also measured speech intelligibility in conditions where the speech was continuous but the noise was interrupted. 16 Hz 128 Hz 512 Hz 8
9 Masking of speech by interrupted noise When the noise is intermittent rather than continuous there is a release from masking. The benefits of non-stationarity depend on the interruption rate and the duty cycle (onoff ratio) of the noise. Masking of speech by interrupted noise Interrupted noise At low interruption rates the effects are similar to speech interrupted by silence. As the interruption rate increases there is a gradual improvement in speech recognition. With 1 interruptions per second, listeners receive several glimpses of each word and can patch together those glimpses to recognize about 75% of the words correctly Interrupted noise When a noise masker is alternated with silence using a 5% duty cycle, there may be considerable masking release, compared to a continuous masker, especially with alternation rates between 1 and 2 per second (Miller and Licklider, 195). Summary: Interrupted noise 1. At alternation rates between about 1 and 2 per second, listeners can patch together cues from the clean segments between the bursts of noise. 2. With slower interruption rates, entire words or phrases are masked; others are noise-free. 3. At rates > 2/sec the masking effect is the same as uninterrupted, continuous noise. 9
10 Picket-fence effect Picket-fence effect Interrupted speech can have a harsh, distorted quality. But when speech and noise are alternated periodically, filling silent gaps with noise, the speech sounded smooth and continuous. Possibly, noise in the gaps enhances the listener s ability to exploit contextual cues. Howard-Jones and Rosen (1993) Checkerboard noise maskers Effects of interruption rate and frequency bandwidth of the checkerboard pattern Howard-Jones and Rosen (1993) Can listeners exploit asynchronous timefrequency glimpses? Yes, but only over broad frequency ranges Frequency Frequency Time Time A glimpsing model of speech perception in noise Martin Cooke Journal of the Acoustical Society of America, Vol. 119, No. 3, pp , March 26 Speech source separation How do the ear and brain separate the target voice from the noise? spatial cues lip-reading semantic context auditory scene analysis (Bregman, 199) glimpsing and tracking 1
11 Auditory scene analysis Bregman (199) The sound that reaches the eardrum of the listener is often a mixture of different sources Acoustic signals originating from different sound sources combine additively Unlike vision, the concept of occlusion is hard to define in audition: sounds overlap but also combine in complex ways. Auditory scene analysis Computational auditory scene analysis Reviewed by Cooke and Ellis (21) Human listeners are good at separating mixtures of sounds, as reflected in speech communication and listening to music in complex listening environments (cocktail parties) Attempts to reproduce this separation process using computational models had limited success (a hard problem!) Glimpsing speech in noise speech is a highly modulated signal in time and frequency, regions of high energy are typically sparsely distributed. Glimpsing speech in noise The information conveyed by the spectrotemporal energy spectrum of clean speech is redundant Redundancy allows speech to be identified based on relatively sparse evidence Frequency (Hz) 1..5 Frequency (Hz) Glimpsing speech in noise Can listeners take advantage of glimpses? direct attention to spectrotemporal regions where the S+N mixture is dominated by the target speech ASR system trained to recognize consonants in noise Maskers differed in glimpse size ASR model developed to exploit non-uniform distribution of SNR in different time-frequency bands Conclusion: model + listeners benefit from glimpsing. Speech + noise mixtures Some regions dominated by target voice Local SNR varies across time and frequency Where the target voice dominates, the problem of source segregation is solved because the signal is effectively clean speech. Clean speech is highly redundant; it remains intelligible after 5% or more of its energy is removed by gating and/or filtering 11
12 STEP model Auditory excitation pattern (Moore, 23) Spectrogram-like representation Reflects non-uniform frequency selectivity in different frequency bands Incorporates a sliding time window reflecting temporal analysis by the auditory system Relative audibility at different frequencies Loudness model Missing data ASR HMM-based speech recognizer Missing-data models Glimpses only Ignore missing information (in masked regions) Glimpses-plus-background Try to fill in missing information (based on masked regions) Sparseness and redundancy Glimpses = spectrotemporal regions where signal exceeds masker by ~3 db. single talker masker target eight-talker masker speech-shaped noise Syllable identification accuracy as a function of the number of competing voices. The level of the target speech (monosyllabic nonsense words) was held constant at 95dB. (After Miller 1947). glimpses Results 12
13 Results Conclusions Best model: Uses information in glimpses and counterevidence in the masked regions Glimpses constrained to a minimum area Treats all regions with local SNR > -5 db as potential glimpses FIG. 4. The correlation between intelligibility and proportion of the target speech in which the local SNR exceeds 3 db. Each point represents a noise condition, and proportions are means across all tokens in the test set. The best linear fit is also shown. The correlation between listeners and these putative glimpses is.955. Conclusions A higher glimpse threshold (e.g. local SNR > db) produces fewer glimpses, but this provides less distorted information than a lower threshold (e.g. -5 db). Conclusions Limitation: local SNR must be known in advance. Is there a way to estimate the local SNR directly from the mixture? Tracking problem: how to integrate glimpses over time? Brungart et al. (21) Brungart et al. (21) 2-talker correct responses (%) Different Modulated talker, talker, Same different talker same noise sex Target-to-Masker Ratio (db) 2-talker correct responses (%) Target-to-Masker Ratio (db) 13
HCS 7367 Speech Perception
Long-term spectrum of speech HCS 7367 Speech Perception Connected speech Absolute threshold Males Dr. Peter Assmann Fall 212 Females Long-term spectrum of speech Vowels Males Females 2) Absolute threshold
More informationAUDL GS08/GAV1 Signals, systems, acoustics and the ear. Pitch & Binaural listening
AUDL GS08/GAV1 Signals, systems, acoustics and the ear Pitch & Binaural listening Review 25 20 15 10 5 0-5 100 1000 10000 25 20 15 10 5 0-5 100 1000 10000 Part I: Auditory frequency selectivity Tuning
More informationComputational Perception /785. Auditory Scene Analysis
Computational Perception 15-485/785 Auditory Scene Analysis A framework for auditory scene analysis Auditory scene analysis involves low and high level cues Low level acoustic cues are often result in
More informationPrelude Envelope and temporal fine. What's all the fuss? Modulating a wave. Decomposing waveforms. The psychophysics of cochlear
The psychophysics of cochlear implants Stuart Rosen Professor of Speech and Hearing Science Speech, Hearing and Phonetic Sciences Division of Psychology & Language Sciences Prelude Envelope and temporal
More informationRole of F0 differences in source segregation
Role of F0 differences in source segregation Andrew J. Oxenham Research Laboratory of Electronics, MIT and Harvard-MIT Speech and Hearing Bioscience and Technology Program Rationale Many aspects of segregation
More informationSPEECH PERCEPTION IN A 3-D WORLD
SPEECH PERCEPTION IN A 3-D WORLD A line on an audiogram is far from answering the question How well can this child hear speech? In this section a variety of ways will be presented to further the teacher/therapist
More informationThe role of periodicity in the perception of masked speech with simulated and real cochlear implants
The role of periodicity in the perception of masked speech with simulated and real cochlear implants Kurt Steinmetzger and Stuart Rosen UCL Speech, Hearing and Phonetic Sciences Heidelberg, 09. November
More informationLinguistic Phonetics Fall 2005
MIT OpenCourseWare http://ocw.mit.edu 24.963 Linguistic Phonetics Fall 2005 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. 24.963 Linguistic Phonetics
More informationLinguistic Phonetics. Basic Audition. Diagram of the inner ear removed due to copyright restrictions.
24.963 Linguistic Phonetics Basic Audition Diagram of the inner ear removed due to copyright restrictions. 1 Reading: Keating 1985 24.963 also read Flemming 2001 Assignment 1 - basic acoustics. Due 9/22.
More informationAcoustics, signals & systems for audiology. Psychoacoustics of hearing impairment
Acoustics, signals & systems for audiology Psychoacoustics of hearing impairment Three main types of hearing impairment Conductive Sound is not properly transmitted from the outer to the inner ear Sensorineural
More informationInfant Hearing Development: Translating Research Findings into Clinical Practice. Auditory Development. Overview
Infant Hearing Development: Translating Research Findings into Clinical Practice Lori J. Leibold Department of Allied Health Sciences The University of North Carolina at Chapel Hill Auditory Development
More informationProviding Effective Communication Access
Providing Effective Communication Access 2 nd International Hearing Loop Conference June 19 th, 2011 Matthew H. Bakke, Ph.D., CCC A Gallaudet University Outline of the Presentation Factors Affecting Communication
More informationEffects of Cochlear Hearing Loss on the Benefits of Ideal Binary Masking
INTERSPEECH 2016 September 8 12, 2016, San Francisco, USA Effects of Cochlear Hearing Loss on the Benefits of Ideal Binary Masking Vahid Montazeri, Shaikat Hossain, Peter F. Assmann University of Texas
More informationUvA-DARE (Digital Academic Repository) Perceptual evaluation of noise reduction in hearing aids Brons, I. Link to publication
UvA-DARE (Digital Academic Repository) Perceptual evaluation of noise reduction in hearing aids Brons, I. Link to publication Citation for published version (APA): Brons, I. (2013). Perceptual evaluation
More informationTopics in Linguistic Theory: Laboratory Phonology Spring 2007
MIT OpenCourseWare http://ocw.mit.edu 24.91 Topics in Linguistic Theory: Laboratory Phonology Spring 27 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationACOUSTIC AND PERCEPTUAL PROPERTIES OF ENGLISH FRICATIVES
ISCA Archive ACOUSTIC AND PERCEPTUAL PROPERTIES OF ENGLISH FRICATIVES Allard Jongman 1, Yue Wang 2, and Joan Sereno 1 1 Linguistics Department, University of Kansas, Lawrence, KS 66045 U.S.A. 2 Department
More informationAsynchronous glimpsing of speech: Spread of masking and task set-size
Asynchronous glimpsing of speech: Spread of masking and task set-size Erol J. Ozmeral, a) Emily Buss, and Joseph W. Hall III Department of Otolaryngology/Head and Neck Surgery, University of North Carolina
More informationAuditory gist perception and attention
Auditory gist perception and attention Sue Harding Speech and Hearing Research Group University of Sheffield POP Perception On Purpose Since the Sheffield POP meeting: Paper: Auditory gist perception:
More informationCHAPTER 1 INTRODUCTION
CHAPTER 1 INTRODUCTION 1.1 BACKGROUND Speech is the most natural form of human communication. Speech has also become an important means of human-machine interaction and the advancement in technology has
More information9/29/14. Amanda M. Lauer, Dept. of Otolaryngology- HNS. From Signal Detection Theory and Psychophysics, Green & Swets (1966)
Amanda M. Lauer, Dept. of Otolaryngology- HNS From Signal Detection Theory and Psychophysics, Green & Swets (1966) SIGNAL D sensitivity index d =Z hit - Z fa Present Absent RESPONSE Yes HIT FALSE ALARM
More informationJuan Carlos Tejero-Calado 1, Janet C. Rutledge 2, and Peggy B. Nelson 3
PRESERVING SPECTRAL CONTRAST IN AMPLITUDE COMPRESSION FOR HEARING AIDS Juan Carlos Tejero-Calado 1, Janet C. Rutledge 2, and Peggy B. Nelson 3 1 University of Malaga, Campus de Teatinos-Complejo Tecnol
More informationBinaural processing of complex stimuli
Binaural processing of complex stimuli Outline for today Binaural detection experiments and models Speech as an important waveform Experiments on understanding speech in complex environments (Cocktail
More informationAssessing Hearing and Speech Recognition
Assessing Hearing and Speech Recognition Audiological Rehabilitation Quick Review Audiogram Types of hearing loss hearing loss hearing loss Testing Air conduction Bone conduction Familiar Sounds Audiogram
More informationUSING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES
USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES Varinthira Duangudom and David V Anderson School of Electrical and Computer Engineering, Georgia Institute of Technology Atlanta, GA 30332
More information11 Music and Speech Perception
11 Music and Speech Perception Properties of sound Sound has three basic dimensions: Frequency (pitch) Intensity (loudness) Time (length) Properties of sound The frequency of a sound wave, measured in
More informationBest Practice Protocols
Best Practice Protocols SoundRecover for children What is SoundRecover? SoundRecover (non-linear frequency compression) seeks to give greater audibility of high-frequency everyday sounds by compressing
More informationThe effect of wearing conventional and level-dependent hearing protectors on speech production in noise and quiet
The effect of wearing conventional and level-dependent hearing protectors on speech production in noise and quiet Ghazaleh Vaziri Christian Giguère Hilmi R. Dajani Nicolas Ellaham Annual National Hearing
More informationIssues faced by people with a Sensorineural Hearing Loss
Issues faced by people with a Sensorineural Hearing Loss Issues faced by people with a Sensorineural Hearing Loss 1. Decreased Audibility 2. Decreased Dynamic Range 3. Decreased Frequency Resolution 4.
More informationPsychoacoustical Models WS 2016/17
Psychoacoustical Models WS 2016/17 related lectures: Applied and Virtual Acoustics (Winter Term) Advanced Psychoacoustics (Summer Term) Sound Perception 2 Frequency and Level Range of Human Hearing Source:
More informationTHE MECHANICS OF HEARING
CONTENTS The mechanics of hearing Hearing loss and the Noise at Work Regulations Loudness and the A weighting network Octave band analysis Hearing protection calculations Worked examples and self assessed
More informationSlow compression for people with severe to profound hearing loss
Phonak Insight February 2018 Slow compression for people with severe to profound hearing loss For people with severe to profound hearing loss, poor auditory resolution abilities can make the spectral and
More informationWhat you re in for. Who are cochlear implants for? The bottom line. Speech processing schemes for
What you re in for Speech processing schemes for cochlear implants Stuart Rosen Professor of Speech and Hearing Science Speech, Hearing and Phonetic Sciences Division of Psychology & Language Sciences
More informationSpeech perception in individuals with dementia of the Alzheimer s type (DAT) Mitchell S. Sommers Department of Psychology Washington University
Speech perception in individuals with dementia of the Alzheimer s type (DAT) Mitchell S. Sommers Department of Psychology Washington University Overview Goals of studying speech perception in individuals
More informationSound localization psychophysics
Sound localization psychophysics Eric Young A good reference: B.C.J. Moore An Introduction to the Psychology of Hearing Chapter 7, Space Perception. Elsevier, Amsterdam, pp. 233-267 (2004). Sound localization:
More informationRobust Neural Encoding of Speech in Human Auditory Cortex
Robust Neural Encoding of Speech in Human Auditory Cortex Nai Ding, Jonathan Z. Simon Electrical Engineering / Biology University of Maryland, College Park Auditory Processing in Natural Scenes How is
More informationInfluence of acoustic complexity on spatial release from masking and lateralization
Influence of acoustic complexity on spatial release from masking and lateralization Gusztáv Lőcsei, Sébastien Santurette, Torsten Dau, Ewen N. MacDonald Hearing Systems Group, Department of Electrical
More informationIsolating the energetic component of speech-on-speech masking with ideal time-frequency segregation
Isolating the energetic component of speech-on-speech masking with ideal time-frequency segregation Douglas S. Brungart a Air Force Research Laboratory, Human Effectiveness Directorate, 2610 Seventh Street,
More informationTemporal offset judgments for concurrent vowels by young, middle-aged, and older adults
Temporal offset judgments for concurrent vowels by young, middle-aged, and older adults Daniel Fogerty Department of Communication Sciences and Disorders, University of South Carolina, Columbia, South
More informationEssential feature. Who are cochlear implants for? People with little or no hearing. substitute for faulty or missing inner hair
Who are cochlear implants for? Essential feature People with little or no hearing and little conductive component to the loss who receive little or no benefit from a hearing aid. Implants seem to work
More informationAuditory-Visual Speech Perception Laboratory
Auditory-Visual Speech Perception Laboratory Research Focus: Identify perceptual processes involved in auditory-visual speech perception Determine the abilities of individual patients to carry out these
More informationHearing. and other senses
Hearing and other senses Sound Sound: sensed variations in air pressure Frequency: number of peaks that pass a point per second (Hz) Pitch 2 Some Sound and Hearing Links Useful (and moderately entertaining)
More informationHearing Lectures. Acoustics of Speech and Hearing. Auditory Lighthouse. Facts about Timbre. Analysis of Complex Sounds
Hearing Lectures Acoustics of Speech and Hearing Week 2-10 Hearing 3: Auditory Filtering 1. Loudness of sinusoids mainly (see Web tutorial for more) 2. Pitch of sinusoids mainly (see Web tutorial for more)
More informationHearing in the Environment
10 Hearing in the Environment Click Chapter to edit 10 Master Hearing title in the style Environment Sound Localization Complex Sounds Auditory Scene Analysis Continuity and Restoration Effects Auditory
More informationSpeech intelligibility in background noise with ideal binary time-frequency masking
Speech intelligibility in background noise with ideal binary time-frequency masking DeLiang Wang a Department of Computer Science & Engineering and Center for Cognitive Science, The Ohio State University,
More informationEffects of noise and filtering on the intelligibility of speech produced during simultaneous communication
Journal of Communication Disorders 37 (2004) 505 515 Effects of noise and filtering on the intelligibility of speech produced during simultaneous communication Douglas J. MacKenzie a,*, Nicholas Schiavetti
More informationWho are cochlear implants for?
Who are cochlear implants for? People with little or no hearing and little conductive component to the loss who receive little or no benefit from a hearing aid. Implants seem to work best in adults who
More informationBINAURAL DICHOTIC PRESENTATION FOR MODERATE BILATERAL SENSORINEURAL HEARING-IMPAIRED
International Conference on Systemics, Cybernetics and Informatics, February 12 15, 2004 BINAURAL DICHOTIC PRESENTATION FOR MODERATE BILATERAL SENSORINEURAL HEARING-IMPAIRED Alice N. Cheeran Biomedical
More informationSpeech Intelligibility Measurements in Auditorium
Vol. 118 (2010) ACTA PHYSICA POLONICA A No. 1 Acoustic and Biomedical Engineering Speech Intelligibility Measurements in Auditorium K. Leo Faculty of Physics and Applied Mathematics, Technical University
More informationEnrique A. Lopez-Poveda Alan R. Palmer Ray Meddis Editors. The Neurophysiological Bases of Auditory Perception
Enrique A. Lopez-Poveda Alan R. Palmer Ray Meddis Editors The Neurophysiological Bases of Auditory Perception 123 The Neurophysiological Bases of Auditory Perception Enrique A. Lopez-Poveda Alan R. Palmer
More informationHearing. Juan P Bello
Hearing Juan P Bello The human ear The human ear Outer Ear The human ear Middle Ear The human ear Inner Ear The cochlea (1) It separates sound into its various components If uncoiled it becomes a tapering
More informationAuditory scene analysis in humans: Implications for computational implementations.
Auditory scene analysis in humans: Implications for computational implementations. Albert S. Bregman McGill University Introduction. The scene analysis problem. Two dimensions of grouping. Recognition
More informationFrequency refers to how often something happens. Period refers to the time it takes something to happen.
Lecture 2 Properties of Waves Frequency and period are distinctly different, yet related, quantities. Frequency refers to how often something happens. Period refers to the time it takes something to happen.
More informationSpeech (Sound) Processing
7 Speech (Sound) Processing Acoustic Human communication is achieved when thought is transformed through language into speech. The sounds of speech are initiated by activity in the central nervous system,
More informationBinaural Hearing. Why two ears? Definitions
Binaural Hearing Why two ears? Locating sounds in space: acuity is poorer than in vision by up to two orders of magnitude, but extends in all directions. Role in alerting and orienting? Separating sound
More informationA. SEK, E. SKRODZKA, E. OZIMEK and A. WICHER
ARCHIVES OF ACOUSTICS 29, 1, 25 34 (2004) INTELLIGIBILITY OF SPEECH PROCESSED BY A SPECTRAL CONTRAST ENHANCEMENT PROCEDURE AND A BINAURAL PROCEDURE A. SEK, E. SKRODZKA, E. OZIMEK and A. WICHER Institute
More informationSpeech perception of hearing aid users versus cochlear implantees
Speech perception of hearing aid users versus cochlear implantees SYDNEY '97 OtorhinolaIYngology M. FLYNN, R. DOWELL and G. CLARK Department ofotolaryngology, The University ofmelbourne (A US) SUMMARY
More information2/25/2013. Context Effect on Suprasegmental Cues. Supresegmental Cues. Pitch Contour Identification (PCI) Context Effect with Cochlear Implants
Context Effect on Segmental and Supresegmental Cues Preceding context has been found to affect phoneme recognition Stop consonant recognition (Mann, 1980) A continuum from /da/ to /ga/ was preceded by
More informationResearch Article The Acoustic and Peceptual Effects of Series and Parallel Processing
Hindawi Publishing Corporation EURASIP Journal on Advances in Signal Processing Volume 9, Article ID 6195, pages doi:1.1155/9/6195 Research Article The Acoustic and Peceptual Effects of Series and Parallel
More informationAuditory Scene Analysis
1 Auditory Scene Analysis Albert S. Bregman Department of Psychology McGill University 1205 Docteur Penfield Avenue Montreal, QC Canada H3A 1B1 E-mail: bregman@hebb.psych.mcgill.ca To appear in N.J. Smelzer
More informationEEL 6586, Project - Hearing Aids algorithms
EEL 6586, Project - Hearing Aids algorithms 1 Yan Yang, Jiang Lu, and Ming Xue I. PROBLEM STATEMENT We studied hearing loss algorithms in this project. As the conductive hearing loss is due to sound conducting
More informationIntelligibility of narrow-band speech and its relation to auditory functions in hearing-impaired listeners
Intelligibility of narrow-band speech and its relation to auditory functions in hearing-impaired listeners VRIJE UNIVERSITEIT Intelligibility of narrow-band speech and its relation to auditory functions
More informationHybrid Masking Algorithm for Universal Hearing Aid System
Hybrid Masking Algorithm for Universal Hearing Aid System H.Hensiba 1,Mrs.V.P.Brila 2, 1 Pg Student, Applied Electronics, C.S.I Institute Of Technology 2 Assistant Professor, Department of Electronics
More informationLecture 3: Perception
ELEN E4896 MUSIC SIGNAL PROCESSING Lecture 3: Perception 1. Ear Physiology 2. Auditory Psychophysics 3. Pitch Perception 4. Music Perception Dan Ellis Dept. Electrical Engineering, Columbia University
More informationMasker-signal relationships and sound level
Chapter 6: Masking Masking Masking: a process in which the threshold of one sound (signal) is raised by the presentation of another sound (masker). Masking represents the difference in decibels (db) between
More informationAuditory principles in speech processing do computers need silicon ears?
* with contributions by V. Hohmann, M. Kleinschmidt, T. Brand, J. Nix, R. Beutelmann, and more members of our medical physics group Prof. Dr. rer.. nat. Dr. med. Birger Kollmeier* Auditory principles in
More informationEvaluating the role of spectral and envelope characteristics in the intelligibility advantage of clear speech
Evaluating the role of spectral and envelope characteristics in the intelligibility advantage of clear speech Jean C. Krause a and Louis D. Braida Research Laboratory of Electronics, Massachusetts Institute
More informationNoise Susceptibility of Cochlear Implant Users: The Role of Spectral Resolution and Smearing
JARO 6: 19 27 (2004) DOI: 10.1007/s10162-004-5024-3 Noise Susceptibility of Cochlear Implant Users: The Role of Spectral Resolution and Smearing QIAN-JIE FU AND GERALDINE NOGAKI Department of Auditory
More informationAmbiguity in the recognition of phonetic vowels when using a bone conduction microphone
Acoustics 8 Paris Ambiguity in the recognition of phonetic vowels when using a bone conduction microphone V. Zimpfer a and K. Buck b a ISL, 5 rue du Général Cassagnou BP 734, 6831 Saint Louis, France b
More informationCombating the Reverberation Problem
Combating the Reverberation Problem Barbara Shinn-Cunningham (Boston University) Martin Cooke (Sheffield University, U.K.) How speech is corrupted by reverberation DeLiang Wang (Ohio State University)
More informationChallenges in microphone array processing for hearing aids. Volkmar Hamacher Siemens Audiological Engineering Group Erlangen, Germany
Challenges in microphone array processing for hearing aids Volkmar Hamacher Siemens Audiological Engineering Group Erlangen, Germany SIEMENS Audiological Engineering Group R&D Signal Processing and Audiology
More informationHearing II Perceptual Aspects
Hearing II Perceptual Aspects Overview of Topics Chapter 6 in Chaudhuri Intensity & Loudness Frequency & Pitch Auditory Space Perception 1 2 Intensity & Loudness Loudness is the subjective perceptual quality
More informationWhat Is the Difference between db HL and db SPL?
1 Psychoacoustics What Is the Difference between db HL and db SPL? The decibel (db ) is a logarithmic unit of measurement used to express the magnitude of a sound relative to some reference level. Decibels
More informationBrian D. Simpson Veridian, 5200 Springfield Pike, Suite 200, Dayton, Ohio 45431
The effects of spatial separation in distance on the informational and energetic masking of a nearby speech signal Douglas S. Brungart a) Air Force Research Laboratory, 2610 Seventh Street, Wright-Patterson
More informationTechnical Discussion HUSHCORE Acoustical Products & Systems
What Is Noise? Noise is unwanted sound which may be hazardous to health, interfere with speech and verbal communications or is otherwise disturbing, irritating or annoying. What Is Sound? Sound is defined
More informationLecture 4: Auditory Perception. Why study perception?
EE E682: Speech & Audio Processing & Recognition Lecture 4: Auditory Perception 1 2 3 4 5 6 Motivation: Why & how Auditory physiology Psychophysics: Detection & discrimination Pitch perception Speech perception
More informationLecture 9: Speech Recognition: Front Ends
EE E682: Speech & Audio Processing & Recognition Lecture 9: Speech Recognition: Front Ends 1 2 Recognizing Speech Feature Calculation Dan Ellis http://www.ee.columbia.edu/~dpwe/e682/
More informationHybridMaskingAlgorithmforUniversalHearingAidSystem. Hybrid Masking Algorithm for Universal Hearing Aid System
Global Journal of Researches in Engineering: Electrical and Electronics Engineering Volume 16 Issue 5 Version 1.0 Type: Double Blind Peer Reviewed International Research Journal Publisher: Global Journals
More informationFREQUENCY COMPRESSION AND FREQUENCY SHIFTING FOR THE HEARING IMPAIRED
FREQUENCY COMPRESSION AND FREQUENCY SHIFTING FOR THE HEARING IMPAIRED Francisco J. Fraga, Alan M. Marotta National Institute of Telecommunications, Santa Rita do Sapucaí - MG, Brazil Abstract A considerable
More informationSPHSC 462 HEARING DEVELOPMENT. Overview Review of Hearing Science Introduction
SPHSC 462 HEARING DEVELOPMENT Overview Review of Hearing Science Introduction 1 Overview of course and requirements Lecture/discussion; lecture notes on website http://faculty.washington.edu/lawerner/sphsc462/
More informationRepresentation of sound in the auditory nerve
Representation of sound in the auditory nerve Eric D. Young Department of Biomedical Engineering Johns Hopkins University Young, ED. Neural representation of spectral and temporal information in speech.
More informationEssential feature. Who are cochlear implants for? People with little or no hearing. substitute for faulty or missing inner hair
Who are cochlear implants for? Essential feature People with little or no hearing and little conductive component to the loss who receive little or no benefit from a hearing aid. Implants seem to work
More informationLateralized speech perception in normal-hearing and hearing-impaired listeners and its relationship to temporal processing
Lateralized speech perception in normal-hearing and hearing-impaired listeners and its relationship to temporal processing GUSZTÁV LŐCSEI,*, JULIE HEFTING PEDERSEN, SØREN LAUGESEN, SÉBASTIEN SANTURETTE,
More informationINTRODUCTION J. Acoust. Soc. Am. 103 (2), February /98/103(2)/1080/5/$ Acoustical Society of America 1080
Perceptual segregation of a harmonic from a vowel by interaural time difference in conjunction with mistuning and onset asynchrony C. J. Darwin and R. W. Hukin Experimental Psychology, University of Sussex,
More informationEffects of slow- and fast-acting compression on hearing impaired listeners consonantvowel identification in interrupted noise
Downloaded from orbit.dtu.dk on: Jan 01, 2018 Effects of slow- and fast-acting compression on hearing impaired listeners consonantvowel identification in interrupted noise Kowalewski, Borys; Zaar, Johannes;
More informationUsing Source Models in Speech Separation
Using Source Models in Speech Separation Dan Ellis Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Eng., Columbia Univ., NY USA dpwe@ee.columbia.edu http://labrosa.ee.columbia.edu/
More informationChanges in the Role of Intensity as a Cue for Fricative Categorisation
INTERSPEECH 2013 Changes in the Role of Intensity as a Cue for Fricative Categorisation Odette Scharenborg 1, Esther Janse 1,2 1 Centre for Language Studies and Donders Institute for Brain, Cognition and
More informationSpeech conveys not only linguistic content but. Vocal Emotion Recognition by Normal-Hearing Listeners and Cochlear Implant Users
Cochlear Implants Special Issue Article Vocal Emotion Recognition by Normal-Hearing Listeners and Cochlear Implant Users Trends in Amplification Volume 11 Number 4 December 2007 301-315 2007 Sage Publications
More informationAuditory nerve. Amanda M. Lauer, Ph.D. Dept. of Otolaryngology-HNS
Auditory nerve Amanda M. Lauer, Ph.D. Dept. of Otolaryngology-HNS May 30, 2016 Overview Pathways (structural organization) Responses Damage Basic structure of the auditory nerve Auditory nerve in the cochlea
More informationSLHS 1301 The Physics and Biology of Spoken Language. Practice Exam 2. b) 2 32
SLHS 1301 The Physics and Biology of Spoken Language Practice Exam 2 Chapter 9 1. In analog-to-digital conversion, quantization of the signal means that a) small differences in signal amplitude over time
More informationUvA-DARE (Digital Academic Repository)
UvA-DARE (Digital Academic Repository) A Speech Intelligibility Index-based approach to predict the speech reception threshold for sentences in fluctuating noise for normal-hearing listeners Rhebergen,
More informationEffect of Consonant Duration Modifications on Speech Perception in Noise-II
International Journal of Electronics Engineering, 2(1), 2010, pp. 75-81 Effect of Consonant Duration Modifications on Speech Perception in Noise-II NH Shobha 1, TG Thomas 2 & K Subbarao 3 1 Research Scholar,
More informationAuditory Scene Analysis: phenomena, theories and computational models
Auditory Scene Analysis: phenomena, theories and computational models July 1998 Dan Ellis International Computer Science Institute, Berkeley CA Outline 1 2 3 4 The computational
More informationRelease from informational masking in a monaural competingspeech task with vocoded copies of the maskers presented contralaterally
Release from informational masking in a monaural competingspeech task with vocoded copies of the maskers presented contralaterally Joshua G. W. Bernstein a) National Military Audiology and Speech Pathology
More informationElements of Effective Hearing Aid Performance (2004) Edgar Villchur Feb 2004 HearingOnline
Elements of Effective Hearing Aid Performance (2004) Edgar Villchur Feb 2004 HearingOnline To the hearing-impaired listener the fidelity of a hearing aid is not fidelity to the input sound but fidelity
More informationThe effects of aging on temporal masking
University of South Florida Scholar Commons Graduate Theses and Dissertations Graduate School 2010 The effects of aging on temporal masking Susan E. Fulton University of South Florida Follow this and additional
More informationMasking release and the contribution of obstruent consonants on speech recognition in noise by cochlear implant users
Masking release and the contribution of obstruent consonants on speech recognition in noise by cochlear implant users Ning Li and Philipos C. Loizou a Department of Electrical Engineering, University of
More informationBuilding Skills to Optimize Achievement for Students with Hearing Loss
PART 2 Building Skills to Optimize Achievement for Students with Hearing Loss Karen L. Anderson, PhD Butte Publications, 2011 1 Look at the ATCAT pg 27-46 Are there areas assessed that you do not do now?
More informationRevisiting the right-ear advantage for speech: Implications for speech displays
INTERSPEECH 2014 Revisiting the right-ear advantage for speech: Implications for speech displays Nandini Iyer 1, Eric Thompson 2, Brian Simpson 1, Griffin Romigh 1 1 Air Force Research Laboratory; 2 Ball
More informationModulation and Top-Down Processing in Audition
Modulation and Top-Down Processing in Audition Malcolm Slaney 1,2 and Greg Sell 2 1 Yahoo! Research 2 Stanford CCRMA Outline The Non-Linear Cochlea Correlogram Pitch Modulation and Demodulation Information
More informationPrescribe hearing aids to:
Harvey Dillon Audiology NOW! Prescribing hearing aids for adults and children Prescribing hearing aids for adults and children Adult Measure hearing thresholds (db HL) Child Measure hearing thresholds
More information