Auditory-Visual Speech Perception Laboratory
|
|
- Eileen McKinney
- 5 years ago
- Views:
Transcription
1 Auditory-Visual Speech Perception Laboratory Research Focus: Identify perceptual processes involved in auditory-visual speech perception Determine the abilities of individual patients to carry out these processes successfully Design intervention strategies using signal processing technologies and training techniques to remedy any deficiencies that may be found. Primary Funding: NIH Grant: DC Al NSF Grant (subcontract): SBR Learning and Intelligent Systems Initiative of the National Science DARPA Grant (subcontract): ONR Award N Staffing: Lab Director: Ken W. Grant Research Associate: Mary T. Cord
2 Auditory-Visual Speech Perception Laboratory Collaborations: Steven Greenberg - The Speech Institute, Oakland, CA David Poeppel - University of Maryland, College Park, MD Virginie van Wassenhove - University of Maryland, College Park, MD
3 Topics Auditory Supplements to Speechreading Bimodal Comodulation Spectro-Temporal Window of Integration
4 Auditory Supplements to Speechreading
5 Selective Needs of Severe-to to-profound Hearing Loss Speech Input Hearing Capacity compressed dynamic range compressed frequency range Speechreading/lipreading as primary channel of spoken language reception
6 Percent Correct Speech Recognition: Sentences B B J B J B J J B B B B B B J J J J J 20 B Auditory-Visual J Auditory J S/N (db) Roughly 6 db improvement in S/N; roughly 30% improvement in intelligibility for NH subjects
7 Speech Recognition: Consonants Percent Correct Recognition Speech-to-Noise Ratio (db) NH - Auditory Consonants HI Auditory Consonants HI-Auditory Sentences ASR Sentences
8 Auditory-Visual vs. Audio Speech Recognition Percent Correct Recognition Speech-to-Noise Ratio (db) NH - Audiovisual Consonants NH - Auditory Consonants HI - Audiovisual Consonants HI Auditory Consonants HI-Auditory Sentences ASR Sentences Roughly 6 db improvement in S/N; roughly 30% improvement in intelligibility for NH subjects.
9 Possible Roles of Speechreading Provide segmental information that is redundant with acoustic information
10 Possible Roles of Speechreading Provide segmental information that is redundant with acoustic information Provide segmental information that is complementary with acoustic information
11 Possible Roles of Speechreading Provide segmental information that is redundant with acoustic information Provide segmental information that is complementary with acoustic information Direct auditory analyses to the target signal who, where, when, what (spectral)
12 Auditory-Visual Speech Recognition: Consonants What information is available through speechreading? Which acoustic signals supplement speechreading? Are there significant audio-visual interactions in speech processing?
13 Visual Feature Recognition 4% 0% 3% Voicing Manner Place Other From Grant, K.W., and Walden, B.E. (1996). J. Acoust. Soc. Am. 100, % %Information Transmitted re: Total Information Received
14 Speech Recognition: Consonants Linguistic feature contributions to visual speech recognition. The top row represents typical feature classifications for speechreading alone (visemes). Each subsequent row represents the effects of adding information about another linguistic feature via an additional input channel (in this case auditory). Note that as additional features are added, consonant confusions associated with speechreading are resolved to a greater and greater extent. Speechreading Voicing Nasality Affrication p,b,m t,d,n g,k f,v T,D s,z S,tS,dZ,Z l r w j p b,m t d,n g k f v T D s z S,tS dz,z l r w j p b m t d n g k f v T D s z S,tS dz,z l r w j p b m t d n g k f v T D s z S ts dz Z l r w j
15 Hypothetical Consonant Recognition - Perfect Feature Transmission 1 Probability Correct A 0 PRE Voicing Manner Place Voicing + Manner Feature(s) Transmitted Auditory consonant recognition based on perfect transmission of indicated feature. Responses within each feature category were uniformly distributed. Predicted AV consonant recognition based on PRE model of integration (Braida, 1991).
16 Designer Acoustic Signals: Minimal Bandwidth, Maximum Benefit Frank and Joe Hardy scanned the wide valley that appeared before them as their yellow sports sedan rounded the crest of a hill. In the distance, a huge cylindrical tower rose from the valley floor. "Looks like a giant barnacle", Joe remarked to his older, dark-haired brother Frank. Biff Hooper, a tall, muscular high school friend of the two amateur detectives, leaned forward from the back seat. "You're looking at the cooling tower. The reactor itself is in the building next to it." Biff's uncle, Jerry Hooper, was a nuclear engineer at the Bayridge Nuclear Power Plant located outside Bayport. He had invited the three boys on a private afternoon tour of the facility. The summer had just begun, and the Hardy brothers were eager for new adventures. WB LPF AM AMFM
17 Designer Acoustic Signals: Minimal Bandwidth, Maximum Benefit Tracking Rate (WPM) SA AM FM AMFM LPF Receiving Condition Tracking Rate (% re: Normal)
18 Designer Acoustic Signals: AM Bands 100 CUNY Sentences Percent Correct / / SA WB 500 Hz 1600 Hz 3150 Hz Prefilter Condition
19 AM Bands - Smoothing-Filter Effects Percent Correct Percent Correct CUNY Sentences 100 IEEE Sentences Smoothing Filter Cutoff (Hz) Prefilter-Carrier SA
20 Problems With Articulation Theory Effective AI With Visual Cues ANSI, Calculated AI
21 Auditory-Visual Spectral Interactions: Consonants A AV Percent Correct Filter Condition Condition (Hz)
22 Feature Distribution re: Center Frequency Percent Information Transmitted (re: Total Received) Voicing + Manner Place From Grant, K.W., and Walden, B.E. (1996). J. Acoust. Soc. Am. 100, Bandwidth Center Frequency (Hz)
23 How It All Works - The Prevailing View Information extracted from both sources independently Integration of extracted information Decision statistic Evaluation Integration Decision From Massaro, 1998
24 A More Expanded View of the Process
25 Auditory Supplements to Speechreading SUMMARY: Speechreading provides information mostly about place-ofarticulation
26 Auditory Supplements to Speechreading SUMMARY: Speechreading provides information mostly about place-ofarticulation Auditory-visual speech recognition is determined primarily by complementary cues between visual and auditory modalities
27 Auditory Supplements to Speechreading SUMMARY: Speechreading provides information mostly about place-ofarticulation Auditory-visual speech recognition is determined primarily by complementary cues between visual and auditory modalities The most intelligible auditory speech signals do not necessarily result in the most intelligible auditory-visual speech signal
28 Auditory Supplements to Speechreading SUMMARY: Speechreading provides information mostly about place-ofarticulation Auditory-visual speech recognition is determined primarily by complementary cues between visual and auditory modalities The most intelligible auditory speech signals do not necessarily result in the most intelligible auditory-visual speech signal Acoustic cues for voicing and manner-or articulation are the best supplements to speechreading
29 Auditory Supplements to Speechreading SUMMARY: Speechreading provides information mostly about place-ofarticulation Auditory-visual speech recognition is determined primarily by complementary cues between visual and auditory modalities The most intelligible auditory speech signals do not necessarily result in the most intelligible auditory-visual speech signal Acoustic cues for voicing and manner-or articulation are the best supplements to speechreading These cues tend to be low frequency
30 Bimodal Coherence: Audio and Visual Comodulation
31 The Independence of Sensory Systems??? Information is extracted independently from A and V modalities Early versus Late Integration Most models are "late integration" models
32 Back to Our Conceptual Framework
33 The Independence of Sensory Systems??? Information is extracted independently from A and V modalities Early versus Late Integration Most models are "late integration" models BUT Speechreading activates primary auditory cortex (cf. Sams et al., 1991)
34 The Independence of Sensory Systems??? Information is extracted independently from A and V modalities Early versus Late Integration Most models are "late integration" models BUT Speechreading activates primary auditory cortex (cf. Sams et al., 1991) Population of neurons in cat Superior Colliculus respond only to bimodal input (cf. Stein and Meredith, 1993)
35 Sensory Integration: Many Questions How is the auditory system modulated by visual speech activity? What is the temporal window governing this interaction?
36 Bimodal Coherence Masking Protection BCMP (Grant and Seitz, 2000, Grant, 2001) Detection of speech in noise is improved by watching a talker (i.e., speechreading) as they produce the target speech signal, provided that the "visible" movement of the lips and acoustic amplitude envelope are highly correlated.
37 Basic Paradigm for BCMP: Exp. 1 Auditory-only speech detection... speech Auditory-visual speech detection... speech noise noise noise noise
38 Methodology for Orthographic BCMP: Exp. 2 Auditory-only speech detection Auditory + orthographic speech detection... speech text... speech noise noise noise noise
39 Methodology for Filtered BCMP: Exp. 3 F1 ( Hz) F2 ( Hz) Auditory-only detection of filtered speech Auditory-visual detection of filtered speech... speech... speech noise noise noise noise
40 Congruent versus Incongruent Speech Masked Threshold (db) A AV M AV UM -21 Sentence 1 Sentence 2 Sentence 3 Average Target Sentence
41 BCMP for Congruent and Incongruent Speech Bimodal Coherence Protection (db) AV M 2.15 AV UM Sentence 1 Sentence 2 Sentence 3 Average Target Sentence
42 Bimodal Coherence Protection (db) BCMP for Orthographic Speech Sentence 1 Sentence 2 Sentence 3 Average Target Sentence
43 Masking Protection (db) BCMP for Filtered Speech AV F1 AV F2 AV WB AV O 0 Sentence 2 Sentence 3 Average Target Sentence
44 Bimodal Masking Release (db) Average BCMP AV WB AV F2 AV F1 AV O Condition Bimodal comodulation masking protection for wideband speech (WB), filtered speech (F2 and F1), and for orthographically cued speech (O).
45 Acoustic Envelope and Lip Area Functions RMS Amplitude (db) WB F1 F2 F3 Lip Area (in 2 ) Watch the log float in the wide river Time (ms)
46 Cross Modality Correlation - Lip Area versus Amplitude Envelope Sentence 3 Sentence 2 WB r=0.52 r=0.35 RMS Envelope r=0.49 r=0.65 F1 F2 r=0.32 r=0.47 r=0.60 F3 r=0.45 Lip Area Lip Area
47 Local Correlations (Lip Area versus F2-Amplitude Envelope Correlation Coefficient Watch the log float in the wide river Time (ms) RMS Amplitude (db)
48 Local Correlations (Lip Area versus F2-Amplitude Envelope Correlation Coefficient Both brothers wear the same size Time (ms) RMS Amplitude (db)
49 Cross Modality Correlations (lip area versus acoustic envelope) Cross Modality Correlations (lip area versus acoustic envelope) S2 S3 Time (ms) S F F S F F2 Time (ms) Correlation Coefficient Peak Amplitude (db) Correlation Coefficient Peak Amplitude (db) Correlation Coefficient Correlation Coefficient Peak Amplitude (db) Peak Amplitude (db)
50 BCMP Summary Speechreading reduces auditory speech detection thresholds by about 1.5 db (range: 1-3 db depending on sentence)
51 BCMP Summary Speechreading reduces auditory speech detection thresholds by about 1.5 db (range: 1-3 db depending on sentence) Amount of BCMP depends on the degree of coherence between acoustic envelope and facial kinematics
52 BCMP Summary Speechreading reduces auditory speech detection thresholds by about 1.5 db (range: 1-3 db depending on sentence) Amount of BCMP depends on the degree of coherence between acoustic envelope and facial kinematics Providing listeners with explicit (orthographic) knowledge of the identity of the target sentence reduces speech detection thresholds by about 0.5 db, independent of the specific target sentence
53 BCMP Summary Speechreading reduces auditory speech detection thresholds by about 1.5 db (range: 1-3 db depending on sentence) Amount of BCMP depends on the degree of coherence between acoustic envelope and facial kinematics Providing listeners with explicit (orthographic) knowledge of the identity of the target sentence reduces speech detection thresholds by about 0.5 db, independent of the specific target sentence Manipulating the degree of coherence between area of mouth opening and acoustic envelope by filtering the target speech has a direct effect on BCMP
54 Temporal Window for Auditory-Visual Integration
55 AUDIO-ALONE ALONE EXPERIMENTS
56 Audio (Alone) Spectral Slit Paradigm The edge of each slit was separated from its nearest neighbor by an octave Can listeners decode spoken sentences using just four narrow (1/3 octave) channels ( slits ) distributed across the spectrum? YES (cf. next slide) What is the intelligibility of each slit alone and in combination with others?
57 Word Intelligibility - Single and Multiple Slits 89% 60% 13% Slit Number CF (Hz) 2% 9% 9% 4% Slit Number CF (Hz)
58 Slit Asynchrony Affects Intelligibility Desynchronizing the slits by more than 25 ms results in a significant decline in intelligibility The effect of asynchrony on intelligibility is relatively symmetrical These data are from a different set of subjects than those participating in the study described earlier - hence slightly different numbers for the baseline conditions
59 Cross-Spectral Temporal Asynchrony Effects TIMIT Word Recognition (%) Bands Bands Slit Asynchrony (ms) From Greenberg, Arai, and Silipo (1998). Proc. ICSLP, Sydney, Dec. 1-4.
60 AUDIO-VISUAL EXPERIMENTS
61 IEEE Sentences Auditory-Visual Tasks Recognition of key words Audio slits Video presented at various temporal asynchronies CV Syllables Recognition of McGurk pairs Audio /pa/, /ba/, /ta/, /da/ Video /ka/, /ga/, /ta/, /da/ Synchrony identification and discrimination Yes/No single interval simultaneity judgments 2IFC adaptive tracking
62 Auditory-Visual Asynchrony - Paradigm
63 Cross-Modality Temporal Asynchrony Effects: Sentences 100 IEEE Word Recognition (%) Audio Alone - Bands Speechreading Alone Audio Delay (ms)
64 Auditory-Visual Integration - by Individual S's These data are complex, but the implications are clear. Video signal leading is better than synchronous for 8 of 9 subjects Audio-visual integration is a complicated, poorly understood process, at least with respect to speech intelligibility Variation across subjects
65 McGurk Synchrony Paradigm
66 Temporal Integration in the McGurk Effect Response Probability /pa/ /ta/ /ka/ Audio Delay (ms)
67 Simultaneity Judgements - Natural vs. McGurk AV Tokens 1 A d V d Probability of Simultaneity A t V t A p V k A b V g Audio Delay (ms)
68 Spectro-Temporal Synchrony Discrimination Filter Band Condition (Hz) WB Audio Lead Audio Lag Audio Delay (ms)
69 Temporal Window of Integration McGurk Synchrony Discrimination McGurk Synchrony Identification Audio Audiovisual McGurk Fusion Identification CV Synchrony Discrimination CV Synchrony Identification IEEE Synchrony Discrimination IEEE Word Identification IEEE Synchrony Discrimination TIMIT Word Identification Audio Delay (ms)
70 Spectro-Temporal Integration: Summary Within Modality (Cross- Spectral Auditory Integration)
71 Spectro-Temporal Integration: Summary Within Modality (Cross- Spectral Auditory Integration) TWI is symmetrical
72 Spectro-Temporal Integration: Summary Within Modality (Cross- Spectral Auditory Integration) TWI is symmetrical TWI roughly 50 ms or less (phoneme?)
73 Spectro-Temporal Integration: Summary Within Modality (Cross- Spectral Auditory Integration) TWI is symmetrical TWI roughly 50 ms or less (phoneme?) Across Modality (Cross-Modal AV Integration)
74 Spectro-Temporal Integration: Summary Within Modality (Cross- Spectral Auditory Integration) TWI is symmetrical TWI roughly 50 ms or less (phoneme?) Across Modality (Cross-Modal AV Integration) TWI is highly asymmetrical favoring visual leads
75 Spectro-Temporal Integration: Summary Within Modality (Cross- Spectral Auditory Integration) TWI is symmetrical TWI roughly 50 ms or less (phoneme?) Across Modality (Cross-Modal AV Integration) TWI is highly asymmetrical favoring visual leads TWI is roughly ms (syllable?)
76 Spectro-Temporal Integration: Summary Within Modality (Cross- Spectral Auditory Integration) TWI is symmetrical TWI roughly 50 ms or less (phoneme?) Across Modality (Cross-Modal AV Integration) TWI is highly asymmetrical favoring visual leads TWI is roughly ms (syllable?) TWI for Incongruent CV's (McGurk Stimuli) is not as wide as TWI for natural congruent CV's
77 Auditory-Visual Speech Perception Laboratory Walter Reed Army Medical Center Army Audiology and Speech Center Washington, DC USA wramc.amedd.army.mil/departments/.army.mil/departments/aasc/avlabavlab
Detection of auditory (cross-spectral) and auditory visual (cross-modal) synchrony
Speech Communication 44 (2004) 43 53 www.elsevier.com/locate/specom Detection of auditory (cross-spectral) and auditory visual (cross-modal) synchrony Ken W. Grant a, *, Virginie van Wassenhove b, David
More informationHCS 7367 Speech Perception
Long-term spectrum of speech HCS 7367 Speech Perception Connected speech Absolute threshold Males Dr. Peter Assmann Fall 212 Females Long-term spectrum of speech Vowels Males Females 2) Absolute threshold
More informationHCS 7367 Speech Perception
Babies 'cry in mother's tongue' HCS 7367 Speech Perception Dr. Peter Assmann Fall 212 Babies' cries imitate their mother tongue as early as three days old German researchers say babies begin to pick up
More informationMULTI-CHANNEL COMMUNICATION
INTRODUCTION Research on the Deaf Brain is beginning to provide a new evidence base for policy and practice in relation to intervention with deaf children. This talk outlines the multi-channel nature of
More informationTHE ROLE OF VISUAL SPEECH CUES IN THE AUDITORY PERCEPTION OF SYNTHETIC STIMULI BY CHILDREN USING A COCHLEAR IMPLANT AND CHILDREN WITH NORMAL HEARING
THE ROLE OF VISUAL SPEECH CUES IN THE AUDITORY PERCEPTION OF SYNTHETIC STIMULI BY CHILDREN USING A COCHLEAR IMPLANT AND CHILDREN WITH NORMAL HEARING Vanessa Surowiecki 1, vid Grayden 1, Richard Dowell
More informationProviding Effective Communication Access
Providing Effective Communication Access 2 nd International Hearing Loop Conference June 19 th, 2011 Matthew H. Bakke, Ph.D., CCC A Gallaudet University Outline of the Presentation Factors Affecting Communication
More informationLanguage Speech. Speech is the preferred modality for language.
Language Speech Speech is the preferred modality for language. Outer ear Collects sound waves. The configuration of the outer ear serves to amplify sound, particularly at 2000-5000 Hz, a frequency range
More informationThe Role of Information Redundancy in Audiovisual Speech Integration. A Senior Honors Thesis
The Role of Information Redundancy in Audiovisual Speech Integration A Senior Honors Thesis Printed in Partial Fulfillment of the Requirement for graduation with distinction in Speech and Hearing Science
More informationSpeech Reading Training and Audio-Visual Integration in a Child with Autism Spectrum Disorder
University of Arkansas, Fayetteville ScholarWorks@UARK Rehabilitation, Human Resources and Communication Disorders Undergraduate Honors Theses Rehabilitation, Human Resources and Communication Disorders
More informationELECTROPHYSIOLOGY OF UNIMODAL AND AUDIOVISUAL SPEECH PERCEPTION
AVSP 2 International Conference on Auditory-Visual Speech Processing ELECTROPHYSIOLOGY OF UNIMODAL AND AUDIOVISUAL SPEECH PERCEPTION Lynne E. Bernstein, Curtis W. Ponton 2, Edward T. Auer, Jr. House Ear
More informationAuditory-Visual Integration of Sine-Wave Speech. A Senior Honors Thesis
Auditory-Visual Integration of Sine-Wave Speech A Senior Honors Thesis Presented in Partial Fulfillment of the Requirements for Graduation with Distinction in Speech and Hearing Science in the Undergraduate
More informationAudio-Visual Integration: Generalization Across Talkers. A Senior Honors Thesis
Audio-Visual Integration: Generalization Across Talkers A Senior Honors Thesis Presented in Partial Fulfillment of the Requirements for graduation with research distinction in Speech and Hearing Science
More informationILLUSIONS AND ISSUES IN BIMODAL SPEECH PERCEPTION
ISCA Archive ILLUSIONS AND ISSUES IN BIMODAL SPEECH PERCEPTION Dominic W. Massaro Perceptual Science Laboratory (http://mambo.ucsc.edu/psl/pslfan.html) University of California Santa Cruz, CA 95064 massaro@fuzzy.ucsc.edu
More informationTactile Communication of Speech
Tactile Communication of Speech RLE Group Sensory Communication Group Sponsor National Institutes of Health/National Institute on Deafness and Other Communication Disorders Grant 2 R01 DC00126, Grant 1
More informationRobust Neural Encoding of Speech in Human Auditory Cortex
Robust Neural Encoding of Speech in Human Auditory Cortex Nai Ding, Jonathan Z. Simon Electrical Engineering / Biology University of Maryland, College Park Auditory Processing in Natural Scenes How is
More informationGick et al.: JASA Express Letters DOI: / Published Online 17 March 2008
modality when that information is coupled with information via another modality (e.g., McGrath and Summerfield, 1985). It is unknown, however, whether there exist complex relationships across modalities,
More informationFREQUENCY COMPRESSION AND FREQUENCY SHIFTING FOR THE HEARING IMPAIRED
FREQUENCY COMPRESSION AND FREQUENCY SHIFTING FOR THE HEARING IMPAIRED Francisco J. Fraga, Alan M. Marotta National Institute of Telecommunications, Santa Rita do Sapucaí - MG, Brazil Abstract A considerable
More informationHearing the Universal Language: Music and Cochlear Implants
Hearing the Universal Language: Music and Cochlear Implants Professor Hugh McDermott Deputy Director (Research) The Bionics Institute of Australia, Professorial Fellow The University of Melbourne Overview?
More informationA Senior Honors Thesis. Brandie Andrews
Auditory and Visual Information Facilitating Speech Integration A Senior Honors Thesis Presented in Partial Fulfillment of the Requirements for graduation with distinction in Speech and Hearing Science
More informationConsonant Perception test
Consonant Perception test Introduction The Vowel-Consonant-Vowel (VCV) test is used in clinics to evaluate how well a listener can recognize consonants under different conditions (e.g. with and without
More informationStudy of perceptual balance for binaural dichotic presentation
Paper No. 556 Proceedings of 20 th International Congress on Acoustics, ICA 2010 23-27 August 2010, Sydney, Australia Study of perceptual balance for binaural dichotic presentation Pandurangarao N. Kulkarni
More informationSpeech perception in individuals with dementia of the Alzheimer s type (DAT) Mitchell S. Sommers Department of Psychology Washington University
Speech perception in individuals with dementia of the Alzheimer s type (DAT) Mitchell S. Sommers Department of Psychology Washington University Overview Goals of studying speech perception in individuals
More informationOptimal Filter Perception of Speech Sounds: Implications to Hearing Aid Fitting through Verbotonal Rehabilitation
Optimal Filter Perception of Speech Sounds: Implications to Hearing Aid Fitting through Verbotonal Rehabilitation Kazunari J. Koike, Ph.D., CCC-A Professor & Director of Audiology Department of Otolaryngology
More informationBest Practice Protocols
Best Practice Protocols SoundRecover for children What is SoundRecover? SoundRecover (non-linear frequency compression) seeks to give greater audibility of high-frequency everyday sounds by compressing
More informationAnalysis of Emotion Recognition using Facial Expressions, Speech and Multimodal Information
Analysis of Emotion Recognition using Facial Expressions, Speech and Multimodal Information C. Busso, Z. Deng, S. Yildirim, M. Bulut, C. M. Lee, A. Kazemzadeh, S. Lee, U. Neumann, S. Narayanan Emotion
More informationTime Varying Comb Filters to Reduce Spectral and Temporal Masking in Sensorineural Hearing Impairment
Bio Vision 2001 Intl Conf. Biomed. Engg., Bangalore, India, 21-24 Dec. 2001, paper PRN6. Time Varying Comb Filters to Reduce pectral and Temporal Masking in ensorineural Hearing Impairment Dakshayani.
More informationModulation and Top-Down Processing in Audition
Modulation and Top-Down Processing in Audition Malcolm Slaney 1,2 and Greg Sell 2 1 Yahoo! Research 2 Stanford CCRMA Outline The Non-Linear Cochlea Correlogram Pitch Modulation and Demodulation Information
More informationSpectrograms (revisited)
Spectrograms (revisited) We begin the lecture by reviewing the units of spectrograms, which I had only glossed over when I covered spectrograms at the end of lecture 19. We then relate the blocks of a
More informationAcoustics, signals & systems for audiology. Psychoacoustics of hearing impairment
Acoustics, signals & systems for audiology Psychoacoustics of hearing impairment Three main types of hearing impairment Conductive Sound is not properly transmitted from the outer to the inner ear Sensorineural
More informationBINAURAL DICHOTIC PRESENTATION FOR MODERATE BILATERAL SENSORINEURAL HEARING-IMPAIRED
International Conference on Systemics, Cybernetics and Informatics, February 12 15, 2004 BINAURAL DICHOTIC PRESENTATION FOR MODERATE BILATERAL SENSORINEURAL HEARING-IMPAIRED Alice N. Cheeran Biomedical
More informationUSING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES
USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES Varinthira Duangudom and David V Anderson School of Electrical and Computer Engineering, Georgia Institute of Technology Atlanta, GA 30332
More informationThe Deaf Brain. Bencie Woll Deafness Cognition and Language Research Centre
The Deaf Brain Bencie Woll Deafness Cognition and Language Research Centre 1 Outline Introduction: the multi-channel nature of language Audio-visual language BSL Speech processing Silent speech Auditory
More informationAUDL GS08/GAV1 Signals, systems, acoustics and the ear. Pitch & Binaural listening
AUDL GS08/GAV1 Signals, systems, acoustics and the ear Pitch & Binaural listening Review 25 20 15 10 5 0-5 100 1000 10000 25 20 15 10 5 0-5 100 1000 10000 Part I: Auditory frequency selectivity Tuning
More informationAsynchronous glimpsing of speech: Spread of masking and task set-size
Asynchronous glimpsing of speech: Spread of masking and task set-size Erol J. Ozmeral, a) Emily Buss, and Joseph W. Hall III Department of Otolaryngology/Head and Neck Surgery, University of North Carolina
More informationACOUSTIC AND PERCEPTUAL PROPERTIES OF ENGLISH FRICATIVES
ISCA Archive ACOUSTIC AND PERCEPTUAL PROPERTIES OF ENGLISH FRICATIVES Allard Jongman 1, Yue Wang 2, and Joan Sereno 1 1 Linguistics Department, University of Kansas, Lawrence, KS 66045 U.S.A. 2 Department
More informationAssessing Hearing and Speech Recognition
Assessing Hearing and Speech Recognition Audiological Rehabilitation Quick Review Audiogram Types of hearing loss hearing loss hearing loss Testing Air conduction Bone conduction Familiar Sounds Audiogram
More information2/25/2013. Context Effect on Suprasegmental Cues. Supresegmental Cues. Pitch Contour Identification (PCI) Context Effect with Cochlear Implants
Context Effect on Segmental and Supresegmental Cues Preceding context has been found to affect phoneme recognition Stop consonant recognition (Mann, 1980) A continuum from /da/ to /ga/ was preceded by
More informationRole of F0 differences in source segregation
Role of F0 differences in source segregation Andrew J. Oxenham Research Laboratory of Electronics, MIT and Harvard-MIT Speech and Hearing Bioscience and Technology Program Rationale Many aspects of segregation
More informationJuan Carlos Tejero-Calado 1, Janet C. Rutledge 2, and Peggy B. Nelson 3
PRESERVING SPECTRAL CONTRAST IN AMPLITUDE COMPRESSION FOR HEARING AIDS Juan Carlos Tejero-Calado 1, Janet C. Rutledge 2, and Peggy B. Nelson 3 1 University of Malaga, Campus de Teatinos-Complejo Tecnol
More informationThe effect of wearing conventional and level-dependent hearing protectors on speech production in noise and quiet
The effect of wearing conventional and level-dependent hearing protectors on speech production in noise and quiet Ghazaleh Vaziri Christian Giguère Hilmi R. Dajani Nicolas Ellaham Annual National Hearing
More informationNoise-Robust Speech Recognition in a Car Environment Based on the Acoustic Features of Car Interior Noise
4 Special Issue Speech-Based Interfaces in Vehicles Research Report Noise-Robust Speech Recognition in a Car Environment Based on the Acoustic Features of Car Interior Noise Hiroyuki Hoshino Abstract This
More informationHearing. and other senses
Hearing and other senses Sound Sound: sensed variations in air pressure Frequency: number of peaks that pass a point per second (Hz) Pitch 2 Some Sound and Hearing Links Useful (and moderately entertaining)
More informationPerceptual congruency of audio-visual speech affects ventriloquism with bilateral visual stimuli
Psychon Bull Rev (2011) 18:123 128 DOI 10.3758/s13423-010-0027-z Perceptual congruency of audio-visual speech affects ventriloquism with bilateral visual stimuli Shoko Kanaya & Kazuhiko Yokosawa Published
More informationLecture 3: Perception
ELEN E4896 MUSIC SIGNAL PROCESSING Lecture 3: Perception 1. Ear Physiology 2. Auditory Psychophysics 3. Pitch Perception 4. Music Perception Dan Ellis Dept. Electrical Engineering, Columbia University
More informationINTRODUCTION J. Acoust. Soc. Am. 103 (2), February /98/103(2)/1080/5/$ Acoustical Society of America 1080
Perceptual segregation of a harmonic from a vowel by interaural time difference in conjunction with mistuning and onset asynchrony C. J. Darwin and R. W. Hukin Experimental Psychology, University of Sussex,
More informationS everal studies indicate that the identification/recognition. Identification Performance by Right- and Lefthanded Listeners on Dichotic CV Materials
J Am Acad Audiol 7 : 1-6 (1996) Identification Performance by Right- and Lefthanded Listeners on Dichotic CV Materials Richard H. Wilson* Elizabeth D. Leigh* Abstract Normative data from 24 right-handed
More informationChapter 40 Effects of Peripheral Tuning on the Auditory Nerve s Representation of Speech Envelope and Temporal Fine Structure Cues
Chapter 40 Effects of Peripheral Tuning on the Auditory Nerve s Representation of Speech Envelope and Temporal Fine Structure Cues Rasha A. Ibrahim and Ian C. Bruce Abstract A number of studies have explored
More informationSlow compression for people with severe to profound hearing loss
Phonak Insight February 2018 Slow compression for people with severe to profound hearing loss For people with severe to profound hearing loss, poor auditory resolution abilities can make the spectral and
More informationPrelude Envelope and temporal fine. What's all the fuss? Modulating a wave. Decomposing waveforms. The psychophysics of cochlear
The psychophysics of cochlear implants Stuart Rosen Professor of Speech and Hearing Science Speech, Hearing and Phonetic Sciences Division of Psychology & Language Sciences Prelude Envelope and temporal
More informationMODALITY, PERCEPTUAL ENCODING SPEED, AND TIME-COURSE OF PHONETIC INFORMATION
ISCA Archive MODALITY, PERCEPTUAL ENCODING SPEED, AND TIME-COURSE OF PHONETIC INFORMATION Philip Franz Seitz and Ken W. Grant Army Audiology and Speech Center Walter Reed Army Medical Center Washington,
More informationCross-Frequency Integration for Consonant and Vowel Identification in Bimodal Hearing
Cross-Frequency Integration for Consonant and Vowel Identification in Bimodal Hearing The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters.
More informationEnrique A. Lopez-Poveda Alan R. Palmer Ray Meddis Editors. The Neurophysiological Bases of Auditory Perception
Enrique A. Lopez-Poveda Alan R. Palmer Ray Meddis Editors The Neurophysiological Bases of Auditory Perception 123 The Neurophysiological Bases of Auditory Perception Enrique A. Lopez-Poveda Alan R. Palmer
More informationOutline.! Neural representation of speech sounds. " Basic intro " Sounds and categories " How do we perceive sounds? " Is speech sounds special?
Outline! Neural representation of speech sounds " Basic intro " Sounds and categories " How do we perceive sounds? " Is speech sounds special? ! What is a phoneme?! It s the basic linguistic unit of speech!
More informationEEL 6586, Project - Hearing Aids algorithms
EEL 6586, Project - Hearing Aids algorithms 1 Yan Yang, Jiang Lu, and Ming Xue I. PROBLEM STATEMENT We studied hearing loss algorithms in this project. As the conductive hearing loss is due to sound conducting
More informationHearing Lectures. Acoustics of Speech and Hearing. Auditory Lighthouse. Facts about Timbre. Analysis of Complex Sounds
Hearing Lectures Acoustics of Speech and Hearing Week 2-10 Hearing 3: Auditory Filtering 1. Loudness of sinusoids mainly (see Web tutorial for more) 2. Pitch of sinusoids mainly (see Web tutorial for more)
More informationPerceptual Effects of Nasal Cue Modification
Send Orders for Reprints to reprints@benthamscience.ae The Open Electrical & Electronic Engineering Journal, 2015, 9, 399-407 399 Perceptual Effects of Nasal Cue Modification Open Access Fan Bai 1,2,*
More informationInfant Hearing Development: Translating Research Findings into Clinical Practice. Auditory Development. Overview
Infant Hearing Development: Translating Research Findings into Clinical Practice Lori J. Leibold Department of Allied Health Sciences The University of North Carolina at Chapel Hill Auditory Development
More information! Can hear whistle? ! Where are we on course map? ! What we did in lab last week. ! Psychoacoustics
2/14/18 Can hear whistle? Lecture 5 Psychoacoustics Based on slides 2009--2018 DeHon, Koditschek Additional Material 2014 Farmer 1 2 There are sounds we cannot hear Depends on frequency Where are we on
More informationSpeech (Sound) Processing
7 Speech (Sound) Processing Acoustic Human communication is achieved when thought is transformed through language into speech. The sounds of speech are initiated by activity in the central nervous system,
More informationSpeech Spectra and Spectrograms
ACOUSTICS TOPICS ACOUSTICS SOFTWARE SPH301 SLP801 RESOURCE INDEX HELP PAGES Back to Main "Speech Spectra and Spectrograms" Page Speech Spectra and Spectrograms Robert Mannell 6. Some consonant spectra
More informationTactual Display of Consonant Voicing to Supplement Lipreading
Tactual Display of Consonant Voicing to Supplement Lipreading by Hanfeng Yuan S.B., Shanghai Jiao Tong University (1995) S.M. Massachusetts Institute of Technology (1999) Submitted to the Department of
More informationThe role of low frequency components in median plane localization
Acoust. Sci. & Tech. 24, 2 (23) PAPER The role of low components in median plane localization Masayuki Morimoto 1;, Motoki Yairi 1, Kazuhiro Iida 2 and Motokuni Itoh 1 1 Environmental Acoustics Laboratory,
More informationCHAPTER 1 INTRODUCTION
CHAPTER 1 INTRODUCTION 1.1 BACKGROUND Speech is the most natural form of human communication. Speech has also become an important means of human-machine interaction and the advancement in technology has
More informationApplying the summation model in audiovisual speech perception
Applying the summation model in audiovisual speech perception Kaisa Tiippana, Ilmari Kurki, Tarja Peromaa Department of Psychology and Logopedics, Faculty of Medicine, University of Helsinki, Finland kaisa.tiippana@helsinki.fi,
More informationSound localization psychophysics
Sound localization psychophysics Eric Young A good reference: B.C.J. Moore An Introduction to the Psychology of Hearing Chapter 7, Space Perception. Elsevier, Amsterdam, pp. 233-267 (2004). Sound localization:
More informationRESEARCH ON SPOKEN LANGUAGE PROCESSING Progress Report No. 27 (2005) Indiana University
AUDIOVISUAL ASYNCHRONY DETECTION RESEARCH ON SPOKEN LANGUAGE PROCESSING Progress Report No. 27 (2005) Indiana University Audiovisual Asynchrony Detection and Speech Perception in Normal-Hearing Listeners
More informationTopic 4. Pitch & Frequency
Topic 4 Pitch & Frequency A musical interlude KOMBU This solo by Kaigal-ool of Huun-Huur-Tu (accompanying himself on doshpuluur) demonstrates perfectly the characteristic sound of the Xorekteer voice An
More informationResearch Article The Acoustic and Peceptual Effects of Series and Parallel Processing
Hindawi Publishing Corporation EURASIP Journal on Advances in Signal Processing Volume 9, Article ID 6195, pages doi:1.1155/9/6195 Research Article The Acoustic and Peceptual Effects of Series and Parallel
More informationRESEARCH ON SPOKEN LANGUAGE PROCESSING Progress Report No. 24 (2000) Indiana University
COMPARISON OF PARTIAL INFORMATION RESEARCH ON SPOKEN LANGUAGE PROCESSING Progress Report No. 24 (2000) Indiana University Use of Partial Stimulus Information by Cochlear Implant Patients and Normal-Hearing
More informationComputational Perception /785. Auditory Scene Analysis
Computational Perception 15-485/785 Auditory Scene Analysis A framework for auditory scene analysis Auditory scene analysis involves low and high level cues Low level acoustic cues are often result in
More informationPerceived Audiovisual Simultaneity in Speech by Musicians and Non-musicians: Preliminary Behavioral and Event-Related Potential (ERP) Findings
The 14th International Conference on Auditory-Visual Speech Processing 25-26 August 2017, Stockholm, Sweden Perceived Audiovisual Simultaneity in Speech by Musicians and Non-musicians: Preliminary Behavioral
More informationWhat you re in for. Who are cochlear implants for? The bottom line. Speech processing schemes for
What you re in for Speech processing schemes for cochlear implants Stuart Rosen Professor of Speech and Hearing Science Speech, Hearing and Phonetic Sciences Division of Psychology & Language Sciences
More informationEvaluating the role of spectral and envelope characteristics in the intelligibility advantage of clear speech
Evaluating the role of spectral and envelope characteristics in the intelligibility advantage of clear speech Jean C. Krause a and Louis D. Braida Research Laboratory of Electronics, Massachusetts Institute
More informationVisual impairment and speech understanding in older adults with acquired hearing loss
Visual impairment and speech understanding in older adults with acquired hearing loss Jean-Pierre Gagné École d orthophonie et d audiologie, Université de Montréal, Montréal, Québec, Canada H3C3J7 jean-pierre.gagne@umontreal.ca
More informationTwenty subjects (11 females) participated in this study. None of the subjects had
SUPPLEMENTARY METHODS Subjects Twenty subjects (11 females) participated in this study. None of the subjects had previous exposure to a tone language. Subjects were divided into two groups based on musical
More informationAn Update on Auditory Neuropathy Spectrum Disorder in Children
An Update on Auditory Neuropathy Spectrum Disorder in Children Gary Rance PhD The University of Melbourne Sound Foundations Through Early Amplification Meeting, Chicago, Dec 2013 Overview Auditory neuropathy
More informationSimulations of high-frequency vocoder on Mandarin speech recognition for acoustic hearing preserved cochlear implant
INTERSPEECH 2017 August 20 24, 2017, Stockholm, Sweden Simulations of high-frequency vocoder on Mandarin speech recognition for acoustic hearing preserved cochlear implant Tsung-Chen Wu 1, Tai-Shih Chi
More informationWIDEXPRESS. no.30. Background
WIDEXPRESS no. january 12 By Marie Sonne Kristensen Petri Korhonen Using the WidexLink technology to improve speech perception Background For most hearing aid users, the primary motivation for using hearing
More informationThe role of periodicity in the perception of masked speech with simulated and real cochlear implants
The role of periodicity in the perception of masked speech with simulated and real cochlear implants Kurt Steinmetzger and Stuart Rosen UCL Speech, Hearing and Phonetic Sciences Heidelberg, 09. November
More informationGrozdana Erjavec 1, Denis Legros 1. Department of Cognition, Language and Interaction, University of Paris VIII, France. Abstract. 1.
INTERSPEECH 2013 Effects of Mouth-Only and Whole-Face Displays on Audio-Visual Speech Perception in Noise: Is the Vision of a Talker s Full Face Truly the Most Efficient Solution? Grozdana Erjavec 1, Denis
More informationCONSTRUCTING TELEPHONE ACOUSTIC MODELS FROM A HIGH-QUALITY SPEECH CORPUS
CONSTRUCTING TELEPHONE ACOUSTIC MODELS FROM A HIGH-QUALITY SPEECH CORPUS Mitchel Weintraub and Leonardo Neumeyer SRI International Speech Research and Technology Program Menlo Park, CA, 94025 USA ABSTRACT
More informationFitting Frequency Compression Hearing Aids to Kids: The Basics
Fitting Frequency Compression Hearing Aids to Kids: The Basics Presenters: Susan Scollie and Danielle Glista Presented at AudiologyNow! 2011, Chicago Support This work was supported by: Canadian Institutes
More informationWho are cochlear implants for?
Who are cochlear implants for? People with little or no hearing and little conductive component to the loss who receive little or no benefit from a hearing aid. Implants seem to work best in adults who
More informationVariability in Word Recognition by Adults with Cochlear Implants: The Role of Language Knowledge
Variability in Word Recognition by Adults with Cochlear Implants: The Role of Language Knowledge Aaron C. Moberly, M.D. CI2015 Washington, D.C. Disclosures ASA and ASHFoundation Speech Science Research
More informationCategorical Perception
Categorical Perception Discrimination for some speech contrasts is poor within phonetic categories and good between categories. Unusual, not found for most perceptual contrasts. Influenced by task, expectations,
More informationSpeech conveys not only linguistic content but. Vocal Emotion Recognition by Normal-Hearing Listeners and Cochlear Implant Users
Cochlear Implants Special Issue Article Vocal Emotion Recognition by Normal-Hearing Listeners and Cochlear Implant Users Trends in Amplification Volume 11 Number 4 December 2007 301-315 2007 Sage Publications
More informationThe development of a modified spectral ripple test
The development of a modified spectral ripple test Justin M. Aronoff a) and David M. Landsberger Communication and Neuroscience Division, House Research Institute, 2100 West 3rd Street, Los Angeles, California
More informationEffects of slow- and fast-acting compression on hearing impaired listeners consonantvowel identification in interrupted noise
Downloaded from orbit.dtu.dk on: Jan 01, 2018 Effects of slow- and fast-acting compression on hearing impaired listeners consonantvowel identification in interrupted noise Kowalewski, Borys; Zaar, Johannes;
More informationEffects of noise and filtering on the intelligibility of speech produced during simultaneous communication
Journal of Communication Disorders 37 (2004) 505 515 Effects of noise and filtering on the intelligibility of speech produced during simultaneous communication Douglas J. MacKenzie a,*, Nicholas Schiavetti
More informationDiagnosis and Management of ANSD: Outcomes of Cochlear Implants versus Hearing Aids
Diagnosis and Management of ANSD: Outcomes of Cochlear Implants versus Hearing Aids Gary Rance PhD The University of Melbourne International Paediatric Conference, Shanghai, April 214 Auditory Neuropathy
More informationAlthough considerable work has been conducted on the speech
Influence of Hearing Loss on the Perceptual Strategies of Children and Adults Andrea L. Pittman Patricia G. Stelmachowicz Dawna E. Lewis Brenda M. Hoover Boys Town National Research Hospital Omaha, NE
More informationAudio-Visual Integration in Multimodal Communication
Audio-Visual Integration in Multimodal Communication TSUHAN CHEN, MEMBER, IEEE, AND RAM R. RAO Invited Paper In this paper, we review recent research that examines audiovisual integration in multimodal
More informationThe functional importance of age-related differences in temporal processing
Kathy Pichora-Fuller The functional importance of age-related differences in temporal processing Professor, Psychology, University of Toronto Adjunct Scientist, Toronto Rehabilitation Institute, University
More informationComment by Delgutte and Anna. A. Dreyer (Eaton-Peabody Laboratory, Massachusetts Eye and Ear Infirmary, Boston, MA)
Comments Comment by Delgutte and Anna. A. Dreyer (Eaton-Peabody Laboratory, Massachusetts Eye and Ear Infirmary, Boston, MA) Is phase locking to transposed stimuli as good as phase locking to low-frequency
More informationRecognition of Multiply Degraded Speech by Young and Elderly Listeners
Journal of Speech and Hearing Research, Volume 38, 1150-1156, October 1995 Recognition of Multiply Degraded Speech by Young and Elderly Listeners Sandra Gordon-Salant University of Maryland College Park
More informationBinaural Hearing for Robots Introduction to Robot Hearing
Binaural Hearing for Robots Introduction to Robot Hearing 1Radu Horaud Binaural Hearing for Robots 1. Introduction to Robot Hearing 2. Methodological Foundations 3. Sound-Source Localization 4. Machine
More informationSound Texture Classification Using Statistics from an Auditory Model
Sound Texture Classification Using Statistics from an Auditory Model Gabriele Carotti-Sha Evan Penn Daniel Villamizar Electrical Engineering Email: gcarotti@stanford.edu Mangement Science & Engineering
More informationBinaural Hearing. Why two ears? Definitions
Binaural Hearing Why two ears? Locating sounds in space: acuity is poorer than in vision by up to two orders of magnitude, but extends in all directions. Role in alerting and orienting? Separating sound
More informationEssential feature. Who are cochlear implants for? People with little or no hearing. substitute for faulty or missing inner hair
Who are cochlear implants for? Essential feature People with little or no hearing and little conductive component to the loss who receive little or no benefit from a hearing aid. Implants seem to work
More information