Automatic audio analysis for content description & indexing

Size: px
Start display at page:

Download "Automatic audio analysis for content description & indexing"

Transcription

1 Automatic audio analysis for content description & indexing Dan Ellis International Computer Science Institute, Berkeley CA Outline Auditory Scene Analysis (ASA) Computational ASA (CASA) Prediction-driven CASA Speech recognition & sound mixtures Implications for content analysis Audio Indexing - Dan Ellis 1998feb04-1

2 1 Auditory Scene Analysis The organization of complex sound scenes according to their inferred sources Sounds rarely occur in isolation - organization required for useful information Human audition is very effective - unexpectedly difficult to model Correct analysis defined by goal - source shows independence, continuity ecological constraints enable organization f/hz city time/s db Audio Indexing - Dan Ellis 1998feb04-2

3 Psychology of ASA Extensive experimental research - organization of simple pieces (sinusoids & white noise) - streaming, pitch perception, double vowels Auditory Scene Analysis [Bregman 1990] grouping rules - common onset/offset/modulation, harmonicity, spatial location Debated... (Darwin, Carlyon, Moore, Remez) (from Darwin 1996) Audio Indexing - Dan Ellis 1998feb04-3

4 2 Computational Auditory Scene Analysis (CASA) Automatic sound organization? - convert an undifferentiated signal into a description in terms of different sources f/hz city horn car noise door crash horn yell time/s Translate psych. rules into programs? - representations to reveal common onset, harmonicity... Motivations & Applications - it s a puzzle: new processing principles? - real-world interactive systems (speech, robots) - hearing prostheses (enhancement, description) - advanced processing (remixing) - multimedia indexing... Audio Indexing - Dan Ellis 1998feb04-4 db

5 CASA survey Early work on co-channel speech - listeners benefit from pitch difference - algorithms for separating periodicities Utterance-sized signals need more - cannot predict number of signals (0, 1, 2...) - birth/death processes Ultimately, more constraints needed - nonperiodic signals - masked cues - ambiguous signals Audio Indexing - Dan Ellis 1998feb04-5

6 CASA1: Periodic pieces Weintraub separate male & female voices - find periodicities in each frequency channel by auto-coincidence - number of voices is hidden state Cooke & Brown (1991-3) - divide time-frequency plane into elements - apply grouping rules to form sources - pull single periodic target out of noise frq/hz brn1h.aif frq/hz brn1h.fi.aif time/s time/s Audio Indexing - Dan Ellis 1998feb04-6

7 CASA2: Hypothesis systems Okuno et al. (1994-) - tracers follow each harmonic + noise agent - residue-driven: account for whole signal Klassner search for a combination of templates - high-level hypotheses permit front-end tuning 3760 Hz 2540 Hz 2230 Hz 1475 Hz 500 Hz 460 Hz 420 Hz Buzzer-Alarm Glass-Clink Phone-Ring Siren-Chirp 2350 Hz 1675 Hz 950 Hz sec TIME (a) sec TIME (b) Ellis model for events perceived in dense scenes - prediction-driven: observations - hypotheses Audio Indexing - Dan Ellis 1998feb04-7

8 CASA3: Other approaches Blind source separation (Bell & Sejnowski) - find exact separation parameters by maximizing statistic e.g. signal independence HMM decomposition (RK Moore) - recover combined source states directly Neural models (Malsburg, Wang & Brown) - avoid implausible AI methods (search, lists) - oscillators substitute for iteration? Audio Indexing - Dan Ellis 1998feb04-8

9 3 Prediction-driven CASA Perception is not direct but a search for plausible hypotheses Data-driven... input mixture Front end signal features Object formation discrete objects Grouping rules Source groups vs. Prediction-driven input mixture Front end signal features Hypothesis management prediction errors Compare & reconcile hypotheses Noise components Periodic components predicted features Predict & combine Motivations - detect non-tonal events (noise & clicks) - support restoration illusions... hooks for high-level knowledge + complete explanation, multiple hypotheses, resynthesis Audio Indexing - Dan Ellis 1998feb04-9

10 Analyzing the continuity illusion Interrupted tone heard as continuous -.. if the interruption could be a masker f/hz ptshort time/s Data-driven just sees gaps Prediction-driven can accommodate - special case or general principle? Audio Indexing - Dan Ellis 1998feb04-10

11 Phonemic Restoration (Warren 1970) Another illusion instance Inference relies on high-level semantics frq/hz 3500 nsoffee.aif time/s Incorporating knowledge into models? Audio Indexing - Dan Ellis 1998feb04-11

12 Subjective ground-truth in mixtures? Listening tests collect perceived events : Consistent answers: f/hz City Horn1 (10/10) S9 horn 2 S10 car horn S4 horn1 S6 double horn S2 first double horn S7 horn S7 horn2 S3 1st horn S5 Honk S8 car horns S1 honk, honk Crash (10/10) S7 gunshot S8 large object crash S6 slam S9 door Slam? S2 crash S4 crash S10 door slamming S5 Trash can S3 crash (not car) S1 slam Horn2 (5/10) S9 horn 5 S8 car horns S2 horn during crash S6 doppler horn S7 horn3 Truck (7/10) S8 truck engine S2 truck accelerating S5 Acceleration S1 rev up/passing S6 acceleration S3 closeup car S10 wheels on road Audio Indexing - Dan Ellis 1998feb04-12

13 PDCASA example: City-street ambience f/hz City f/hz Wefts Horn1 (10/10) Horn2 (5/10) Horn3 (5/10) Horn4 (8/10) Horn5 (10/10) Weft5 Wefts6,7 Weft8 Wefts9 12 f/hz Noise2,Click1 0 0 Crash (10/10) Squeal (6/10) Truck (7/10) f/hz Noise time/s db Problems - error allocation - rating hypotheses - source hierarchy - resynthesis Audio Indexing - Dan Ellis 1998feb04-13

14 4 Speech recognition & sound mixtures Conventional speech recognition: signal Feature extraction low-dim. features Phoneme classifier phoneme probabilities HMM decoder words - signal assumed entirely speech - find valid labelling by discrete labels - class models from training data Some problems: - need to ignore lexically-irrelevant variation (microphone, voice pitch etc.) - compact feature space everything speech-like Very fragile to nonspeech, background - scene-analysis methods very attractive... Audio Indexing - Dan Ellis 1998feb04-14

15 CASA for speech recognition Data-driven: CASA as preprocessor - problems with holes (but: Cooke, Okuno) - doesn t exploit knowledge of speech structure Prediction-driven: speech as component - same reconciliation of speech hypotheses - need to express predictions in signal domain Speech components Hypothesis management Noise components Predict & combine input mixture Front end Compare & reconcile Periodic components Audio Indexing - Dan Ellis 1998feb04-15

16 Example of speech & nonspeech (a) Clap (clap8k env.pf) f/bark db (b) Speech plus clap (223cl env.pf) (c) Recognizer output h# w n ay n tclt uw f ay ah s ay ow h# v s eh v ah n h# h# n ay n t uw f ay v ow h# s eh v ax n tcl <SIL> nine two five oh <SIL> seven (d) Reconstruction from labels alone (223cl renvg.pf) (e) Slowly varying portion of original (223cl envg.pf) (f) Predicted speech element ( = (d)+(e) ) (223cl renv.pf) (g) Click5 from nonspeech analysis (justclick.pf) (h) Spurious elements from nonspeech analysis (nonclicks.pf) Problems: - undoing classification & normalization - finding a starting hypothesis - granularity of integration Audio Indexing - Dan Ellis 1998feb04-16

17 5 Implications for content analysis: Using CASA to index soundtracks f/hz city horn car noise door crash horn yell time/s db What are the objects in a soundtrack? - subjective definition need auditory model Segmentation vs. classification - low-level cues locate events - higher-level learned knowledge to give semantic label (footstep, crash)... AI complete? But: hard to separate - illusion phenomena suggest auditory organization depends on interpretation Audio Indexing - Dan Ellis 1998feb04-17

18 Using speech recognition for indexing Active research area: Access to news broadcast databases - e.g. Informedia (CMU), ThisL (BBC+...) - use LV-CSR to transcribe, then text-retrieval to find % word error rate, still works OK Several systems at NIST TREC workshop Tricks to ignore nonspeech/poor speech Audio Indexing - Dan Ellis 1998feb04-18

19 Open issues in automatic indexing How to do ASA? Explanation/description hierarchy - PDCASA: generic primitives + constraining hierarchy - subjective & task-dependent Classification - connecting subjective & objective properties finding subjective invariants, prominence - representation of sound-object classes Resynthesis? - a good description should be adequate - provided in PDCASA, but low quality - requires good knowledge-based constraints Audio Indexing - Dan Ellis 1998feb04-19

20 6 Conclusions Auditory organization is required in real environments We don t know how listeners do it! - plenty of modeling interest Prediction-reconciliation can account for illusions - use knowledge when signal is inadequate - important in a wider range of circumstances? Speech recognizers are a good source of knowledge Automatic indexing implies synthetic listener - need to solve a lot of modeling issues Audio Indexing - Dan Ellis 1998feb04-20

Computational Auditory Scene Analysis: An overview and some observations. CASA survey. Other approaches

Computational Auditory Scene Analysis: An overview and some observations. CASA survey. Other approaches CASA talk - Haskins/NUWC - Dan Ellis 1997oct24/5-1 The importance of auditory illusions for artificial listeners 1 Dan Ellis International Computer Science Institute, Berkeley CA

More information

Computational Auditory Scene Analysis: Principles, Practice and Applications

Computational Auditory Scene Analysis: Principles, Practice and Applications Computational Auditory Scene Analysis: Principles, Practice and Applications Dan Ellis International Computer Science Institute, Berkeley CA Outline 1 2 3 4 5 Auditory Scene Analysis

More information

Auditory Scene Analysis: phenomena, theories and computational models

Auditory Scene Analysis: phenomena, theories and computational models Auditory Scene Analysis: phenomena, theories and computational models July 1998 Dan Ellis International Computer Science Institute, Berkeley CA Outline 1 2 3 4 The computational

More information

Recognition & Organization of Speech and Audio

Recognition & Organization of Speech and Audio Recognition & Organization of Speech and Audio Dan Ellis Electrical Engineering, Columbia University http://www.ee.columbia.edu/~dpwe/ Outline 1 2 3 4 5 Introducing Tandem modeling

More information

Recognition & Organization of Speech and Audio

Recognition & Organization of Speech and Audio Recognition & Organization of Speech and Audio Dan Ellis Electrical Engineering, Columbia University Outline 1 2 3 4 5 Sound organization Background & related work Existing projects

More information

Sound, Mixtures, and Learning

Sound, Mixtures, and Learning Sound, Mixtures, and Learning Dan Ellis Laboratory for Recognition and Organization of Speech and Audio (LabROSA) Electrical Engineering, Columbia University http://labrosa.ee.columbia.edu/

More information

Recognition & Organization of Speech and Audio

Recognition & Organization of Speech and Audio Recognition & Organization of Speech and Audio Dan Ellis Electrical Engineering, Columbia University http://www.ee.columbia.edu/~dpwe/ Outline 1 2 3 4 Introducing Robust speech recognition

More information

Sound, Mixtures, and Learning

Sound, Mixtures, and Learning Sound, Mixtures, and Learning Dan Ellis oratory for Recognition and Organization of Speech and Audio () Electrical Engineering, Columbia University http://labrosa.ee.columbia.edu/

More information

Recognition & Organization of Speech & Audio

Recognition & Organization of Speech & Audio Recognition & Organization of Speech & Audio Dan Ellis http://labrosa.ee.columbia.edu/ Outline 1 2 3 Introducing Projects in speech, music & audio Summary overview - Dan Ellis 21-9-28-1 1 Sound organization

More information

Using Source Models in Speech Separation

Using Source Models in Speech Separation Using Source Models in Speech Separation Dan Ellis Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Eng., Columbia Univ., NY USA dpwe@ee.columbia.edu http://labrosa.ee.columbia.edu/

More information

General Soundtrack Analysis

General Soundtrack Analysis General Soundtrack Analysis Dan Ellis oratory for Recognition and Organization of Speech and Audio () Electrical Engineering, Columbia University http://labrosa.ee.columbia.edu/

More information

Computational Perception /785. Auditory Scene Analysis

Computational Perception /785. Auditory Scene Analysis Computational Perception 15-485/785 Auditory Scene Analysis A framework for auditory scene analysis Auditory scene analysis involves low and high level cues Low level acoustic cues are often result in

More information

Modulation and Top-Down Processing in Audition

Modulation and Top-Down Processing in Audition Modulation and Top-Down Processing in Audition Malcolm Slaney 1,2 and Greg Sell 2 1 Yahoo! Research 2 Stanford CCRMA Outline The Non-Linear Cochlea Correlogram Pitch Modulation and Demodulation Information

More information

Auditory Scene Analysis

Auditory Scene Analysis 1 Auditory Scene Analysis Albert S. Bregman Department of Psychology McGill University 1205 Docteur Penfield Avenue Montreal, QC Canada H3A 1B1 E-mail: bregman@hebb.psych.mcgill.ca To appear in N.J. Smelzer

More information

Using Speech Models for Separation

Using Speech Models for Separation Using Speech Models for Separation Dan Ellis Comprising the work of Michael Mandel and Ron Weiss Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Eng., Columbia Univ., NY

More information

Auditory scene analysis in humans: Implications for computational implementations.

Auditory scene analysis in humans: Implications for computational implementations. Auditory scene analysis in humans: Implications for computational implementations. Albert S. Bregman McGill University Introduction. The scene analysis problem. Two dimensions of grouping. Recognition

More information

Robustness, Separation & Pitch

Robustness, Separation & Pitch Robustness, Separation & Pitch or Morgan, Me & Pitch Dan Ellis Columbia / ICSI dpwe@ee.columbia.edu http://labrosa.ee.columbia.edu/ 1. Robustness and Separation 2. An Academic Journey 3. Future COLUMBIA

More information

Lecture 4: Auditory Perception. Why study perception?

Lecture 4: Auditory Perception. Why study perception? EE E682: Speech & Audio Processing & Recognition Lecture 4: Auditory Perception 1 2 3 4 5 6 Motivation: Why & how Auditory physiology Psychophysics: Detection & discrimination Pitch perception Speech perception

More information

HCS 7367 Speech Perception

HCS 7367 Speech Perception Babies 'cry in mother's tongue' HCS 7367 Speech Perception Dr. Peter Assmann Fall 212 Babies' cries imitate their mother tongue as early as three days old German researchers say babies begin to pick up

More information

USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES

USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES Varinthira Duangudom and David V Anderson School of Electrical and Computer Engineering, Georgia Institute of Technology Atlanta, GA 30332

More information

HCS 7367 Speech Perception

HCS 7367 Speech Perception Long-term spectrum of speech HCS 7367 Speech Perception Connected speech Absolute threshold Males Dr. Peter Assmann Fall 212 Females Long-term spectrum of speech Vowels Males Females 2) Absolute threshold

More information

Auditory gist perception and attention

Auditory gist perception and attention Auditory gist perception and attention Sue Harding Speech and Hearing Research Group University of Sheffield POP Perception On Purpose Since the Sheffield POP meeting: Paper: Auditory gist perception:

More information

Lecture 9: Speech Recognition: Front Ends

Lecture 9: Speech Recognition: Front Ends EE E682: Speech & Audio Processing & Recognition Lecture 9: Speech Recognition: Front Ends 1 2 Recognizing Speech Feature Calculation Dan Ellis http://www.ee.columbia.edu/~dpwe/e682/

More information

Auditory Scene Analysis in Humans and Machines

Auditory Scene Analysis in Humans and Machines Auditory Scene Analysis in Humans and Machines Dan Ellis Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Eng., Columbia Univ., NY USA dpwe@ee.columbia.edu http://labrosa.ee.columbia.edu/

More information

INTRODUCTION J. Acoust. Soc. Am. 103 (2), February /98/103(2)/1080/5/$ Acoustical Society of America 1080

INTRODUCTION J. Acoust. Soc. Am. 103 (2), February /98/103(2)/1080/5/$ Acoustical Society of America 1080 Perceptual segregation of a harmonic from a vowel by interaural time difference in conjunction with mistuning and onset asynchrony C. J. Darwin and R. W. Hukin Experimental Psychology, University of Sussex,

More information

AUDL GS08/GAV1 Signals, systems, acoustics and the ear. Pitch & Binaural listening

AUDL GS08/GAV1 Signals, systems, acoustics and the ear. Pitch & Binaural listening AUDL GS08/GAV1 Signals, systems, acoustics and the ear Pitch & Binaural listening Review 25 20 15 10 5 0-5 100 1000 10000 25 20 15 10 5 0-5 100 1000 10000 Part I: Auditory frequency selectivity Tuning

More information

Auditory Scene Analysis. Dr. Maria Chait, UCL Ear Institute

Auditory Scene Analysis. Dr. Maria Chait, UCL Ear Institute Auditory Scene Analysis Dr. Maria Chait, UCL Ear Institute Expected learning outcomes: Understand the tasks faced by the auditory system during everyday listening. Know the major Gestalt principles. Understand

More information

Role of F0 differences in source segregation

Role of F0 differences in source segregation Role of F0 differences in source segregation Andrew J. Oxenham Research Laboratory of Electronics, MIT and Harvard-MIT Speech and Hearing Bioscience and Technology Program Rationale Many aspects of segregation

More information

Lecture 3: Perception

Lecture 3: Perception ELEN E4896 MUSIC SIGNAL PROCESSING Lecture 3: Perception 1. Ear Physiology 2. Auditory Psychophysics 3. Pitch Perception 4. Music Perception Dan Ellis Dept. Electrical Engineering, Columbia University

More information

Chapter 16: Chapter Title 221

Chapter 16: Chapter Title 221 Chapter 16: Chapter Title 221 Chapter 16 The History and Future of CASA Malcolm Slaney IBM Almaden Research Center malcolm@ieee.org 1 INTRODUCTION In this chapter I briefly review the history and the future

More information

A Consumer-friendly Recap of the HLAA 2018 Research Symposium: Listening in Noise Webinar

A Consumer-friendly Recap of the HLAA 2018 Research Symposium: Listening in Noise Webinar A Consumer-friendly Recap of the HLAA 2018 Research Symposium: Listening in Noise Webinar Perry C. Hanavan, AuD Augustana University Sioux Falls, SD August 15, 2018 Listening in Noise Cocktail Party Problem

More information

Sensation and Perception -- Team Problem Solving Assignments

Sensation and Perception -- Team Problem Solving Assignments Sensation and Perception -- Team Solving Assignments Directions: As a group choose 3 of the following reports to complete. The combined final report must be typed and is due, Monday, October 17. 1. Dark

More information

group by pitch: similar frequencies tend to be grouped together - attributed to a common source.

group by pitch: similar frequencies tend to be grouped together - attributed to a common source. Pattern perception Section 1 - Auditory scene analysis Auditory grouping: the sound wave hitting out ears is often pretty complex, and contains sounds from multiple sources. How do we group sounds together

More information

= + Auditory Scene Analysis. Week 9. The End. The auditory scene. The auditory scene. Otherwise known as

= + Auditory Scene Analysis. Week 9. The End. The auditory scene. The auditory scene. Otherwise known as Auditory Scene Analysis Week 9 Otherwise known as Auditory Grouping Auditory Streaming Sound source segregation The auditory scene The auditory system needs to make sense of the superposition of component

More information

! Can hear whistle? ! Where are we on course map? ! What we did in lab last week. ! Psychoacoustics

! Can hear whistle? ! Where are we on course map? ! What we did in lab last week. ! Psychoacoustics 2/14/18 Can hear whistle? Lecture 5 Psychoacoustics Based on slides 2009--2018 DeHon, Koditschek Additional Material 2014 Farmer 1 2 There are sounds we cannot hear Depends on frequency Where are we on

More information

Acoustics, signals & systems for audiology. Psychoacoustics of hearing impairment

Acoustics, signals & systems for audiology. Psychoacoustics of hearing impairment Acoustics, signals & systems for audiology Psychoacoustics of hearing impairment Three main types of hearing impairment Conductive Sound is not properly transmitted from the outer to the inner ear Sensorineural

More information

Hearing in the Environment

Hearing in the Environment 10 Hearing in the Environment Click Chapter to edit 10 Master Hearing title in the style Environment Sound Localization Complex Sounds Auditory Scene Analysis Continuity and Restoration Effects Auditory

More information

ELL 788 Computational Perception & Cognition July November 2015

ELL 788 Computational Perception & Cognition July November 2015 ELL 788 Computational Perception & Cognition July November 2015 Module 8 Audio and Multimodal Attention Audio Scene Analysis Two-stage process Segmentation: decomposition to time-frequency segments Grouping

More information

Date: April 19, 2017 Name of Product: Cisco Spark Board Contact for more information:

Date: April 19, 2017 Name of Product: Cisco Spark Board Contact for more information: Date: April 19, 2017 Name of Product: Cisco Spark Board Contact for more information: accessibility@cisco.com Summary Table - Voluntary Product Accessibility Template Criteria Supporting Features Remarks

More information

2/25/2013. Context Effect on Suprasegmental Cues. Supresegmental Cues. Pitch Contour Identification (PCI) Context Effect with Cochlear Implants

2/25/2013. Context Effect on Suprasegmental Cues. Supresegmental Cues. Pitch Contour Identification (PCI) Context Effect with Cochlear Implants Context Effect on Segmental and Supresegmental Cues Preceding context has been found to affect phoneme recognition Stop consonant recognition (Mann, 1980) A continuum from /da/ to /ga/ was preceded by

More information

Sound Interfaces Engineering Interaction Technologies. Prof. Stefanie Mueller HCI Engineering Group

Sound Interfaces Engineering Interaction Technologies. Prof. Stefanie Mueller HCI Engineering Group Sound Interfaces 6.810 Engineering Interaction Technologies Prof. Stefanie Mueller HCI Engineering Group what is sound? if a tree falls in the forest and nobody is there does it make sound?

More information

J Jeffress model, 3, 66ff

J Jeffress model, 3, 66ff Index A Absolute pitch, 102 Afferent projections, inferior colliculus, 131 132 Amplitude modulation, coincidence detector, 152ff inferior colliculus, 152ff inhibition models, 156ff models, 152ff Anatomy,

More information

Noise-Robust Speech Recognition Technologies in Mobile Environments

Noise-Robust Speech Recognition Technologies in Mobile Environments Noise-Robust Speech Recognition echnologies in Mobile Environments Mobile environments are highly influenced by ambient noise, which may cause a significant deterioration of speech recognition performance.

More information

The role of periodicity in the perception of masked speech with simulated and real cochlear implants

The role of periodicity in the perception of masked speech with simulated and real cochlear implants The role of periodicity in the perception of masked speech with simulated and real cochlear implants Kurt Steinmetzger and Stuart Rosen UCL Speech, Hearing and Phonetic Sciences Heidelberg, 09. November

More information

Hear Better With FM. Get more from everyday situations. Life is on

Hear Better With FM. Get more from everyday situations. Life is on Hear Better With FM Get more from everyday situations Life is on We are sensitive to the needs of everyone who depends on our knowledge, ideas and care. And by creatively challenging the limits of technology,

More information

Auditory Physiology PSY 310 Greg Francis. Lecture 30. Organ of Corti

Auditory Physiology PSY 310 Greg Francis. Lecture 30. Organ of Corti Auditory Physiology PSY 310 Greg Francis Lecture 30 Waves, waves, waves. Organ of Corti Tectorial membrane Sits on top Inner hair cells Outer hair cells The microphone for the brain 1 Hearing Perceptually,

More information

iclicker+ Student Remote Voluntary Product Accessibility Template (VPAT)

iclicker+ Student Remote Voluntary Product Accessibility Template (VPAT) iclicker+ Student Remote Voluntary Product Accessibility Template (VPAT) Date: May 22, 2017 Product Name: iclicker+ Student Remote Product Model Number: RLR15 Company Name: Macmillan Learning, iclicker

More information

Tandem modeling investigations

Tandem modeling investigations Tandem modeling investigations Dan Ellis International Computer Science Institute, Berkeley CA Outline 1 2 3 What makes Tandem successful? Can we make Tandem better? Does Tandem

More information

Product Model #:ASTRO Digital Spectra Consolette W7 Models (Local Control)

Product Model #:ASTRO Digital Spectra Consolette W7 Models (Local Control) Subpart 1194.25 Self-Contained, Closed Products When a timed response is required alert user, allow sufficient time for him to indicate that he needs additional time to respond [ N/A ] For touch screen

More information

Speech recognition in noisy environments: A survey

Speech recognition in noisy environments: A survey T-61.182 Robustness in Language and Speech Processing Speech recognition in noisy environments: A survey Yifan Gong presented by Tapani Raiko Feb 20, 2003 About the Paper Article published in Speech Communication

More information

IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 15, NO. 5, SEPTEMBER A Computational Model of Auditory Selective Attention

IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 15, NO. 5, SEPTEMBER A Computational Model of Auditory Selective Attention IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 15, NO. 5, SEPTEMBER 2004 1151 A Computational Model of Auditory Selective Attention Stuart N. Wrigley, Member, IEEE, and Guy J. Brown Abstract The human auditory

More information

Advanced Audio Interface for Phonetic Speech. Recognition in a High Noise Environment

Advanced Audio Interface for Phonetic Speech. Recognition in a High Noise Environment DISTRIBUTION STATEMENT A Approved for Public Release Distribution Unlimited Advanced Audio Interface for Phonetic Speech Recognition in a High Noise Environment SBIR 99.1 TOPIC AF99-1Q3 PHASE I SUMMARY

More information

Real-World Sound Recognition: A Recipe

Real-World Sound Recognition: A Recipe Real-World Sound Recognition: A Recipe Tjeerd C. Andringa and Maria E. Niessen Department of Artificial Intelligence, University of Groningen, Gr. Kruisstr. 2/1, 9712 TS Groningen, The Netherlands {T.Andringa,M.Niessen}@ai.rug.nl

More information

Gender Based Emotion Recognition using Speech Signals: A Review

Gender Based Emotion Recognition using Speech Signals: A Review 50 Gender Based Emotion Recognition using Speech Signals: A Review Parvinder Kaur 1, Mandeep Kaur 2 1 Department of Electronics and Communication Engineering, Punjabi University, Patiala, India 2 Department

More information

Discrete Signal Processing

Discrete Signal Processing 1 Discrete Signal Processing C.M. Liu Perceptual Lab, College of Computer Science National Chiao-Tung University http://www.cs.nctu.edu.tw/~cmliu/courses/dsp/ ( Office: EC538 (03)5731877 cmliu@cs.nctu.edu.tw

More information

Summary Table Voluntary Product Accessibility Template. Criteria Supporting Features Remarks and explanations

Summary Table Voluntary Product Accessibility Template. Criteria Supporting Features Remarks and explanations Plantronics VPAT-12 Summary Table Voluntary Product Accessibility Template Section 1194.21 Software Applications and Operating Systems Section 1194.22 Web-based intranet and Internet Information and Applications

More information

Binaural Hearing. Why two ears? Definitions

Binaural Hearing. Why two ears? Definitions Binaural Hearing Why two ears? Locating sounds in space: acuity is poorer than in vision by up to two orders of magnitude, but extends in all directions. Role in alerting and orienting? Separating sound

More information

Consonant Perception test

Consonant Perception test Consonant Perception test Introduction The Vowel-Consonant-Vowel (VCV) test is used in clinics to evaluate how well a listener can recognize consonants under different conditions (e.g. with and without

More information

Hearing II Perceptual Aspects

Hearing II Perceptual Aspects Hearing II Perceptual Aspects Overview of Topics Chapter 6 in Chaudhuri Intensity & Loudness Frequency & Pitch Auditory Space Perception 1 2 Intensity & Loudness Loudness is the subjective perceptual quality

More information

Oticon Agil. Connectivity Matters - A Whitepaper. Agil Improves ConnectLine Sound Quality

Oticon Agil. Connectivity Matters - A Whitepaper. Agil Improves ConnectLine Sound Quality Oticon Agil Connectivity Matters - A Whitepaper Agil Improves ConnectLine Sound Quality M. Lisa Sjolander, AuD Oticon A/S, Smørum, Denmark Agil Improves ConnectLine Sound Quality Wireless technology is

More information

Product Model #: Digital Portable Radio XTS 5000 (Std / Rugged / Secure / Type )

Product Model #: Digital Portable Radio XTS 5000 (Std / Rugged / Secure / Type ) Rehabilitation Act Amendments of 1998, Section 508 Subpart 1194.25 Self-Contained, Closed Products The following features are derived from Section 508 When a timed response is required alert user, allow

More information

iclicker2 Student Remote Voluntary Product Accessibility Template (VPAT)

iclicker2 Student Remote Voluntary Product Accessibility Template (VPAT) iclicker2 Student Remote Voluntary Product Accessibility Template (VPAT) Date: May 22, 2017 Product Name: i>clicker2 Student Remote Product Model Number: RLR14 Company Name: Macmillan Learning, iclicker

More information

Speech Spatial Qualities

Speech Spatial Qualities Speech Spatial Qualities Advice about answering the questions The following questions inquire about aspects of your ability and experience hearing and listening in different situations. For each question,

More information

Auditory System & Hearing

Auditory System & Hearing Auditory System & Hearing Chapters 9 and 10 Lecture 17 Jonathan Pillow Sensation & Perception (PSY 345 / NEU 325) Spring 2015 1 Cochlea: physical device tuned to frequency! place code: tuning of different

More information

Optical Illusions 4/5. Optical Illusions 2/5. Optical Illusions 5/5 Optical Illusions 1/5. Reading. Reading. Fang Chen Spring 2004

Optical Illusions 4/5. Optical Illusions 2/5. Optical Illusions 5/5 Optical Illusions 1/5. Reading. Reading. Fang Chen Spring 2004 Optical Illusions 2/5 Optical Illusions 4/5 the Ponzo illusion the Muller Lyer illusion Optical Illusions 5/5 Optical Illusions 1/5 Mauritz Cornelis Escher Dutch 1898 1972 Graphical designer World s first

More information

Musical Instrument Classification through Model of Auditory Periphery and Neural Network

Musical Instrument Classification through Model of Auditory Periphery and Neural Network Musical Instrument Classification through Model of Auditory Periphery and Neural Network Ladislava Jankø, Lenka LhotskÆ Department of Cybernetics, Faculty of Electrical Engineering Czech Technical University

More information

HOW TO USE THE SHURE MXA910 CEILING ARRAY MICROPHONE FOR VOICE LIFT

HOW TO USE THE SHURE MXA910 CEILING ARRAY MICROPHONE FOR VOICE LIFT HOW TO USE THE SHURE MXA910 CEILING ARRAY MICROPHONE FOR VOICE LIFT Created: Sept 2016 Updated: June 2017 By: Luis Guerra Troy Jensen The Shure MXA910 Ceiling Array Microphone offers the unique advantage

More information

21/01/2013. Binaural Phenomena. Aim. To understand binaural hearing Objectives. Understand the cues used to determine the location of a sound source

21/01/2013. Binaural Phenomena. Aim. To understand binaural hearing Objectives. Understand the cues used to determine the location of a sound source Binaural Phenomena Aim To understand binaural hearing Objectives Understand the cues used to determine the location of a sound source Understand sensitivity to binaural spatial cues, including interaural

More information

Effects of Location, Frequency Region, and Time Course of Selective Attention on Auditory Scene Analysis

Effects of Location, Frequency Region, and Time Course of Selective Attention on Auditory Scene Analysis Journal of Experimental Psychology: Human Perception and Performance 2004, Vol. 30, No. 4, 643 656 Copyright 2004 by the American Psychological Association 0096-1523/04/$12.00 DOI: 10.1037/0096-1523.30.4.643

More information

Virtual Sensors: Transforming the Way We Think About Accommodation Stevens Institute of Technology-Hoboken, New Jersey Katherine Grace August, Avi

Virtual Sensors: Transforming the Way We Think About Accommodation Stevens Institute of Technology-Hoboken, New Jersey Katherine Grace August, Avi Virtual Sensors: Transforming the Way We Think About Accommodation Stevens Institute of Technology-Hoboken, New Jersey Katherine Grace August, Avi Hauser, Dave Nall Fatimah Shehadeh-Grant, Jennifer Chen,

More information

Binaural Hearing for Robots Introduction to Robot Hearing

Binaural Hearing for Robots Introduction to Robot Hearing Binaural Hearing for Robots Introduction to Robot Hearing 1Radu Horaud Binaural Hearing for Robots 1. Introduction to Robot Hearing 2. Methodological Foundations 3. Sound-Source Localization 4. Machine

More information

Effects of Cochlear Hearing Loss on the Benefits of Ideal Binary Masking

Effects of Cochlear Hearing Loss on the Benefits of Ideal Binary Masking INTERSPEECH 2016 September 8 12, 2016, San Francisco, USA Effects of Cochlear Hearing Loss on the Benefits of Ideal Binary Masking Vahid Montazeri, Shaikat Hossain, Peter F. Assmann University of Texas

More information

SPEECH ANALYSIS 3/3 PERCEPTION

SPEECH ANALYSIS 3/3 PERCEPTION NATURAL LANGUAGE PROCESSING Prof. L. Sbattella SPEECH ANALYSIS 3/3 PERCEPTION Dott. Ing. Sonia Cenceschi Sonia.cenceschi@polimi.it Psychoacoustics and perception Philip G. Zimbardo, Richard J. Gerrig,

More information

Robust Neural Encoding of Speech in Human Auditory Cortex

Robust Neural Encoding of Speech in Human Auditory Cortex Robust Neural Encoding of Speech in Human Auditory Cortex Nai Ding, Jonathan Z. Simon Electrical Engineering / Biology University of Maryland, College Park Auditory Processing in Natural Scenes How is

More information

Human cogition. Human Cognition. Optical Illusions. Human cognition. Optical Illusions. Optical Illusions

Human cogition. Human Cognition. Optical Illusions. Human cognition. Optical Illusions. Optical Illusions Human Cognition Fang Chen Chalmers University of Technology Human cogition Perception and recognition Attention, emotion Learning Reading, speaking, and listening Problem solving, planning, reasoning,

More information

Speech Spatial Qualities -C

Speech Spatial Qualities -C Speech Spatial Qualities -C Advice about answering the questions The following questions inquire about aspects of your ability and experience hearing and listening in different situations. You answered

More information

Voice Detection using Speech Energy Maximization and Silence Feature Normalization

Voice Detection using Speech Energy Maximization and Silence Feature Normalization , pp.25-29 http://dx.doi.org/10.14257/astl.2014.49.06 Voice Detection using Speech Energy Maximization and Silence Feature Normalization In-Sung Han 1 and Chan-Shik Ahn 2 1 Dept. of The 2nd R&D Institute,

More information

Competing Frameworks in Perception

Competing Frameworks in Perception Competing Frameworks in Perception Lesson II: Perception module 08 Perception.08. 1 Views on perception Perception as a cascade of information processing stages From sensation to percept Template vs. feature

More information

Competing Frameworks in Perception

Competing Frameworks in Perception Competing Frameworks in Perception Lesson II: Perception module 08 Perception.08. 1 Views on perception Perception as a cascade of information processing stages From sensation to percept Template vs. feature

More information

CONSTRUCTING TELEPHONE ACOUSTIC MODELS FROM A HIGH-QUALITY SPEECH CORPUS

CONSTRUCTING TELEPHONE ACOUSTIC MODELS FROM A HIGH-QUALITY SPEECH CORPUS CONSTRUCTING TELEPHONE ACOUSTIC MODELS FROM A HIGH-QUALITY SPEECH CORPUS Mitchel Weintraub and Leonardo Neumeyer SRI International Speech Research and Technology Program Menlo Park, CA, 94025 USA ABSTRACT

More information

VPAT Summary. VPAT Details. Section Telecommunications Products - Detail. Date: October 8, 2014 Name of Product: BladeCenter HS23

VPAT Summary. VPAT Details. Section Telecommunications Products - Detail. Date: October 8, 2014 Name of Product: BladeCenter HS23 Date: October 8, 2014 Name of Product: BladeCenter HS23 VPAT Summary Criteria Status Remarks and Explanations Section 1194.21 Software Applications and Operating Systems Section 1194.22 Web-based Internet

More information

Summary Table Voluntary Product Accessibility Template

Summary Table Voluntary Product Accessibility Template PLANTRONICS VPAT 3 Product: Wireless Hearing Aid Compatible (HAC) Headsets Operated with dedicated Base assembly Over the Head Voice Tube: Pulsar 590, CS351 Over the Head Noise-Canceling: CS351N Convertible

More information

Summary Table Voluntary Product Accessibility Template. Supporting Features. Supports. Supports. Supports. Supports

Summary Table Voluntary Product Accessibility Template. Supporting Features. Supports. Supports. Supports. Supports Date: March 31, 2016 Name of Product: ThinkServer TS450, TS550 Summary Table Voluntary Product Accessibility Template Section 1194.21 Software Applications and Operating Systems Section 1194.22 Web-based

More information

Infant Hearing Development: Translating Research Findings into Clinical Practice. Auditory Development. Overview

Infant Hearing Development: Translating Research Findings into Clinical Practice. Auditory Development. Overview Infant Hearing Development: Translating Research Findings into Clinical Practice Lori J. Leibold Department of Allied Health Sciences The University of North Carolina at Chapel Hill Auditory Development

More information

SOLUTIONS Homework #3. Introduction to Engineering in Medicine and Biology ECEN 1001 Due Tues. 9/30/03

SOLUTIONS Homework #3. Introduction to Engineering in Medicine and Biology ECEN 1001 Due Tues. 9/30/03 SOLUTIONS Homework #3 Introduction to Engineering in Medicine and Biology ECEN 1001 Due Tues. 9/30/03 Problem 1: a) Where in the cochlea would you say the process of "fourier decomposition" of the incoming

More information

Acoustic Sensing With Artificial Intelligence

Acoustic Sensing With Artificial Intelligence Acoustic Sensing With Artificial Intelligence Bowon Lee Department of Electronic Engineering Inha University Incheon, South Korea bowon.lee@inha.ac.kr bowon.lee@ieee.org NVIDIA Deep Learning Day Seoul,

More information

Categorical Perception

Categorical Perception Categorical Perception Discrimination for some speech contrasts is poor within phonetic categories and good between categories. Unusual, not found for most perceptual contrasts. Influenced by task, expectations,

More information

This American Life Transcript. Prologue. Broadcast June 25, Episode #411: First Contact. So, Scott, you were born without hearing, right?

This American Life Transcript. Prologue. Broadcast June 25, Episode #411: First Contact. So, Scott, you were born without hearing, right? Scott Krepel Interview from TAL #411 1 This American Life Transcript Prologue Broadcast June 25, 2010 Episode #411: First Contact Is that Marc? Yes, that s Marc speaking for Scott. So, Scott, you were

More information

Fitting Decisions and their Impact on Hearing Aid User Benefit. Mallory Maine, AuD Audiologist, GN ReSound

Fitting Decisions and their Impact on Hearing Aid User Benefit. Mallory Maine, AuD Audiologist, GN ReSound Fitting Decisions and their Impact on Hearing Aid User Benefit Mallory Maine, AuD Audiologist, GN ReSound Agenda Common Fitting Oversights #1 Setting the coupler type in fitting software To set or not

More information

The functional importance of age-related differences in temporal processing

The functional importance of age-related differences in temporal processing Kathy Pichora-Fuller The functional importance of age-related differences in temporal processing Professor, Psychology, University of Toronto Adjunct Scientist, Toronto Rehabilitation Institute, University

More information

Auditory System & Hearing

Auditory System & Hearing Auditory System & Hearing Chapters 9 part II Lecture 16 Jonathan Pillow Sensation & Perception (PSY 345 / NEU 325) Spring 2019 1 Phase locking: Firing locked to period of a sound wave example of a temporal

More information

Binaural processing of complex stimuli

Binaural processing of complex stimuli Binaural processing of complex stimuli Outline for today Binaural detection experiments and models Speech as an important waveform Experiments on understanding speech in complex environments (Cocktail

More information

The effect of wearing conventional and level-dependent hearing protectors on speech production in noise and quiet

The effect of wearing conventional and level-dependent hearing protectors on speech production in noise and quiet The effect of wearing conventional and level-dependent hearing protectors on speech production in noise and quiet Ghazaleh Vaziri Christian Giguère Hilmi R. Dajani Nicolas Ellaham Annual National Hearing

More information

Avaya G450 Branch Gateway, Release 7.1 Voluntary Product Accessibility Template (VPAT)

Avaya G450 Branch Gateway, Release 7.1 Voluntary Product Accessibility Template (VPAT) Avaya G450 Branch Gateway, Release 7.1 Voluntary Product Accessibility Template (VPAT) can be administered via a graphical user interface or via a text-only command line interface. The responses in this

More information

Chapter 5. Summary and Conclusions! 131

Chapter 5. Summary and Conclusions! 131 ! Chapter 5 Summary and Conclusions! 131 Chapter 5!!!! Summary of the main findings The present thesis investigated the sensory representation of natural sounds in the human auditory cortex. Specifically,

More information

Fundamentals of Cognitive Psychology, 3e by Ronald T. Kellogg Chapter 2. Multiple Choice

Fundamentals of Cognitive Psychology, 3e by Ronald T. Kellogg Chapter 2. Multiple Choice Multiple Choice 1. Which structure is not part of the visual pathway in the brain? a. occipital lobe b. optic chiasm c. lateral geniculate nucleus *d. frontal lobe Answer location: Visual Pathways 2. Which

More information

Speech Separation in Humans and Machines

Speech Separation in Humans and Machines Speech Separation in Humans and Machines Dan Ellis Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Eng., Columbia Univ., NY USA dpwe@ee.columbia.edu http://labrosa.ee.columbia.edu/

More information

Computational Models of Mammalian Hearing:

Computational Models of Mammalian Hearing: Computational Models of Mammalian Hearing: Frank Netter and his Ciba paintings An Auditory Image Approach Dick Lyon For Tom Dean s Cortex Class Stanford, April 14, 2010 Breschet 1836, Testut 1897 167 years

More information

Issues faced by people with a Sensorineural Hearing Loss

Issues faced by people with a Sensorineural Hearing Loss Issues faced by people with a Sensorineural Hearing Loss Issues faced by people with a Sensorineural Hearing Loss 1. Decreased Audibility 2. Decreased Dynamic Range 3. Decreased Frequency Resolution 4.

More information

Summary Table Voluntary Product Accessibility Template. Supporting Features Not Applicable Not Applicable. Supports with Exceptions.

Summary Table Voluntary Product Accessibility Template. Supporting Features Not Applicable Not Applicable. Supports with Exceptions. Plantronics/ Clarity Summary Table Voluntary Product Accessibility Template Criteria Section 1194.21 Software Applications and Operating Systems Section 1194.22 Web-based intranet and Internet Information

More information