Non-Verbal Eliza-like Human Behaviors in Human-Robot Interaction through Real-Time Auditory and Visual Multiple-Talker Tracking

Size: px
Start display at page:

Download "Non-Verbal Eliza-like Human Behaviors in Human-Robot Interaction through Real-Time Auditory and Visual Multiple-Talker Tracking"

Transcription

1 Non-Verbal Eliza-like Human Behaviors in Human-Robot Interaction through Real-Time Auditory and Visual Multiple-Talker Tracking Hiroshi G. Okuno y;z, Kazuhiro Nakadai z, and Hiroaki Kitano z;λ y Graduate School of Informatics, Kyoto University, Sakyo, Kyoto , Japan, okuno@nue.org z Kitano Symbiotic Systems Project, Japan Science and Technology Inc., M-31 #6A, Jingumae, Shibuya, Tokyo , Japan, nakadai@symbio.jst.go.jp Λ Sony Computer Science Labratories, Shinagawa, Tokyo, Japan, kitano@csl.sony.co.jp Abstract Sound is essential to enhance visual experience and human robot interaction, but usually most research and development efforts are made mainly towards sound generation, speech synthesis and speech recognition. The reason why only a little attention has been paid on auditory scene analysis is that real-time perception of a mixture of sounds is difficult. Recently, Nakadai et al have developed real-time auditory and visual multiple-talker tracking technology. In this paper, this technology is applied to human-robot interaction including a receptionist robot and a companion robot at a party. The system implemented on the upper-torso humanoid SIG includes face identification, speech recognition, focusof-attention control, and sensorimotor task in tracking multiple talkers. Focus-of-attention is controlled by associating auditory and visual streams with using the sound source direction and talker position as a clue. Once an association is established, the humanoid keeps its face to the direction of the associated talker. It also provides task-oriented focus-ofattention control. We demonstrate that the resulting behavior of SIG invites the users participation in interaction and encourages the users to explore SIG s behaviors. These behaviors are quite similar to those caused by Eliza. Introduction Social interaction is essential for humanoid robots, because they are getting more common in social and home environments, such as a pet robot in a living room, a service robot at office, or a robot serving people at a party (Brooks et al. 1998). Social skills of such robots require robust complex perceptual abilities; for example, it identifies people in the room, pays attention to their voice and looks at them to identify, and associates voice and visual images. Intelligent behavior of social interaction should emerge from rich channels of input sensors; vision, audition, tactile, and others. Perception of various kinds of sensory inputs should be active in the sense that we hear and see things and events that are important to us as individuals, not sound waves or light rays (Handel 1989). In other words, selective attention of sensors represented as looking versus seeing or listening versus hearing plays an important role in social interaction. Other important factors in social interaction are recognition and synthesis of emotion in face expression and voice tones (Breazeal & Scassellati 1999; Breazeal 2001). In the 21st century, autonomous robots are more common in social and home environments, such as a pet robot at living room, a service robot at office, or a robot serving people at a party. The robot shall identify people in the room, pay attention to their voice and look at them to identify visually, and associate voice and visual images, so that highly robust event identification can be accomplished. These are minimum requirements for social interaction (Brooks et al. 1998). Sound has been recently recognized as essential in order to enhance visual experience and human computer interaction, and thus not a few contributions have been done by academia and industries at IROS and its related conferences (Breazeal & Scassellati 1999; Brooks et al. 1999; Nakadai et al. 2000b). This is because people can dance with sounds but not with images (Handel 1989). Handel also pointed out the importance of selective attention by writing We hear and see things and events that are important to us as individuals, not sound waves or light rays (Handel 1989). Sound, however, has not been utilized so much as input media except speech recognition. There are at least two reasons for this tendency: 1. Handling of a mixture of sounds We hear a mixture of sounds, not a sound of single sound source. Automatic speech recognition (ASR) assumes that the input is a voiced speech and this assumption holds as long as a microphone is set close to the mouth of a speaker. Of course, speech recognition community develops robust ASR to make this assumption hold on wider fields (Hansen, Mammone, & Young 1994). 2. Real-time processing Some studies with computational auditory scene analysis (CASA) to understand a mixture of sounds has been done (Rosenthal & Okuno 1998). However, one of its critical problems in applying CASA techniques to a real-world system is a lack of real-time processing. Usually, people hear a mixture of sounds, and people with normal hearing can separate sounds from the mixture and focus on a particular voice or sound in a noisy environment. This capability is known as the cocktail party effect (Cherry 1953). Real-time processing is essential to incorporate cocktail party effect into a robot.

2 Task of Speaker Tracking Real-time object tracking is applied to some applications that will run under auditorily and visually noisy environments. For example, at a party, many people are talking and moving. In this situation, strong reverberations (echoes) occur and speeches are interfered by other sounds or talks. Not only reverberations, but also lighting of illuminating conditions change dynamically, and people are often occluded by other people and reappear. a) Cover made of FRP c) A microphone installed in an ear Figure 1: SIG the Humanoid: Its cover, mechanical structure, and a microphone Nakadai et al developed real-time auditory and visual multiple-tracking system (Nakadai et al. 2001). The key idea of their work is to integrate auditory and visual information to track several things simultaneously. In this paper, we apply the real-time auditory and visual multipletracking system to a receptionist robot and a companion robot of a party in order to demonstrate the feasibility of a cocktail party robot. The system is composed of face identification, speech separation, automatic speech recognition, speech synthesis, dialog control as well as the auditory and visual tracking. Some robots realize social interaction, in particular, in visual and dialogue processing. Ono et al. use the robot named Robovie to make common attention between human and robot by using gestures (Ono, Imai, & Ishiguro 2000). Breazeal incorporates the capabilities of recognition and synthesis of emotion in face expression and voice tones into the robot named Kismet (Breazeal & Scassellati 1999; Breazeal 2001). Waldherr et al. makes the robot named AMELLA that can recognize pose and motion gestures (Waldherr et al. 1998). Matsusaka et al. built the robot named Hadaly that can localize the talker as well as recognize speeches by speech-recognition system so that it can interact with multiple people (Matsusaka et al. 1999). Nakadai et al developed real-time auditory and visual multiple-tracking system for the upper-torso humanoid named SIG (Nakadai et al. 2001). The rest of the paper is organized as follows: Section 2 describes the design of the whole system. Section 3 describes the details of each subsystem, in particular, real-time auditory and visual multiple-tracking system. Section 4 demonstrates the system behavior. Section 5 discusses the observations of the experiments and future work, and Section 6 concludes the paper. Robots at a party To design the system for such a noisy environment and prove its feasibility, we take a robot at a party as an example. Its task is the following two cases: 1) Receptionist robot At the entrance of a party room, a robot interacts with a participant as a receptionist. It talks to the participant according to whether it knows the participant. If the robot knows a participant, the task is very simple; it will confirm the name of the participant by asking Hello. Are you XXX-san?. If it does not know a participant, it asks the name and then registers the participant s face with his/her name to the face database. During this conversation, the robot should look at the participant and should not turn to any direction during the conversation. 2) Companion robot In the party room, a robot plays a role of a passive companion. It does not speak to a participant, but sees and listens to people. It identifies people s face and the position and turns its body to face the speaker. The issue is to design and develop the tracking system which localizes the speakers and participants in real-time by integrating face identification and localization, sound source localization, and its motor-control. In this situation, we don t make the robot to interact with the participants, because we believe that a silent companion is more suitable for the party and such attitude is more socially acceptable. Please note that the robot knows all the participants, because they registered at the receptionist desk. SIG the humanoid As a test bed of integration of perceptual information to control motor of high degree of freedom (DOF), we designed a humanoid robot (hereafter, referred as SIG) with the following components (Kitano et al. 2000): ffl 4 DOFs of body driven by 4 DC motors Its mechanical structure is shown in Figure 1b. Each DC motor has a potentiometer to measure the direction. ffl A pair of CCD cameras of Sony EVI-G20 for visual stereo input Each camera has 3 DOFs, that is, pan, tilt and zoom. Focus is automatically adjusted. The offset of camera position can be obtained from each camera (Figure 1b). ffl Two pairs of omni-directional microphones (Sony ECM- 77S) (Figure 1c). One pair of microphones are installed at the ear position of the head to collect sounds from the external world. Each microphone is shielded by the cover to prevent from capturing internal noises. The other pair of microphones is to collect sounds within a cover. ffl A cover of the body (Figure 1a) reduces sounds to be emitted to external environments, which is expected to reduce the complexity of sound processing.

3 This cover, made of FRP, is designed by our professional designer for making human robot interaction smoother as well (Nakadai et al. 2000b). System Description Fig. 2 depicts the logical structure of the system based on client/server model. Each server or client executes the following modules: 1. Audition extracts auditory events by pitch extraction, sound source separation and localization, and sends those events to Association, 2. Vision extracts visual events by face extraction, identification and localization, and then sends visual events to Association, 3. Motor Control generates PWM (Pulse Width Modulation) signals to DC motors and sends motor events to Association, 4. Association integrates various events to create streams, 5. Focus-of-Attention makes a plan of motor control, 6. Dialog Control communicates with people by speech synthesis and speech recognition, 7. Face Database maintains the face database, and 8. Viewer instruments various streams and data. Instrumentation is implemented distributed on each node. SIG server displays the radar chart of objects and the stream chart. Motion client displays the radar chart of the body direction. Audition client displays the spectrogram of input sound and pitch (frequency) vs. sound source direction chart. Vision client displays the image of the camera and the status of face identification and tracking. Among these subsystems, Focus-of-Attention, and Dialog Control are newly developed for the system reported in this paper. We use the automatic speech recognition system, Julian developed by Kyoto University (Kawahara et al. 1999). Since the system should run in real-time, the above clients are physically distributed to three Linux nodes connected by TCP/IP over 100Base-TX network and run asynchronously. Vision is placed on a node of Pentium-III 733 MHz, Audition is on a node of Pentium-III 600 MHz, and the rest modules are on a node of Pentium-III 450 MHz. Active Audition Module To understand sound in general, not restricted to a specific sound, a mixture of sound should be analyzed. There are lots of techniques for CASA developed so far, but only the realtime active audition proposed by Nakadai et al runs in realtime (Nakadai et al. 2001). Therefore, we use this system as the base of the receptionist and companion robots. To localize sound sources with two microphones, first a set of peaks are extracted for left and right channels, respectively. Then, the same or similar peaks of left and right channels are identified as a pair and each pair is used to calculate interaural phase difference (IPD) and interaural intensity difference (IID). IPD is calculated from frequencies of less than 1500 Hz, while IID is from frequency of more than 1500 Hz. Since auditory and visual tracking involves motor movements, which cause motor and mechanical noises, audition should suppress or at least reduce such noises. In human robot interaction, when a robot is talking, it should suppress its own speeches. Nakadai et al presented the active audition for humanoids to improve sound source tracking by integrating audition, vision, and motor controls (Nakadai et al. 2000a). We also use their heuristics to reduce internal burst noises caused by motor movements. From IPD and IID, the epipolar geometry is used to obtain the direction of sound source (Nakadai et al. 2000a). The key ideas of their real-time active audition system are twofold; one is to exploit the property of the harmonic structure (fundamental frequency, F 0, and its overtones) to find a more accurate pair of peaks in left and right channels. The other is to search the sound source direction by combining the belief factors of IPD and IID based on Dempster-Shafer theory. Finally, audition module sends an auditory event consisting of pitch (F 0) and a list of 20-best direction ( ) with reliability for each harmonics. Vision: Face identification Module Since the visual processing detects several faces, extracts, identifies and tracks each face simultaneously, the size, direction and brightness of each face changes frequently. The key idea of this task is the combination of skin-color extraction, correlation based matching, and multiple scale images generation (Hidai et al. 2000). The face identification module (see Fig. 2) projects each extracted face into the discrimination space, and calculates its distance d to each registered face. Since this distance depends on the degree (L, the number of registered faces) of discrimination space, it is converted to a parameterindependent probability P v as follows. Z 1 P v = e t t L 2 1 dt (1) d 2 2 The discrimination matrix is created in advance or on demand by using a set of variation of the face with an ID (name). This analysis is done by using Online Linear Discriminant Analysis (Hiraoka et al. 2000). The face localization module converts a face position in 2-D image plane into 3-D world space. Suppose that a face is w w pixels located in (x, y) in the image plane, whose width and height are X and Y, respectively (see screen shots shown in Fig. 4). Then the face position in the world space is obtained as a set of azimuth, elevation ffi, and distance r as follows: r = C 1 w ; =sin 1 ψ x X 2 C 2 r! 1ψ Y 2 ; ffi =sin y C 2 r where C 1 and C 2 are constants defined by the size of the image plane and the image angle of the camera. Finally, vision module sends a visual event consisting of a list of 5-best Face ID (Name) with its reliability and position (distance r, azimuth and elevation ffi) for each face.!

4 Figure 2: Logical organization of the system composed of SIG server, Motion, Audition, Vision, and Reception clients from left to right. Stream Formation and Association Association synchronizes the results (events) given by other modules. It forms an auditory, visual or associated stream by their proximity. Events are stored in the short-term memory only for 2 seconds. Synchronization process runs with the delay of 200 msec, which is the largest delay of the system, that is, vision module. An auditory event is connected to the nearest auditory stream within ±10 ffi and with common or harmonic pitch. A visual event is connected to the nearest visual stream within 40 cm and with common face ID. In either case, if there are plural candidates, the most reliable one is selected. If any appropriate stream is found, such an event becomes a new stream. In case that no event is connected to an existing stream, such a stream remains alive for up to 500 msec. After 500 msec of keep-alive state, the stream terminates. An auditory and a visual streams are associated if their direction difference is within ±10 ffi and this situation continues for more than 50% of the 1 sec period. If either auditory or visual event has not been found for more than 3 sec, such an associated stream is disassociated and only existing auditory or visual stream remains. If the auditory and visual direction difference has been more than 30 ffi for 3 sec, such an associated stream is disassociated to two separate streams. Focus-of-Attention and Dialog Control Focus-of-Attention Control is based on continuity and triggering. By continuity, the system tries to keep the same status, while by triggering, the system tries to track the most interesting object. Attention has two modes, sociallyoriented and task-oriented. In this paper, a receptionist robot adopts task-oriented, while a companion robot adopts socially-oriented focus-of-attention control. Dialog control is a mixed architecture of bottom-up and top-down control. By bottom-up, the most plausible stream means the one that has the highest belief factors. By topdown, the plausibility is defined by the applications. For a receptionist robot, the continuity of the current focus-ofattention has the highest priority. For a companion robot, on the contrary, the stream that is associated the most recently is focused. Experiments and Performance Evaluation The width, length and height of the room of experiment are about 3 m, 3 m, and 2 m, respectively. The room has 6 downlights embedded on the ceiling. For evaluation of the behavior of SIG, one scenario for the receptionist robot and one for the companion robot are executed. The first scenario examines whether an auditory stream triggers Focus-of-Attention to make a plan for SIG to turn to a speaker, and whether SIG can ignore the sound it generates by itself. The second scenario examines how many people SIG can discriminate by integrating auditory and visual streams. SIG as a receptionist robot The precedence of streams selected by focus-of-attention control as a receptionist robot is specified from higher to lower as follows:

5 b) When a participant comes, SIG turns toward him. a) Initial state c) He says Hello and introduces himself to SIG. d) After registration, SIG registers his faces and name. Figure 3: Temporal sequence of snapshots of SIG s interaction as a receptionist robot a) The leftmost man is being tracked. b) He starts talking. c) Associated stream is created. d) The next right man is being tracked. e) Face and voice are associated. f) He stops talking. g) The second right man starts talking. h) The leftmost man starts talking. Figure 4: Temporal sequence of snapshots for a companion robot: scene, radar and sequence chart, spectrogram and pitch-vsdirection chart, and the image of the camera. associated stream auditory stream visual stream One scenario to evaluate the above control is specified as follows: (1) A known participant comes to the receptionist robot. His face has been registered in the face database. (2) He says Hello to SIG. (3) SIG replies Hello. You are XXXsan, aren t you? (4) He says yes. (5) SIG says XXX-san, Welcome to the party. Please enter the room.. Fig. 3 illustrates four snapshots of this scenario. Fig. 3 a) shows the initial state. The speaker on the stand is the mouth of SIG s. Fig. 3 b) shows when a participant comes to the receptionist, but SIG has not noticed him yet, because he is out of SIG s sight. When he speaks to SIG, Audition generates an auditory event with sound source direction, and sends it to Association, which creates an auditory stream. This stream triggers Focus-of-Attention to make a plan that SIG should turn to him. Fig. 3 c) shows the result of the turning. In addition, Audition gives the input to Speech Recognition, which gives the result of speech recognition to Dialog Control. It generates a synthesized speech. Although Audition notices that it hears the sound, SIG will not change the attention, because association of his face and speech keeps SIG s attention on him. Finally, he enters the room while SIG tracks his walking. This scenario shows that SIG takes two interesting behaviors. One is voice-triggered tracking shown in Fig. 3 c). The other is that SIG does not pay attention to its own speech. This is attained naturally by the current association algorithm, because this algorithm is designed by taking into account the fact that conversation is conducted by alternate initiatives. The variant of this scenario is also used to check whether the system works well. (1 ) A participant comes to the receptionist robot, whose face has not been registered in the face database. (2) He says Hello to SIG. (3) SIG replies Hello. Could you give me your name? (4) He says his name. (5) SIG says XXX-san, Welcome to the party. Please enter the room. After giving his name to the system, Face Database module is invoked. SIG as a companion robot The precedence of streams selected by focus-of-attention control as a companion robot is as follows: auditory stream associated stream visual stream

6 There is no explicit scenario. Four speakers actually talks spontaneously in attendance of SIG. Then SIG tracks some speaker and then changes focus-of-attention to others. The observed behavior is evaluating by consulting the internal states of SIG, that is, auditory and visual localization shown in the radar chart, auditory, visual, and associated streams shown in the stream chart, and peak extraction. The top-left image in each snapshot shows the scene of this experiment recorded by a video camera. The top-right image consists of the radar chart (left) and the stream chart (right) updated in real-time. The former shows the environment recognized by SIG at the moment of the snapshot. A pink sector indicates a visual field of SIG. Because of using the absolute coordinate, the pink sector rotates as SIG turns. A green point with a label is the direction and the face ID of a visual stream. A blue sector is the direction of an auditory stream. Green, blue and red lines indicate the direction of visual, auditory and associated stream, respectively. Blue and green thin lines indicate auditory and visual streams, respectively. Blue, green and red thick lines indicate associated streams with only auditory, only visual, and both information, respectively. The bottom-left image shows the auditory viewer consisting of the power spectrum and auditory event viewer. The latter shows an auditory event as a filled circle with its pitch in X axis and its direction in Y axis. The bottom-right image shows the visual viewer captured by the SIG s left eye. A detected face is displayed with a red rectangle. The current system does not use a stereo vision. The temporal sequence of SIG s recognition and actions shows that the design of companion robot works well and pays its attention to a new talker. The current system has attained a passive companion. To design and develop an active companion may be important future work. Observations As a receptionist robot, once an association is established, SIG keeps its face fixed to the direction of the speaker of the associated stream. Therefore, even when SIG utters via a loud speaker on the left, SIG does not pay an attention to the sound source, that is, its own speech. This phenomenon of focus-of-attention results in an automatic suppression of self-generated sounds. Of course, this kind of suppression is observed by another benchmark which contains the situation that SIG and the human speaker utter at the same time. As a companion robot, SIG pays attention to a speaker appropriately. SIG also tracks the same person well when two moving talkers cross and their faces are out of sight of SIG. These results prove that the proposed system succeeds in real-time sensorimotor tasks of tracking with face identification. The current system has attained a passive companion. To design and develop an active companion may be important future work. SIG as a non-verbal Eliza As socially-oriented attention control, interesting human behaviors are observed. The mechanism of associating auditory and visual streams and that of socially-oriented attention control are explained in advance to the user. 1. Some people walk around talking with their hand covering SIG s eyes in order to confirm the performance of auditory tracking. 2. Some people creep on the floor with talking in order to confirm the performance of auditory tracking. 3. Some people play hide-and-seek games with SIG. 4. Some people play sounds from a pair of loud speakers with changing the balance control of pre-amplifier in order to confirm the performance of auditory tracking. 5. When one person reads loud a book and then another person starts to read loud a book, SIG turns its head to the second talker for a short time and then is back to the first talker and keeps its attention on him/her. Above observations remind us of Eliza (Weizenbaum 1966), although SIG does not say anything except a receptionist robot. When the user says something to SIG, it turns to him/her, which invites the participation of the user into interaction. SIG also invites exploration of the principles of its functioning, that is, the user is drawn in to see how SIG will respond to variations in behavior. Since SIG takes only passive behaviors, it does not arouse higher expectations of verisimilitude that it can deliver on. Needless to say, there are lots of works remaining to validate the proposed approach for personality of artifacts. We are currently working to incorporate active social interaction by developing the capability of listening to simultaneous speeches. Conclusions and Future Works In this paper, we demonstrate that auditory and visual multiple-object tracking subsystem can augment the functionality of human robot interaction. Although a simple scheme of behavior is implemented, human robot interaction is drastically improved by real-time multiple-person tracking. We can pleasantly spend an hour with SIG as a companion robot even if its attitude is quite passive. Since the application of auditory and visual multipleobject tracking is not restricted to robots or humanoids, auditory capability can be transferred to software agents or systems. As discussed in the introduction section, auditory information should not be ignored in computer graphics or human computer interaction. By integrating audition and vision, more cross-modal perception can be attained. Future work includes applications such as listening to several things simultaneously (Okuno, Nakatani, & Kawabata 1999), cocktail party computer, integration of auditory and visual tracking and pose and gesture recognition, and other novel areas. Since the event-level communication is less expensive than the low-level data representation, say signals itself, auditory and visual multiple-object tracking can be applied to tele-existence or virtual reality. Acknowledgments We thank our colleagues of Symbiotic Intelligence Group, Kitano Symbiotic Systems Project, Tatsuya Matsui, and former colloeague, Dr. Tino Lourens, for their discussions. We

7 also thank Prof. Tatsuya Kawahara of Kyoto University for allowing us to use Julian automatic speech recognition system. References Breazeal, C., and Scassellati, B A context-dependent attention system for a social robot. In Proceedints of the Sixteenth International Joint Conference on Atificial Intelligence (IJCAI- 99), Breazeal, C Emotive qualities in robot speech. In Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS-2001), IEEE. Brooks, R. A.; Breazeal, C.; Irie, R.; Kemp, C. C.; Marjanovic, M.; Scassellati, B.; and Williamson, M. M Alternative essences of intelligence. In Proceedings of 15th National Conference on Artificial Intelligence (AAAI-98), AAAI. Brooks, R.; Breazeal, C.; Marjanovie, M.; Scassellati, B.; and Williamson, M The cog project: Building a humanoid robot. Technical report, MIT. Cherry, E. C Some experiments on the recognition of speech, with one and with two ears. Journal of Acoustic Society of America 25: Handel, S Listening. MA.: The MIT Press. Hansen, J.; Mammone, R.; and Young, S Editorial for the special issue on robust speech processing. IEEE Transactions on Speech and Audio Processing 2(4): Hidai, K.; Mizoguchi, H.; Hiraoka, K.; Tanaka, M.; Shigehara, T.; and Mishima, T Robust face detection against brightness fluctuation and size variation. In Proceedings of IEEE/RAS International Conference on Intelligent Robots and Systems (IROS 2000), IEEE. Hiraoka, K.; Hamahira, M.; Hidai, K.; Mizoguchi, H.; Mishima, T.; and Yoshizawa, S Fast algorithm for online linear discriminant analysis. In Proceedings of ITC-2000, IEEK/IEICE. Kawahara, T.; Lee, A.; Kobayashi, T.; Takeda, K.; Minematsu, N.; Itou, K.; Ito, A.; Yamamoto, M.; Yamada, A.; Utsuro, T.; and Shikano, K Japanese dictation toolkit 1997 version. Journal of Acoustic Society Japan (E) 20(3): Kitano, H.; Okuno, H. G.; Nakadai, K.; Fermin, I.; Sabish, T.; Nakagawa, Y.; and Matsui, T Designing a humanoid head for robocup challenge. In Proceedings of the Fourth International Conference on Autonomous Agents (Agents 2000), ACM. Matsusaka, Y.; Tojo, T.; Kuota, S.; Furukawa, K.; Tamiya, D.; Hayata, K.; Nakano, Y.; and Kobayashi, T Multi-person conversation via multi-modal interface a robot who communicates with multi-user. In Proceedings of 6th European Conference on Speech Communication Technology (EUROSPEECH- 99), ESCA. Nakadai, K.; Lourens, T.; Okuno, H. G.; and Kitano, H. 2000a. Active audition for humanoid. In Proceedings of 17th National Conference on Artificial Intelligence (AAAI-2000), AAAI. Nakadai, K.; Matsui, T.; Okuno, H. G.; and Kitano, H. 2000b. Active audition system and humanoid exterior design. In Proceedings of IEEE/RAS International Conference on Intelligent Robots and Systems (IROS 2000), IEEE. Nakadai, K. Hidai, K.; Mizoguchi, H.; Okuno, H. G.; and Kitano, H Real-time auditory and visual multiple-object tracking for robots. In Proceedings of the Seventeenth International Joint Conference on Artificial Intelligence (IJCAI-01), volume II, IJCAI. Okuno, H. G.; Nakatani, T.; and Kawabata, T Listening to two simultaneous speeches. Speech Communication 27(3-4): Ono, T.; Imai, M.; and Ishiguro, H A model of embodied communications with gestures between humans and robots. In Proceedings of Twenty-third Annual Meeting of the Cognitive Science Society (CogSci2001), AAAI. Rosenthal, D., and Okuno, H. G., eds Computational Auditory Scene Analysis. Mahwah, New Jersey: Lawrence Erlbaum Associates. Waldherr, S.; Thrun, S.; Romero, R.; and Margaritis, D Template-based recoginition of pose and motion gestures on a mobile robot. In Proceedings of 15th National Conference on Artificial Intelligence (AAAI-98), AAAI. Weizenbaum, J Eliza a computer program for the study of natural language communication between man and machine. Communications of the ACM 9(1):36 45.

Human-Robot Interaction through Real-Time Auditory and Visual Multiple-Talker Tracking

Human-Robot Interaction through Real-Time Auditory and Visual Multiple-Talker Tracking Proceedings of the 2001 IEEE/RSJ International Conference on Intelligent Robots and Systems Maui, Hawaii, USA, Oct. 29 - Nov. 03, 2001 Human-Robot Interaction through Real-Time Auditory and Visual Multiple-Talker

More information

Realizing Personality in Audio-Visually Triggered Non-verbal Behaviors

Realizing Personality in Audio-Visually Triggered Non-verbal Behaviors Proceedings orthe 2003 IEEE loternatiood Conference on Robotics &Automation Taipei, Taiwan, September 14-19, 2003 Realizing Personality in Audio-Visually Triggered Non-verbal Behaviors I Hiroshi G. Okunot,*,

More information

A tension-moderating mechanism for promoting speech-based human-robot interaction

A tension-moderating mechanism for promoting speech-based human-robot interaction A tension-moderating mechanism for promoting speech-based human-robot interaction Takayuki Kanda *, Kayoko Iwase *, Masahiro Shiomi *, Hiroshi Ishiguro * * Department of Communication Robots Faculty of

More information

A context-dependent attention system for a social robot

A context-dependent attention system for a social robot A context-dependent attention system for a social robot Cynthia Breazeal and Brian Scassellati MIT Artificial Intelligence Lab 545 Technology Square Cambridge, MA 02139 U. S. A. Abstract This paper presents

More information

Auditory Scene Analysis

Auditory Scene Analysis 1 Auditory Scene Analysis Albert S. Bregman Department of Psychology McGill University 1205 Docteur Penfield Avenue Montreal, QC Canada H3A 1B1 E-mail: bregman@hebb.psych.mcgill.ca To appear in N.J. Smelzer

More information

Noise-Robust Speech Recognition Technologies in Mobile Environments

Noise-Robust Speech Recognition Technologies in Mobile Environments Noise-Robust Speech Recognition echnologies in Mobile Environments Mobile environments are highly influenced by ambient noise, which may cause a significant deterioration of speech recognition performance.

More information

Hearing in the Environment

Hearing in the Environment 10 Hearing in the Environment Click Chapter to edit 10 Master Hearing title in the style Environment Sound Localization Complex Sounds Auditory Scene Analysis Continuity and Restoration Effects Auditory

More information

Computational Perception /785. Auditory Scene Analysis

Computational Perception /785. Auditory Scene Analysis Computational Perception 15-485/785 Auditory Scene Analysis A framework for auditory scene analysis Auditory scene analysis involves low and high level cues Low level acoustic cues are often result in

More information

Learning Utility for Behavior Acquisition and Intention Inference of Other Agent

Learning Utility for Behavior Acquisition and Intention Inference of Other Agent Learning Utility for Behavior Acquisition and Intention Inference of Other Agent Yasutake Takahashi, Teruyasu Kawamata, and Minoru Asada* Dept. of Adaptive Machine Systems, Graduate School of Engineering,

More information

Auditory scene analysis in humans: Implications for computational implementations.

Auditory scene analysis in humans: Implications for computational implementations. Auditory scene analysis in humans: Implications for computational implementations. Albert S. Bregman McGill University Introduction. The scene analysis problem. Two dimensions of grouping. Recognition

More information

AVR Based Gesture Vocalizer Using Speech Synthesizer IC

AVR Based Gesture Vocalizer Using Speech Synthesizer IC AVR Based Gesture Vocalizer Using Speech Synthesizer IC Mr.M.V.N.R.P.kumar 1, Mr.Ashutosh Kumar 2, Ms. S.B.Arawandekar 3, Mr.A. A. Bhosale 4, Mr. R. L. Bhosale 5 Dept. Of E&TC, L.N.B.C.I.E.T. Raigaon,

More information

Analysis of Emotion Recognition using Facial Expressions, Speech and Multimodal Information

Analysis of Emotion Recognition using Facial Expressions, Speech and Multimodal Information Analysis of Emotion Recognition using Facial Expressions, Speech and Multimodal Information C. Busso, Z. Deng, S. Yildirim, M. Bulut, C. M. Lee, A. Kazemzadeh, S. Lee, U. Neumann, S. Narayanan Emotion

More information

Computational Auditory Scene Analysis: An overview and some observations. CASA survey. Other approaches

Computational Auditory Scene Analysis: An overview and some observations. CASA survey. Other approaches CASA talk - Haskins/NUWC - Dan Ellis 1997oct24/5-1 The importance of auditory illusions for artificial listeners 1 Dan Ellis International Computer Science Institute, Berkeley CA

More information

SENSATION AND PERCEPTION KEY TERMS

SENSATION AND PERCEPTION KEY TERMS SENSATION AND PERCEPTION KEY TERMS BOTTOM-UP PROCESSING BOTTOM-UP PROCESSING refers to processing sensory information as it is coming in. In other words, if I flash a random picture on the screen, your

More information

Speech Group, Media Laboratory

Speech Group, Media Laboratory Speech Group, Media Laboratory The Speech Research Group of the M.I.T. Media Laboratory is concerned with understanding human speech communication and building systems capable of emulating conversational

More information

Analysis of Speech Recognition Techniques for use in a Non-Speech Sound Recognition System

Analysis of Speech Recognition Techniques for use in a Non-Speech Sound Recognition System Analysis of Recognition Techniques for use in a Sound Recognition System Michael Cowling, Member, IEEE and Renate Sitte, Member, IEEE Griffith University Faculty of Engineering & Information Technology

More information

Spectrograms (revisited)

Spectrograms (revisited) Spectrograms (revisited) We begin the lecture by reviewing the units of spectrograms, which I had only glossed over when I covered spectrograms at the end of lecture 19. We then relate the blocks of a

More information

USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES

USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES Varinthira Duangudom and David V Anderson School of Electrical and Computer Engineering, Georgia Institute of Technology Atlanta, GA 30332

More information

Speaker 2. Speaker 3. Speaker m~2m. Machine Audition System. Speaker 1 says... Speaker 2 says... Speaker 3 says...

Speaker 2. Speaker 3. Speaker m~2m. Machine Audition System. Speaker 1 says... Speaker 2 says... Speaker 3 says... Challenge Problem for Computational Auditory Scene Analysis: Understanding Three Simultaneous Speeches Hiroshi G. Okuno, Tomohiro Nakatani, and Takeshi Kawabata NTT Basic Research Laboratories 3-1 Morinosato-Wakamiya,

More information

Effects of speaker's and listener's environments on speech intelligibili annoyance. Author(s)Kubo, Rieko; Morikawa, Daisuke; Akag

Effects of speaker's and listener's environments on speech intelligibili annoyance. Author(s)Kubo, Rieko; Morikawa, Daisuke; Akag JAIST Reposi https://dspace.j Title Effects of speaker's and listener's environments on speech intelligibili annoyance Author(s)Kubo, Rieko; Morikawa, Daisuke; Akag Citation Inter-noise 2016: 171-176 Issue

More information

Motion Control for Social Behaviours

Motion Control for Social Behaviours Motion Control for Social Behaviours Aryel Beck a.beck@ntu.edu.sg Supervisor: Nadia Magnenat-Thalmann Collaborators: Zhang Zhijun, Rubha Shri Narayanan, Neetha Das 10-03-2015 INTRODUCTION In order for

More information

CS343: Artificial Intelligence

CS343: Artificial Intelligence CS343: Artificial Intelligence Introduction: Part 2 Prof. Scott Niekum University of Texas at Austin [Based on slides created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All materials

More information

ITU-T. FG AVA TR Version 1.0 (10/2013) Part 3: Using audiovisual media A taxonomy of participation

ITU-T. FG AVA TR Version 1.0 (10/2013) Part 3: Using audiovisual media A taxonomy of participation International Telecommunication Union ITU-T TELECOMMUNICATION STANDARDIZATION SECTOR OF ITU FG AVA TR Version 1.0 (10/2013) Focus Group on Audiovisual Media Accessibility Technical Report Part 3: Using

More information

Binaural Hearing for Robots Introduction to Robot Hearing

Binaural Hearing for Robots Introduction to Robot Hearing Binaural Hearing for Robots Introduction to Robot Hearing 1Radu Horaud Binaural Hearing for Robots 1. Introduction to Robot Hearing 2. Methodological Foundations 3. Sound-Source Localization 4. Machine

More information

Speech as HCI. HCI Lecture 11. Human Communication by Speech. Speech as HCI(cont. 2) Guest lecture: Speech Interfaces

Speech as HCI. HCI Lecture 11. Human Communication by Speech. Speech as HCI(cont. 2) Guest lecture: Speech Interfaces HCI Lecture 11 Guest lecture: Speech Interfaces Hiroshi Shimodaira Institute for Communicating and Collaborative Systems (ICCS) Centre for Speech Technology Research (CSTR) http://www.cstr.ed.ac.uk Thanks

More information

Chapter 6. Attention. Attention

Chapter 6. Attention. Attention Chapter 6 Attention Attention William James, in 1890, wrote Everyone knows what attention is. Attention is the taking possession of the mind, in clear and vivid form, of one out of what seem several simultaneously

More information

Visi-Pitch IV is the latest version of the most widely

Visi-Pitch IV is the latest version of the most widely APPLICATIONS Voice Disorders Motor Speech Disorders Voice Typing Fluency Selected Articulation Training Hearing-Impaired Speech Professional Voice Accent Reduction and Second Language Learning Importance

More information

HOW TO USE THE SHURE MXA910 CEILING ARRAY MICROPHONE FOR VOICE LIFT

HOW TO USE THE SHURE MXA910 CEILING ARRAY MICROPHONE FOR VOICE LIFT HOW TO USE THE SHURE MXA910 CEILING ARRAY MICROPHONE FOR VOICE LIFT Created: Sept 2016 Updated: June 2017 By: Luis Guerra Troy Jensen The Shure MXA910 Ceiling Array Microphone offers the unique advantage

More information

Sensation and Perception

Sensation and Perception 1 Sensation and Perception DR. ARNEL BANAGA SALGADO, Doctor of Psychology (USA) FPM (Ph.D.) Psychology (India) Doctor of Education (Phl) Master of Arts in Nursing (Phl) Master of Arts in Teaching Psychology

More information

Sound Interfaces Engineering Interaction Technologies. Prof. Stefanie Mueller HCI Engineering Group

Sound Interfaces Engineering Interaction Technologies. Prof. Stefanie Mueller HCI Engineering Group Sound Interfaces 6.810 Engineering Interaction Technologies Prof. Stefanie Mueller HCI Engineering Group what is sound? if a tree falls in the forest and nobody is there does it make sound?

More information

Sign Language Interpretation Using Pseudo Glove

Sign Language Interpretation Using Pseudo Glove Sign Language Interpretation Using Pseudo Glove Mukul Singh Kushwah, Manish Sharma, Kunal Jain and Anish Chopra Abstract The research work presented in this paper explores the ways in which, people who

More information

Hearing II Perceptual Aspects

Hearing II Perceptual Aspects Hearing II Perceptual Aspects Overview of Topics Chapter 6 in Chaudhuri Intensity & Loudness Frequency & Pitch Auditory Space Perception 1 2 Intensity & Loudness Loudness is the subjective perceptual quality

More information

TWO HANDED SIGN LANGUAGE RECOGNITION SYSTEM USING IMAGE PROCESSING

TWO HANDED SIGN LANGUAGE RECOGNITION SYSTEM USING IMAGE PROCESSING 134 TWO HANDED SIGN LANGUAGE RECOGNITION SYSTEM USING IMAGE PROCESSING H.F.S.M.Fonseka 1, J.T.Jonathan 2, P.Sabeshan 3 and M.B.Dissanayaka 4 1 Department of Electrical And Electronic Engineering, Faculty

More information

HCS 7367 Speech Perception

HCS 7367 Speech Perception Long-term spectrum of speech HCS 7367 Speech Perception Connected speech Absolute threshold Males Dr. Peter Assmann Fall 212 Females Long-term spectrum of speech Vowels Males Females 2) Absolute threshold

More information

Accessible Computing Research for Users who are Deaf and Hard of Hearing (DHH)

Accessible Computing Research for Users who are Deaf and Hard of Hearing (DHH) Accessible Computing Research for Users who are Deaf and Hard of Hearing (DHH) Matt Huenerfauth Raja Kushalnagar Rochester Institute of Technology DHH Auditory Issues Links Accents/Intonation Listening

More information

Member 1 Member 2 Member 3 Member 4 Full Name Krithee Sirisith Pichai Sodsai Thanasunn

Member 1 Member 2 Member 3 Member 4 Full Name Krithee Sirisith Pichai Sodsai Thanasunn Microsoft Imagine Cup 2010 Thailand Software Design Round 1 Project Proposal Template PROJECT PROPOSAL DUE: 31 Jan 2010 To Submit to proposal: Register at www.imaginecup.com; select to compete in Software

More information

HearIntelligence by HANSATON. Intelligent hearing means natural hearing.

HearIntelligence by HANSATON. Intelligent hearing means natural hearing. HearIntelligence by HANSATON. HearIntelligence by HANSATON. Intelligent hearing means natural hearing. Acoustic environments are complex. We are surrounded by a variety of different acoustic signals, speech

More information

General Soundtrack Analysis

General Soundtrack Analysis General Soundtrack Analysis Dan Ellis oratory for Recognition and Organization of Speech and Audio () Electrical Engineering, Columbia University http://labrosa.ee.columbia.edu/

More information

Echo Canceller with Noise Reduction Provides Comfortable Hands-free Telecommunication in Noisy Environments

Echo Canceller with Noise Reduction Provides Comfortable Hands-free Telecommunication in Noisy Environments Canceller with Reduction Provides Comfortable Hands-free Telecommunication in Noisy Environments Sumitaka Sakauchi, Yoichi Haneda, Manabu Okamoto, Junko Sasaki, and Akitoshi Kataoka Abstract Audio-teleconferencing,

More information

Digital. hearing instruments have burst on the

Digital. hearing instruments have burst on the Testing Digital and Analog Hearing Instruments: Processing Time Delays and Phase Measurements A look at potential side effects and ways of measuring them by George J. Frye Digital. hearing instruments

More information

Auditory Scene Analysis. Dr. Maria Chait, UCL Ear Institute

Auditory Scene Analysis. Dr. Maria Chait, UCL Ear Institute Auditory Scene Analysis Dr. Maria Chait, UCL Ear Institute Expected learning outcomes: Understand the tasks faced by the auditory system during everyday listening. Know the major Gestalt principles. Understand

More information

PERCEPTION OF UNATTENDED SPEECH. University of Sussex Falmer, Brighton, BN1 9QG, UK

PERCEPTION OF UNATTENDED SPEECH. University of Sussex Falmer, Brighton, BN1 9QG, UK PERCEPTION OF UNATTENDED SPEECH Marie Rivenez 1,2, Chris Darwin 1, Anne Guillaume 2 1 Department of Psychology University of Sussex Falmer, Brighton, BN1 9QG, UK 2 Département Sciences Cognitives Institut

More information

Binaural Hearing. Why two ears? Definitions

Binaural Hearing. Why two ears? Definitions Binaural Hearing Why two ears? Locating sounds in space: acuity is poorer than in vision by up to two orders of magnitude, but extends in all directions. Role in alerting and orienting? Separating sound

More information

Artificial Emotions to Assist Social Coordination in HRI

Artificial Emotions to Assist Social Coordination in HRI Artificial Emotions to Assist Social Coordination in HRI Jekaterina Novikova, Leon Watts Department of Computer Science University of Bath Bath, BA2 7AY United Kingdom j.novikova@bath.ac.uk Abstract. Human-Robot

More information

Research Proposal on Emotion Recognition

Research Proposal on Emotion Recognition Research Proposal on Emotion Recognition Colin Grubb June 3, 2012 Abstract In this paper I will introduce my thesis question: To what extent can emotion recognition be improved by combining audio and visual

More information

C H A N N E L S A N D B A N D S A C T I V E N O I S E C O N T R O L 2

C H A N N E L S A N D B A N D S A C T I V E N O I S E C O N T R O L 2 C H A N N E L S A N D B A N D S Audibel hearing aids offer between 4 and 16 truly independent channels and bands. Channels are sections of the frequency spectrum that are processed independently by the

More information

PSYC 441 Cognitive Psychology II

PSYC 441 Cognitive Psychology II PSYC 441 Cognitive Psychology II Session 4 Background of Object Recognition Lecturer: Dr. Benjamin Amponsah, Dept., of Psychology, UG, Legon Contact Information: bamponsah@ug.edu.gh College of Education

More information

An active unpleasantness control system for indoor noise based on auditory masking

An active unpleasantness control system for indoor noise based on auditory masking An active unpleasantness control system for indoor noise based on auditory masking Daisuke Ikefuji, Masato Nakayama, Takanabu Nishiura and Yoich Yamashita Graduate School of Information Science and Engineering,

More information

VPAT Summary. VPAT Details. Section Telecommunications Products - Detail. Date: October 8, 2014 Name of Product: BladeCenter HS23

VPAT Summary. VPAT Details. Section Telecommunications Products - Detail. Date: October 8, 2014 Name of Product: BladeCenter HS23 Date: October 8, 2014 Name of Product: BladeCenter HS23 VPAT Summary Criteria Status Remarks and Explanations Section 1194.21 Software Applications and Operating Systems Section 1194.22 Web-based Internet

More information

Product Model #: Digital Portable Radio XTS 5000 (Std / Rugged / Secure / Type )

Product Model #: Digital Portable Radio XTS 5000 (Std / Rugged / Secure / Type ) Rehabilitation Act Amendments of 1998, Section 508 Subpart 1194.25 Self-Contained, Closed Products The following features are derived from Section 508 When a timed response is required alert user, allow

More information

The Use of Voice Recognition and Speech Command Technology as an Assistive Interface for ICT in Public Spaces.

The Use of Voice Recognition and Speech Command Technology as an Assistive Interface for ICT in Public Spaces. The Use of Voice Recognition and Speech Command Technology as an Assistive Interface for ICT in Public Spaces. A whitepaper published by Peter W Jarvis (Senior Executive VP, Storm Interface) and Nicky

More information

Development of an interactive digital signage based on F-formation system

Development of an interactive digital signage based on F-formation system Development of an interactive digital signage based on F-formation system Yu Kobayashi 1, Masahide Yuasa 2, and Daisuke Katagami 1 1 Tokyo Polytechnic University, Japan 2 Shonan Institute of Technology,

More information

AUDL GS08/GAV1 Signals, systems, acoustics and the ear. Pitch & Binaural listening

AUDL GS08/GAV1 Signals, systems, acoustics and the ear. Pitch & Binaural listening AUDL GS08/GAV1 Signals, systems, acoustics and the ear Pitch & Binaural listening Review 25 20 15 10 5 0-5 100 1000 10000 25 20 15 10 5 0-5 100 1000 10000 Part I: Auditory frequency selectivity Tuning

More information

whether or not the fundamental is actually present.

whether or not the fundamental is actually present. 1) Which of the following uses a computer CPU to combine various pure tones to generate interesting sounds or music? 1) _ A) MIDI standard. B) colored-noise generator, C) white-noise generator, D) digital

More information

Lecture 3: Perception

Lecture 3: Perception ELEN E4896 MUSIC SIGNAL PROCESSING Lecture 3: Perception 1. Ear Physiology 2. Auditory Psychophysics 3. Pitch Perception 4. Music Perception Dan Ellis Dept. Electrical Engineering, Columbia University

More information

Sign Language in the Intelligent Sensory Environment

Sign Language in the Intelligent Sensory Environment Sign Language in the Intelligent Sensory Environment Ákos Lisztes, László Kővári, Andor Gaudia, Péter Korondi Budapest University of Science and Technology, Department of Automation and Applied Informatics,

More information

Ear Beamer. Aaron Lucia, Niket Gupta, Matteo Puzella, Nathan Dunn. Department of Electrical and Computer Engineering

Ear Beamer. Aaron Lucia, Niket Gupta, Matteo Puzella, Nathan Dunn. Department of Electrical and Computer Engineering Ear Beamer Aaron Lucia, Niket Gupta, Matteo Puzella, Nathan Dunn The Team Aaron Lucia CSE Niket Gupta CSE Matteo Puzella EE Nathan Dunn CSE Advisor: Prof. Mario Parente 2 A Familiar Scenario 3 Outlining

More information

Acoustic Sensing With Artificial Intelligence

Acoustic Sensing With Artificial Intelligence Acoustic Sensing With Artificial Intelligence Bowon Lee Department of Electronic Engineering Inha University Incheon, South Korea bowon.lee@inha.ac.kr bowon.lee@ieee.org NVIDIA Deep Learning Day Seoul,

More information

Single-Channel Sound Source Localization Based on Discrimination of Acoustic Transfer Functions

Single-Channel Sound Source Localization Based on Discrimination of Acoustic Transfer Functions 3 Single-Channel Sound Source Localization Based on Discrimination of Acoustic Transfer Functions Ryoichi Takashima, Tetsuya Takiguchi and Yasuo Ariki Graduate School of System Informatics, Kobe University,

More information

Some methodological aspects for measuring asynchrony detection in audio-visual stimuli

Some methodological aspects for measuring asynchrony detection in audio-visual stimuli Some methodological aspects for measuring asynchrony detection in audio-visual stimuli Pacs Reference: 43.66.Mk, 43.66.Lj Van de Par, Steven ; Kohlrausch, Armin,2 ; and Juola, James F. 3 ) Philips Research

More information

Sound localization psychophysics

Sound localization psychophysics Sound localization psychophysics Eric Young A good reference: B.C.J. Moore An Introduction to the Psychology of Hearing Chapter 7, Space Perception. Elsevier, Amsterdam, pp. 233-267 (2004). Sound localization:

More information

Frequency Tracking: LMS and RLS Applied to Speech Formant Estimation

Frequency Tracking: LMS and RLS Applied to Speech Formant Estimation Aldebaro Klautau - http://speech.ucsd.edu/aldebaro - 2/3/. Page. Frequency Tracking: LMS and RLS Applied to Speech Formant Estimation ) Introduction Several speech processing algorithms assume the signal

More information

VISUAL PERCEPTION & COGNITIVE PROCESSES

VISUAL PERCEPTION & COGNITIVE PROCESSES VISUAL PERCEPTION & COGNITIVE PROCESSES Prof. Rahul C. Basole CS4460 > March 31, 2016 How Are Graphics Used? Larkin & Simon (1987) investigated usefulness of graphical displays Graphical visualization

More information

What you re in for. Who are cochlear implants for? The bottom line. Speech processing schemes for

What you re in for. Who are cochlear implants for? The bottom line. Speech processing schemes for What you re in for Speech processing schemes for cochlear implants Stuart Rosen Professor of Speech and Hearing Science Speech, Hearing and Phonetic Sciences Division of Psychology & Language Sciences

More information

Binaural processing of complex stimuli

Binaural processing of complex stimuli Binaural processing of complex stimuli Outline for today Binaural detection experiments and models Speech as an important waveform Experiments on understanding speech in complex environments (Cocktail

More information

SPHSC 462 HEARING DEVELOPMENT. Overview Review of Hearing Science Introduction

SPHSC 462 HEARING DEVELOPMENT. Overview Review of Hearing Science Introduction SPHSC 462 HEARING DEVELOPMENT Overview Review of Hearing Science Introduction 1 Overview of course and requirements Lecture/discussion; lecture notes on website http://faculty.washington.edu/lawerner/sphsc462/

More information

Perception Lie Paradox: Mathematically Proved Uncertainty about Humans Perception Similarity

Perception Lie Paradox: Mathematically Proved Uncertainty about Humans Perception Similarity Perception Lie Paradox: Mathematically Proved Uncertainty about Humans Perception Similarity Ahmed M. Mahran Computer and Systems Engineering Department, Faculty of Engineering, Alexandria University,

More information

Learning Objectives. AT Goals. Assistive Technology for Sensory Impairments. Review Course for Assistive Technology Practitioners & Suppliers

Learning Objectives. AT Goals. Assistive Technology for Sensory Impairments. Review Course for Assistive Technology Practitioners & Suppliers Assistive Technology for Sensory Impairments Review Course for Assistive Technology Practitioners & Suppliers Learning Objectives Define the purpose of AT for persons who have sensory impairment Identify

More information

Definition Slides. Sensation. Perception. Bottom-up processing. Selective attention. Top-down processing 11/3/2013

Definition Slides. Sensation. Perception. Bottom-up processing. Selective attention. Top-down processing 11/3/2013 Definition Slides Sensation = the process by which our sensory receptors and nervous system receive and represent stimulus energies from our environment. Perception = the process of organizing and interpreting

More information

= add definition here. Definition Slide

= add definition here. Definition Slide = add definition here Definition Slide Definition Slides Sensation = the process by which our sensory receptors and nervous system receive and represent stimulus energies from our environment. Perception

More information

Product Model #:ASTRO Digital Spectra Consolette W7 Models (Local Control)

Product Model #:ASTRO Digital Spectra Consolette W7 Models (Local Control) Subpart 1194.25 Self-Contained, Closed Products When a timed response is required alert user, allow sufficient time for him to indicate that he needs additional time to respond [ N/A ] For touch screen

More information

Prof. Greg Francis 7/7/08

Prof. Greg Francis 7/7/08 Perceptual development IIE 366: Developmental Psychology Chapter 5: Perceptual and Motor Development Module 5.1 Basic Sensory and Perceptual Processes Greg Francis Lecture 11 Children and Their Development,

More information

Skin color detection for face localization in humanmachine

Skin color detection for face localization in humanmachine Research Online ECU Publications Pre. 2011 2001 Skin color detection for face localization in humanmachine communications Douglas Chai Son Lam Phung Abdesselam Bouzerdoum 10.1109/ISSPA.2001.949848 This

More information

Facial expression recognition with spatiotemporal local descriptors

Facial expression recognition with spatiotemporal local descriptors Facial expression recognition with spatiotemporal local descriptors Guoying Zhao, Matti Pietikäinen Machine Vision Group, Infotech Oulu and Department of Electrical and Information Engineering, P. O. Box

More information

Summary Table: Voluntary Product Accessibility Template

Summary Table: Voluntary Product Accessibility Template Date: August 16 th, 2011 Name of Product: Cisco Unified Wireless IP Phone 7921G, 7925G, 7925G-EX and 7926G Contact for more information: Conrad Price, cprice@cisco.com Summary Table: Voluntary Product

More information

A visual concomitant of the Lombard reflex

A visual concomitant of the Lombard reflex ISCA Archive http://www.isca-speech.org/archive Auditory-Visual Speech Processing 2005 (AVSP 05) British Columbia, Canada July 24-27, 2005 A visual concomitant of the Lombard reflex Jeesun Kim 1, Chris

More information

Modulation and Top-Down Processing in Audition

Modulation and Top-Down Processing in Audition Modulation and Top-Down Processing in Audition Malcolm Slaney 1,2 and Greg Sell 2 1 Yahoo! Research 2 Stanford CCRMA Outline The Non-Linear Cochlea Correlogram Pitch Modulation and Demodulation Information

More information

Roger TM. for Education Bridging the understanding gap

Roger TM. for Education Bridging the understanding gap Roger TM for Education Bridging the understanding gap Active participation in the modern classroom For all children, especially those with hearing difficulties, it s important to hear well in school for

More information

Networx Enterprise Proposal for Internet Protocol (IP)-Based Services. Supporting Features. Remarks and explanations. Criteria

Networx Enterprise Proposal for Internet Protocol (IP)-Based Services. Supporting Features. Remarks and explanations. Criteria Section 1194.21 Software Applications and Operating Systems Internet Protocol Telephony Service (IPTelS) Detail Voluntary Product Accessibility Template (a) When software is designed to run on a system

More information

1. INTRODUCTION. Vision based Multi-feature HGR Algorithms for HCI using ISL Page 1

1. INTRODUCTION. Vision based Multi-feature HGR Algorithms for HCI using ISL Page 1 1. INTRODUCTION Sign language interpretation is one of the HCI applications where hand gesture plays important role for communication. This chapter discusses sign language interpretation system with present

More information

Linguistic Phonetics Fall 2005

Linguistic Phonetics Fall 2005 MIT OpenCourseWare http://ocw.mit.edu 24.963 Linguistic Phonetics Fall 2005 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. 24.963 Linguistic Phonetics

More information

Section Web-based Internet information and applications VoIP Transport Service (VoIPTS) Detail Voluntary Product Accessibility Template

Section Web-based Internet information and applications VoIP Transport Service (VoIPTS) Detail Voluntary Product Accessibility Template Section 1194.22 Web-based Internet information and applications VoIP Transport Service (VoIPTS) Detail Voluntary Product Accessibility Template Remarks and explanations (a) A text equivalent for every

More information

Recognition of sign language gestures using neural networks

Recognition of sign language gestures using neural networks Recognition of sign language gestures using neural s Peter Vamplew Department of Computer Science, University of Tasmania GPO Box 252C, Hobart, Tasmania 7001, Australia vamplew@cs.utas.edu.au ABSTRACT

More information

Practical Approaches to Comforting Users with Relational Agents

Practical Approaches to Comforting Users with Relational Agents Practical Approaches to Comforting Users with Relational Agents Timothy Bickmore, Daniel Schulman Northeastern University College of Computer and Information Science 360 Huntington Avenue, WVH 202, Boston,

More information

Body-conductive acoustic sensors in human-robot communication

Body-conductive acoustic sensors in human-robot communication Body-conductive acoustic sensors in human-robot communication Panikos Heracleous, Carlos T. Ishi, Takahiro Miyashita, and Norihiro Hagita Intelligent Robotics and Communication Laboratories Advanced Telecommunications

More information

Avaya G450 Branch Gateway, Release 7.1 Voluntary Product Accessibility Template (VPAT)

Avaya G450 Branch Gateway, Release 7.1 Voluntary Product Accessibility Template (VPAT) Avaya G450 Branch Gateway, Release 7.1 Voluntary Product Accessibility Template (VPAT) can be administered via a graphical user interface or via a text-only command line interface. The responses in this

More information

Summary Table Voluntary Product Accessibility Template. Supporting Features. Supports. Supports. Supports. Supports

Summary Table Voluntary Product Accessibility Template. Supporting Features. Supports. Supports. Supports. Supports Date: March 31, 2016 Name of Product: ThinkServer TS450, TS550 Summary Table Voluntary Product Accessibility Template Section 1194.21 Software Applications and Operating Systems Section 1194.22 Web-based

More information

Artificial Intelligence

Artificial Intelligence Artificial Intelligence COMP-241, Level-6 Mohammad Fahim Akhtar, Dr. Mohammad Hasan Department of Computer Science Jazan University, KSA Chapter 2: Intelligent Agents In which we discuss the nature of

More information

IAT 355 Perception 1. Or What You See is Maybe Not What You Were Supposed to Get

IAT 355 Perception 1. Or What You See is Maybe Not What You Were Supposed to Get IAT 355 Perception 1 Or What You See is Maybe Not What You Were Supposed to Get Why we need to understand perception The ability of viewers to interpret visual (graphical) encodings of information and

More information

HCS 7367 Speech Perception

HCS 7367 Speech Perception Babies 'cry in mother's tongue' HCS 7367 Speech Perception Dr. Peter Assmann Fall 212 Babies' cries imitate their mother tongue as early as three days old German researchers say babies begin to pick up

More information

Providing Effective Communication Access

Providing Effective Communication Access Providing Effective Communication Access 2 nd International Hearing Loop Conference June 19 th, 2011 Matthew H. Bakke, Ph.D., CCC A Gallaudet University Outline of the Presentation Factors Affecting Communication

More information

Communication (Journal)

Communication (Journal) Chapter 2 Communication (Journal) How often have you thought you explained something well only to discover that your friend did not understand? What silly conversational mistakes have caused some serious

More information

How do you design an intelligent agent?

How do you design an intelligent agent? Intelligent Agents How do you design an intelligent agent? Definition: An intelligent agent perceives its environment via sensors and acts rationally upon that environment with its effectors. A discrete

More information

Sperling conducted experiments on An experiment was conducted by Sperling in the field of visual sensory memory.

Sperling conducted experiments on An experiment was conducted by Sperling in the field of visual sensory memory. Levels of category Basic Level Category: Subordinate Category: Superordinate Category: Stages of development of Piaget 1. Sensorimotor stage 0-2 2. Preoperational stage 2-7 3. Concrete operational stage

More information

Institute of Communication Acoustics Chair of Information Technology. Ruhr-Universität Bochum

Institute of Communication Acoustics Chair of Information Technology. Ruhr-Universität Bochum Institute of Communication Acoustics Chair of Information Technology Ruhr-Universität Bochum Institute of Communication Acoustics Institute of Communication Acoustics Ruhr-Universität Bochum 2 Institute

More information

Multimodal interactions: visual-auditory

Multimodal interactions: visual-auditory 1 Multimodal interactions: visual-auditory Imagine that you are watching a game of tennis on television and someone accidentally mutes the sound. You will probably notice that following the game becomes

More information

Hear Better With FM. Get more from everyday situations. Life is on

Hear Better With FM. Get more from everyday situations. Life is on Hear Better With FM Get more from everyday situations Life is on We are sensitive to the needs of everyone who depends on our knowledge, ideas and care. And by creatively challenging the limits of technology,

More information

Infant Hearing Development: Translating Research Findings into Clinical Practice. Auditory Development. Overview

Infant Hearing Development: Translating Research Findings into Clinical Practice. Auditory Development. Overview Infant Hearing Development: Translating Research Findings into Clinical Practice Lori J. Leibold Department of Allied Health Sciences The University of North Carolina at Chapel Hill Auditory Development

More information