Recalibration of temporal order perception by exposure to audio-visual asynchrony Vroomen, Jean; Keetels, Mirjam; de Gelder, Bea; Bertelson, P.

Similar documents
Tilburg University. Auditory grouping occurs prior to intersensory pairing Keetels, Mirjam; Stekelenburg, Jeroen; Vroomen, Jean

Visual motion influences the contingent auditory motion aftereffect Vroomen, Jean; de Gelder, Beatrice

Tilburg University. The spatial constraint in intersensory pairing Vroomen, Jean; Keetels, Mirjam

Adaptation to motor-visual and motor-auditory temporal lags transfer across modalities

Tilburg University. Directing Spatial Attention towards the Illusory Location of a Ventriloquized Sound Vroomen, Jean; Bertelson, P.

Quantifying temporal ventriloquism in audiovisual synchrony perception

Recent studies have shown that the brain can adjust the processing

Quantifying temporal ventriloquism in audio-visual synchrony perception

Multimodal interactions: visual-auditory

Report. Direction of Visual Apparent Motion Driven Solely by Timing of a Static Sound. Elliot Freeman 1,2, * and Jon Driver 2 1

PERCEPTION OF AUDITORY-VISUAL SIMULTANEITY CHANGES BY ILLUMINANCE AT THE EYES

Perception of Synchrony between the Senses

The effects of subthreshold synchrony on the perception of simultaneity. Ludwig-Maximilians-Universität Leopoldstr 13 D München/Munich, Germany

Modality Differences in Timing: Testing the Pacemaker Speed Explanation

New insights in audio-visual synchrony perception

Effects of stimulus duration on audio-visual synchrony perception

UNCORRECTED PROOF. Multisensory integration, perception and ecological validity. Beatrice De Gelder 1,2 and Paul Bertelson 1,3

Perceptual congruency of audio-visual speech affects ventriloquism with bilateral visual stimuli

Bayesian Inference Explains Perception of Unity and Ventriloquism Aftereffect: Identification of Common Sources of Audiovisual Stimuli

Auditory temporal modulation of the visual Ternus effect: the influence of time interval

Direction of visual apparent motion driven by perceptual organization of cross-modal signals

Some methodological aspects for measuring asynchrony detection in audio-visual stimuli

Audio-visual synchrony perception

Neuroscience Letters

On the physiological location of otoacoustic emissions

Visual and Auditory Velocity Perception and Multimodal Illusions. Katherine S. Gasaway. Advisor: Paul M. Corballis, PhD

Supplemental Information: Task-specific transfer of perceptual learning across sensory modalities

Simultaneity constancy

TEMPORAL ORDER JUDGEMENTS A SENSITIVE MEASURE FOR MEASURING PERCEPTUAL LATENCY?

Audio-Visual Speech Timing Sensitivity Is Enhanced in Cluttered Conditions

Illusory reversal of temporal order: the bias to report a dimmer stimulus as the first

The Attentional Blink is Modulated by First Target Contrast: Implications of an Attention Capture Hypothesis

Published in: Proceedings of 4th Congress of the Federation of Acoustical Societies of Europe, FASE '84, Sandefjord, August 21-24, 1984

Lihan Chen. Crossmodal Temporal Capture in Visual and Tactile Apparent Motion: Influences of temporal structure and crossmodal grouping

Ten recommendations for Osteoarthritis and Cartilage (OAC) manuscript preparation, common for all types of studies.

Visual dominance and attention: The Colavita effect revisited

Published in: Proceedings of 15th International Congress on Acoustics, ICA'95, Trondheim, Norway, June 26-30, 1995

Selective bias in temporal bisection task by number exposition

Auditory Scene Analysis

Tri-modal integration of visual, tactile and auditory signals for the perception of sequences of events

Fundamentals of Psychophysics

Evaluation of a Danish speech corpus for assessment of spatial unmasking

Low-Level Visual Processing Speed Modulates Judgment of Audio-Visual Simultaneity

Why Seeing Is Believing: Merging Auditory and Visual Worlds

Coexistence of Multiple Modal Dominances

Attention enhances feature integration

Seeing Sound: Changing Visual Perception Through Cross-Modal Interaction. Tracey D. Berger. New York University. Department of Psychology

A Multimodal Paradigm for Investigating the Perisaccadic Temporal Inversion Effect in Vision

Grouping and Segregation of Sensory Events by Actions in Temporal Audio-Visual Recalibration

Effect of pitch space correspondence on sound induced visual motion perception

Categorical Perception

Exploring a brightness-drag illusion. Author. Published. Journal Title DOI. Copyright Statement. Downloaded from. Griffith Research Online

The influence of regularities on temporal order judgments

Auditory dominance over vision in the perception of interval duration

Behavioural evidence for separate mechanisms of audiovisual temporal binding as a function of leading sensory modality

Perceived Audiovisual Simultaneity in Speech by Musicians and Non-musicians: Preliminary Behavioral and Event-Related Potential (ERP) Findings

Conscious control of movements: increase of temporal precision in voluntarily delayed actions

INTRODUCTION J. Acoust. Soc. Am. 103 (2), February /98/103(2)/1080/5/$ Acoustical Society of America 1080

Influences of intra- and crossmodal grouping on visual and

Recalibrating the body: visuotactile ventriloquism aftereffect

Introduction to the Special Issue on Multimodality of Early Sensory Processing: Early Visual Maps Flexibly Encode Multimodal Space

Semantic factors in uence multisensory pairing: a transcranial magnetic stimulation study

Theoretical Neuroscience: The Binding Problem Jan Scholz, , University of Osnabrück

Segregation from direction differences in dynamic random-dot stimuli

The role of modality congruence in the presentation and recognition of taskirrelevant stimuli in dual task paradigms.

Vision Research. Audiovisual integration of stimulus transients. Tobias S. Andersen a,b, *, Pascal Mamassian c. abstract

Effects of Auditory Input on a Spatial Serial Response Time Task

Psychophysical Methods

The ventriloquist effect does not depend on the direction of automatic visual attention

Auditory modulation of visual apparent motion with short spatial and temporal intervals

Time Order as Psychological Bias

The role of multisensory integration in the bottom-up and top-down control of attentional object selection

From where does the content of a certain geo-communication come? semiotics in web-based geo-communication Brodersen, Lars

Significance of a notch in the otoacoustic emission stimulus spectrum.

Sensory Adaptation within a Bayesian Framework for Perception

Optimal Audiovisual Integration in People with One Eye

Supporting Information

A model of parallel time estimation

Consciousness and Cognition

Time- order effects in timed brightness and size discrimination of paired visual stimuli

The perception of motion transparency: A signal-to-noise limit

Three methods for measuring perception. Magnitude estimation. Steven s power law. P = k S n

University of Dundee. Volunteering Scott, Rosalind. Publication date: Link to publication in Discovery Research Portal

Acta Psychologica 134 (2010) Contents lists available at ScienceDirect. Acta Psychologica. journal homepage:

Coordination in Sensory Integration

Impaired perception of temporal order in auditory extinction

Comparing Bayesian models for multisensory cue combination without mandatory integration

Sequential Effects in Spatial Exogenous Cueing: Theoretical and Methodological Issues

Neural basis of auditory-induced shifts in visual time-order perception

Changing expectations about speed alters perceived motion direction

Object Substitution Masking: When does Mask Preview work?

PERFORMANCE OPERATING CHARACTERISTICS FOR SPATIAL AND TEMPORAL DISCRIMINATIONS: COMMON OR SEPARATE CAPACITIES?

STUDY OF THE BISECTION OPERATION

Hearing What the Eyes See

USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES

REACTION TIME AS A MEASURE OF INTERSENSORY FACILITATION l

Audio-Visual Information Processing Across. Different Age Groups

Validity of Haptic Cues and Its Effect on Priming Visual Spatial Attention

TIME-ORDER EFFECTS FOR AESTHETIC PREFERENCE

Transcription:

Tilburg University Recalibration of temporal order perception by exposure to audio-visual asynchrony Vroomen, Jean; Keetels, Mirjam; de Gelder, Bea; Bertelson, P. Published in: Cognitive Brain Research Publication date: 2004 Link to publication Citation for published version (APA): Vroomen, J., Keetels, M. N., de Gelder, B., & Bertelson, P. (2004). Recalibration of temporal order perception by exposure to audio-visual asynchrony. Cognitive Brain Research, 22(1), 32-35. General rights Copyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright owners and it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights. - Users may download and print one copy of any publication from the public portal for the purpose of private study or research - You may not further distribute the material or use it for any profit-making activity or commercial gain - You may freely distribute the URL identifying the publication in the public portal Take down policy If you believe that this document breaches copyright, please contact us providing details, and we will remove access to the work immediately and investigate your claim. Download date: 28. dec. 2017

Cognitive Brain Research 22 (2004) 32 35 Research report Recalibration of temporal order perception by exposure to audio-visual asynchrony Jean Vroomen a, *, Mirjam Keetels a, Beatrice de Gelder a,b, Paul Bertelson a,b a Department of Psychology, Tilburg University, Warandelaan 2, Tilburg 500 LE, The Netherlands b Université Libre de Bruxelles, Brussels, Belgium Accepted 6 July 2004 Available online 26 August 2004 www.elsevier.com/locate/cogbrainres Abstract The perception of simultaneity between auditory and visual information is of crucial importance for maintaining a coordinated representation of a multisensory event. Here we show that the perceptual system is able to adaptively recalibrate itself to audio-visual temporal asynchronies. Participants were exposed to a train of sounds and light flashes with a constant time lag ranging from 200 (sound first) to +200 ms (light first). Following this exposure, a temporal order judgement (TOJ) task was performed in which a sound and light were presented with a stimulus onset asynchrony (SOA) chosen from 11 values between 240 and +240 ms. Participants either judged whether the sound or the light was presented first, or whether the sound and light were presented simultaneously or successively. The point of subjective simultaneity (PSS) was, in both cases, shifted in the direction of the exposure lag, indicative of recalibration. D 2004 Elsevier B.V. All rights reserved. Theme: Sensory systems Topic: Visual psychophysics and behavior Keywords: Intersensory perception; Recalibration; Aftereffects; Temporal order judgement 1. Introduction Most natural events are processed by a number of different neural mechanisms. For example, seeing and hearing a talker provides multisensory information that is processed by specialized visual and auditory neural pathways. Several behavioural and neurophysiological studies have now highlighted the crucial role that temporal synchrony plays in binding such intersensory information so that a coherent representation of an event is obtained [1,2,5]. If temporal co-occurrence is indeed of crucial importance, the question arises how synchronization in the brain is achieved, as there are not only differences in physical transmission time between sound and light, but neural pathways also often differ considerably in processing speed. Given that neural architectures change over lifetime * Corresponding author. Tel.: +31 13 4662394; fax: +31 13 4662067. E-mail address: j.vroomen@uvt.nl (J. Vroomen). and that experience or attention can alter the neural response time to preferred or attended stimuli (e.g., the law of prior entry states that attended objects are perceived more rapidly than unattended objects [8,9]), it seems that any synchronization mechanism would need to be flexible in order to properly perform its function. Here, we explore whether the perceptual system does indeed adapt to changes in the timing of intersensory events. The experiments build explicitly upon our work on adaptation to audio-visual spatial conflict [2,10]. The logic of the spatial conflict situation is that two sets of data, delivered in the auditory and visual modality, specify different locations but that other parameters, in particular synchronized timing, are those normally associated with a single event, and thus favour pairing of the two conflicting data. Provided that the conflicting data are indeed paired, their disagreement about location may be considered a misalignment of the sensory systems. Such misalignment manifests itself as an immediate bias of the perceived 0926-6410/$ - see front matter D 2004 Elsevier B.V. All rights reserved. doi:10.1016/j.cogbrainres.2004.07.003

J. Vroomen et al. / Cognitive Brain Research 22 (2004) 32 35 33 auditory location towards the visual distracter, and following prolonged exposure to intersensory discrepancy, leads to adaptation or recalibration that is observable as an aftereffect [6]. Recalibration, in essence, reduces the conflict between vision and audition by shifting the least reliable modality for spatial information the auditory one towards the more precise modality, in this case vision [12]. Following the same logic, we predicted that if temporally misaligned but spatially co-localized auditory and visual data are paired, then the intersensory temporal misalignment might also manifest itself both as an immediate bias and as an aftereffect. Immediate temporal bias has indeed been demonstrated, as for example the perceived occurrence of a flash is attracted towards a temporally misaligned sound burst [5,11] (note that in the temporal domain, sound attracts vision as time is more accurately coded in audition). Temporal aftereffects, though, have not been explored. Here we therefore examined whether aftereffects indicative of temporal recalibration might be observed as well. Participants were exposed to a series of sounds and light flashes with a fixed temporal offset for some time. Following this exposure phase, the point of subjective simultaneity of a sound and light flash was measured by obtaining psychometric functions on temporal order judgements (TOJ) in two different tasks. Participants either judged on test trials whether a sound or a light had appeared first, or they judged whether a sound and light were presented simultaneously or successively. Two different tasks were used instead of one to check whether strategic effects rather than adaptive sensory changes influenced the results. Strategic adjustments are likely to be different in one or the other task, but not so for sensory changes. True temporal recalibration should therefore manifest itself in that in both tasks, the point of subjective simultaneity is shifted towards the previously experienced temporal conflict. 2. Method 2.1. Participants Twenty students from Tilburg University participated. Half of them judged which modality appeared first (sound or light), the other half judged whether sound or light were presented simultaneously or successively. Participants reported normal hearing and normal or corrected-to-normal seeing. They were tested individually and were unaware of the purpose of the experiment. 2.2. Stimuli The auditory stimulus consisted of a 2000 Hz tone of 20 ms duration (5 ms fade-in and fade-out) presented at 82 db(a). Sounds were presented via a hidden loudspeaker placed at eye-level, 50 cm in front of the participant at a central location. The visual stimulus consisted of a 20 ms flash of a red LED (diameter of 1 cm, luminance of 40 cd/m 2 ), placed directly in front of the loudspeaker. 2.3. Design The experiment had two within-subjects factors: Exposure lag during the exposure phase ( 200, 100, 0, +100 and +200 ms with negative values indicating that the sound was presented first) and stimulus onset asynchrony (SOA) between the sound and light of the test stimulus ( 240, 120, 90, 60, 30, 0, +30, +60, +90, +120 and +240 ms). These factors yielded 55 equi-probable conditions for the experimental trials. Each condition was presented 12 times for a total of 660 trials. Trials were presented in 15 blocks of 44 trials each (four repetitions for each SOA), preceded by two warm-up trials. Exposure lag was constant within a block and SOA varied randomly. Exposure lag varied between successive blocks with order counterbalanced across participants. 2.4. Procedure Participants sat at a table in a dark and sound-proof booth. Head movements were precluded by a chin- and forehead-rest. Each block of experimental trials started with an exposure phase of 240 repetitions (3 min) of the sound light stimulus pair (ISI=750 ms) with a constant time lag between the sound and the light. To ensure that participants were looking at the light during exposure, they had to attend the position where the light was presented. Unpredictably, a small green LED (a catch trial) could at that position be flashed once for 50 ms, and this occurred two, three, or four times during the exposure phase. Participants had to count and report at the end of exposure the number of catch trials. After a 10-s delay, the first trial then started. Each trial consisted of two parts: an audio-visual reexposure phase followed by the presentation of a sound and light whose temporal order had to be judged. The reexposure phase consisted of a train of eight sounds and lights with the same time lag as was used during the exposure phase. After 1000 ms, the sound and light of the test stimulus were presented (with a variable SOA between them). The participants task was to judge either whether the sound or the light of the test stimulus was presented first, or to judge whether the sound and light were presented simultaneously or successively. Participants made an unspeeded response by pressing one of two designated keys on a response box. The next trial started 2000 ms after a response. Training was given prior to testing. Trials of the training session were not preceded by an exposure or re-exposure phase. Participants were either trained to distinguish soundfirst from light-first trials (one block of 240 versus +240 ms SOA, and one block of 120 versus +120 ms SOA), or to distinguish simultaneous from successive trials (one

34 J. Vroomen et al. / Cognitive Brain Research 22 (2004) 32 35 block of 240 and +240 ms versus 0 ms SOA, and one block of +120 and 120 versus 0 ms SOA). Each block consisted of 32 trials where each of the SOAs was presented equally often in random order. Whenever participants made an erroneous response, they received corrective feedback (a green LED flickering three times). Training continued until fewer than three erroneous responses were made within a block. Following this initial training, participants were exposed to the full range of SOAs without feedback in a single block of 154 trials (14 times the 11 SOAs used in the experiment proper). Testing lasted about 4 h, and was run on two consecutive days. 3. Results Trials of the training session and warm-up trials were excluded from further analyses. Performance on catch trials was flawless, indicating that participants were indeed looking at the light during exposure. For participants who judged whether the sound or the light appeared first, the percentage of dlight-firstt responses was calculated for each participant and for each combination of exposure lag ( 200, 100, 0, +100 and +200 ms) and SOA (from 240 to +240 ms). For each of the thus obtained distribution of responses, an individually determined psychometric function was calculated by fitting a cumulative normal distribution using maximum likelihood estimation. The mean of the resulting distribution (the interpolated 50% crossover point) is the point of subjective simultaneity (henceforth the PSS), and the slope is a measure of the sharpness with which stimuli are distinguished from one another. For participants who judged whether the sound and the light were presented simultaneously or successively, the percentage of dsimultaneoust responses was calculated for each participant and for each combination of exposure lag and SOA. As the SOA varied from 240 to +240 ms, the probability of responding dsimultaneoust increased and then decreased. The resultant distribution of responses were fitted with a Gaussian function using maximum-likelihood estimation. Three defining parameters were estimated for each participant: the mean of the distribution providing a measure of the PSS, the peak height and the standard deviation. Results showed, as expected, that in both tasks exposure to an audio-visual asynchrony shifted the PSS in the direction of the lag (Fig. 1), while there was no effect of exposure lag on the other estimated parameters (all F sb1). Thus, following exposure to sound-first adapters ( 200 or 100 ms exposure), a sound of the test stimulus had to be presented earlier to the light in order to be perceived as simultaneous if compared with exposure to light-first adapters (+100 and +200 ms). In the overall ANOVA on the PSS, there was a highly significant effect of exposure lag, F(4,72)=24.72, pb0.001, with no significant overall difference between the two tasks ( Fb1). The interaction between exposure lag and Fig. 1. The average point of subjective simultaneity (PSS) as a function of the audio-visual lag in the exposure phase. Error bars represent S.E.M. across participants. Participants either judged whether the sound or the light of the test stimulus was presented first (continuous line), or they judged whether the sound and light of the test stimulus were presented simultaneously or successively (dotted line). The PSS shifted, in both cases, towards the lag of the exposure phase. task was not significant ( F(4,72)=2.6, p=0.07), although there was a tendency that, at an exposure lag of 200 ms, the effect of exposure was somewhat bigger in the simultaneous/successive task than in the sound-first/lightfirst task. Possibly, this may reflect that the simultaneous/ successive task is more sensitive to shifts in criterion. When the effect of exposure lag was partitioned into linear, quadratic, cubic and higher order trends, there was a significant linear trend, F(1,18)=57.98, pb0.001, indicating that the PSS shifted, on average, about 6.7% in the direction of the lag. There was also a significant cubic trend F(1,18)=7.94, pb0.01, indicating that the effect levelled off when the exposure lag reached ~+200 or ~ 200 ms. 4. Discussion The present study showed for the first time that the point of subjective audio-visual simultaneity can be shifted towards a previously experienced temporal lag. This shift is interpreted as a manifestation of temporal recalibration: That is, when temporally offset but spatially co-localized audio-visual stimuli are paired, the criterion for simultaneity is shifted accordingly. The size of the shift is of the same order of magnitude as has been reported for recalibration in the spatial domain [7]. Moreover, the effect of exposure lag levelled off around ~F100 ms, which is also around the limit where a sound can capture the perceived onset of a light [5,11]. Beyond this limit, it becomes more likely that

J. Vroomen et al. / Cognitive Brain Research 22 (2004) 32 35 35 data in the two modalities are not paired anymore, in which case there is no need for the perceptual system to recalibrate. 1 Similar audio-visual induced aftereffects have now also been observed in the perception of space and speech [1,2,3]. The rule in all these cases is that whenever there is a moderate conflict between what is heard and what is seen, the brain takes advantage of the strength of each modality, such that the information that is most accurately coded in one modality changes the information encoded in the other, less accurate modality (see also Ref. [4]). In the present case, one might thus expect that since temporal information is more accurately coded in audition, vision has shifted towards audition. Admittedly, though, several other possibilities remain valid, as one might also argue that audition has been shifted towards vision, or it might even be the case that only the specific relation between the two modalities has changed. In future research, one may distinguish between these alternatives by checking whether adaptation to audio-visual temporal conflict generalizes to visual, auditory, or audio-visual tasks that measure the temporal occurrence of a stimulus. One reasonable explanation how intersensory temporal recalibration might occur is that multi-modal neurons in the brain shift their temporal alignment preference during the exposure period to the discrepant stimuli. This implies that the intersensory representation of an object or event is dynamic, presumably to account for differences in the processing capacities of each sensory system. This may suggests that the way in which different sensory modalities are coordinated in time may not be because the brain employs a wide temporal window of integration, but because the window is actively changed, depending on previous experience. References [1] P. Bertelson, Ventriloquism: a case of crossmodal perceptual grouping, in: G. Aschersleben, T. Bachmann, J. Mqsseler (Eds.), Cognitive Contributions to the Perception of Spatial and Temporal Events, Elsevier, North Holland, 1999, pp. 347 363. [2] P. Bertelson, B. De Gelder, The psychology of multimodal perception, in: C. Spence, J.D. Driver (Eds.), Crossmodal Space and Crossmodal Attention, Oxford University Press, Oxford, 2004. [3] P. Bertelson, J. Vroomen, B. De Gelder, Visual recalibration of auditory speech identification: a McGurk aftereffect, Psychol. Sci. 14 (2003) 592 597. [4] M.O. Ernst, M.S. Banks, Humans integrate visual and haptic information in a statistically optimal fashion, Nature 415 (2002) 429 433. [5] S. Morein-Zamir, S. Soto-Faraco, A. Kingstone, Auditory capture of vision: examining temporal ventriloquism, Cogn. Brain Res. 17 (2003) 154 163. [6] M. Radeau, P. Bertelson, The after-effects of ventriloquism, Q. J. Exp. Psychol. 26 (1974) 63 71. [7] M. Radeau, P. Bertelson, Auditory visual interaction and the timing of inputs: Thomas (1941) revisited, Psychol. Res. 49 (1987) 17 22. [8] C. Spence, D.I. Shore, R.M. Klein, Multisensory prior entry, J. Exp. Psychol. Gen. 130 (2001) 799 832. [9] E.B. Titchener, Lectures on the Elementary Psychology of Feeling and Attention, Macmillan, New York, 1908. [10] J. Vroomen, B. De Gelder, Perceptual effects of cross-modal stimulation: the cases of ventriloquism and the freezing phenomenon, in: G. Calvert, C. Spence, B.E. Stein (Eds.), Handbook of Multisensory Processes, MIT Press, Cambridge, 2004, pp. 141 150. [11] J. Vroomen, B. De Gelder, Temporal ventriloquism: sound modulates the flash-lag effect, J. Exp. Psychol. Hum. Percept. Perform. 30 (2004) 513 518. [12] R.B. Welch, Perceptual Modification: Adapting to Altered Sensory Environments, Academic Press, New York, 1978. 1 In a control experiment (N=10) in which participants judged which modality appeared first, we examined dexcessivet lags of the adaptors at SOAs of +350 ms (light-much-before-sound) and 350 ms (sound-muchbefore-light). In this case, there was no shift of the PSS ( Fb1) confirming that the audio-visual lag in the exposure phase has to be within limits for recalibration to occur.