Estimating sparse spectro-temporal receptive fields with natural stimuli

Size: px
Start display at page:

Download "Estimating sparse spectro-temporal receptive fields with natural stimuli"

Transcription

1 Network: Computation in Neural Systems September 2007; 18(3): Estimating sparse spectro-temporal receptive fields with natural stimuli STEPHEN V. DAVID, NIMA MESGARANI, & SHIHAB A. SHAMMA Institute for Systems Research, University of Maryland, College Park, MD 20742, USA (Received 14 March 2007; revised 5 July 2007; accepted 2 August 2007) Abstract Several algorithms have been proposed to characterize the spectro-temporal tuning properties of auditory neurons during the presentation of natural stimuli. Algorithms designed to work at realistic signal-to-noise levels must make some prior assumptions about tuning in order to produce accurate fits, and these priors can introduce bias into estimates of tuning. We compare a new, computationally efficient algorithm for estimating tuning properties, boosting, to a more commonly used algorithm, normalized reverse correlation. These algorithms employ the same functional model and cost function, differing only in their priors. We use both algorithms to estimate spectro-temporal tuning properties of neurons in primary auditory cortex during the presentation of continuous human speech. Models estimated using either algorithm, have similar predictive power, although fits by boosting are slightly more accurate. More strikingly, neurons characterized with boosting appear tuned to narrower spectral bandwidths and higher temporal modulation rates than when characterized with normalized reverse correlation. These differences have little impact on responses to speech, which is spectrally broadband and modulated at low rates. However, we find that models estimated by boosting also predict responses to non-speech stimuli more accurately. These findings highlight the crucial role of priors in characterizing neuronal response properties with natural stimuli. Keywords: Auditory cortex, speech, reverse correlation, boosting Correspondence: Shihab A. Shamma, 2203 AV Williams Building, College Park, MD, sas@ eng.umd.edu ISSN X print/issn online/07/ ß 2007 Informa UK Ltd. DOI: /

2 192 S. V. David et al. Introduction Recently, interest has grown in characterizing the response properties of sensory neurons under natural stimulus conditions (Theunissen et al. 2001; David and Gallant 2005; Wu et al. 2006). This approach is useful for understanding how well functional models developed using simple synthetic stimuli generalize to conditions encountered during natural behavior (Theunissen et al. 2000; David et al. 2004). In this study, we evaluate a new method for characterizing functional properties from their responses to natural stimuli: boosting (Friedman et al. 2000; Willmore et al. 2005; Zhang and Yu 2005; Willmore et al. 2006). In auditory systems, the functional properties of single neurons are often described by a spectro-temporal receptive field (STRF) (Kowalski et al. 1996; Theunissen et al. 2001; Sahani and Linden 2003; Machens et al. 2004). The STRF is a linear model in the spectral domain of sound and is well-suited for systems identification methodologies. Classically, measurements of STRFs have been restricted to the domain of white noise stimuli because spike-triggered averaging, the most commonly used estimation method, produces biased estimates for nonwhite stimuli, including natural stimuli (Theunissen et al. 2001). More recently, a new method, normalized reverse correlation (NRC), has been developed that converges on unbiased STRF estimates for arbitrary, non-white stimuli (Theunissen et al. 2001), along with some slight variants (Willmore and Smyth 2003; Machens et al. 2004). Studies using NRC have identified large significant changes in tuning under natural stimulus conditions that are not observed during stimulation by synthetic stimuli both in auditory (Theunissen et al. 2001; Woolley et al. 2005) and visual systems (David et al. 2004). Because natural stimuli contain strong autocorrelations, some dimensions have much lower variance than others (Theunissen et al. 2001). Signal-to-noise levels for response properties along the low-variance dimensions tend to be poor. Algorithms for STRF estimation such as NRC require regularization to prevent overfitting to noise along the low-variance dimensions. Regularization introduces a prior that imposes a constraint on the STRF estimate, independent of the data to be fit. Under noisy conditions, a prior can bias STRF estimates, and it is critical to account for the effects of this bias when drawing conclusions about tuning (David et al. 2004; Woolley et al. 2005). Since the development of NRC, several other methods have been developed for estimating receptive field properties using natural stimuli (Ringach et al. 2002; Willmore and Smyth 2003; Prenger et al. 2004; Sharpee et al. 2004; Touryan et al. 2005; Willmore et al. 2005; Rapela et al. 2006). These algorithms are largely variants of gradient descent techniques, in which STRF estimates are repeatedly adjusted by small amounts in the direction that minimizes fit error. Gradient descent algorithms can be used to fit data with the same functional model (linear STRF) and cost function (minimum mean-squared error) as NRC; thus they converge on the same theoretical solution as NRC. However, each algorithm uses a different prior than NRC, which can bias STRF estimates for finite noisy data (Wu et al. 2006). One new method for STRF estimation, boosting, is appealing because it requires a relatively light computational load and provides easily interpretable results (Friedman et al. 2000; Zhang and Yu 2005). This method has been shown to be effective for characterizing spatiotemporal responses properties in extrastriate visual

3 Sparse receptive fields from natural stimuli 193 cortex (Willmore et al. 2005, 2006). However, a direct comparison has not yet been made between boosting and other STRF estimation methods. In this study, we compare STRFs estimated by boosting and NRC under natural stimulus conditions. We estimate STRFs from the responses of isolated neurons in primary auditory cortex (A1) to continuous human speech (Garfolo 1988). Speech, while not a complete sampling of all possible sounds, is a complex mammalian vocalization that shares high-order statistical properties with a broad class of natural sounds (Smith and Lewicki 2006). We compare boosted and NRC STRFs in terms of their predictive power and their basic tuning properties. We find that boosted STRFs predict responses to speech and synthetic non-speech stimuli significantly better than NRC STRFs. The differences in predictive power result directly from differences in the tuning bandwidth of STRFs estimated by the two methods. These results suggest that boosting provides a more accurate characterization of spectro-temporal tuning under these experimental conditions. Methods Experimental procedures Auditory response properties were recorded from single neurons in primary auditory cortex (A1) of five awake passive ferrets. All experimental procedures were conformed to standards specified by the National Institutes of Health and the University of Maryland Animal Care and Use Committee. Surgical preparation. Animals were implanted with a stainless steel head post to allow for stable recording. While under anesthesia (nembutol and halothane), the skin and muscles on the top of the head were retracted from the central 4 cm diameter of skull. Several titanium bone screws were attached to the skull and the head post was glued on the midline. The entire site was then covered with dental acrylic and the skin around the implant was allowed to heal. Analgesics (banamine) and antibiotics (baytril) were administered under veterinary supervision until recovery. After recovery from surgery, animals were acclimatized to a custom restrainer where their heads were held fixed. A small craniotomy (1 mm diameter) was then drilled through the acrylic and skull over A1. The craniotomy was cleaned daily to prevent infection. After recordings were completed in one hemisphere, the site was closed with a layer of dental acrylic, and the same procedure was repeated in the other hemisphere. Neurophysiology. Single unit activity was recorded using tungsten micro-electrodes (1 5 MO, FHC). For three animals a single electrode was used and its position was controlled using a pneumatic microdrive (Narishige). Electrical activity was amplified with a custom amplifier, converted to a digital signal (National Instruments), and recorded using a custom data acquisition system. For the remaining two animals, one to four electrodes were controlled by independent microdrives and activity was recorded using a commercial data acquisition system

4 194 S. V. David et al. (Alpha-Omega). For all experiments, spike events were extracted from the raw electrophysiological recordings using custom template matching software. Only single units with clear (>95%) isolation were included in the analysis data set. Stimuli were presented from digital recordings using custom software. The digital signals were transformed to analog (National Instruments), equalized to achieve flat gain (Rane), and amplified (Radio Shack) to the desired sound level. These signals were presented through a small speaker (Etymotics) inserted in the ear contralateral to the recording site. Before each recording, the equalizer was calibrated to the acoustical properties of the speaker insertion. Upon identification of a recording site with isolatable units, a sequence of random tones (100 ms duration, 500 ms separation) was used to measure latency and spectral tuning. Neurons were verified as being in A1 by their tonotopic organization, latency and simple frequency tuning curves (Kowalski et al. 1996; Bizley et al. 2005). Speech stimuli. Continuous speech stimuli were drawn from a standard speech database (TIMIT; Garfolo 1988). Stimuli were sentences (3 4 sec), each sampled from a different speaker (balanced across gender). For each neuron, different sentences were presented at db SPL for 5 repetitions. The original stimuli were recorded at 16 khz but were upsampled to 44 khz before presentation. Temporally orthogonal ripple stimuli. In order to test how well STRFs generalize across stimulus classes, responses were also recorded to temporally orthogonal ripple combinations (TORCs; Klein et al. 2000). TORCs provide an efficient probe of spectro-temporal response properties by approximating white noise in the spectral domain. The set of 30 TORCs probed tuning over 5 octaves ( Hz or Hz, chosen to span the range of frequencies at which pure tones drive responses), with a spectral resolution of 1.2 cycles per octave and temporal envelope resolution of 24 Hz or 48 Hz (5 repetitions, 3 s duration, 44 khz sampling). Spectro-temporal receptive field estimation Linear spectro-temporal model. Neurons in A1 are tuned to stimulus frequency but are rarely phase-locked to oscillations of the sound waveform (Kowalski et al. 1996; Bizley et al. 2005). To describe tuning properties of such neurons, it is useful to represent auditory stimuli in terms of their spectrogram. The spectrogram of a sound waveform transforms the stimulus into a time-varying function of energy in each frequency band (Figure 1). This transformation removes the phase of the carrier signal so that the mapping from the spectrogram to neuronal firing rate can be described by a linear function. For a stimulus spectrogram s(x, t) and instantaneous neuronal firing rate r(t) sampled at times t ¼ 1,..., T, the linear STRF is defined as the linear mapping

5 Sparse receptive fields from natural stimuli 195 (A) (B) Frequency (Hz) Response (spikes/sec) Time (ms) max Power Figure 1. (A) Continuous speech stimulus presented to neurons in primary auditory cortex (A1). To estimate the STRF, the raw stimulus waveform (top) was filtered by a bank of logarithmically spaced gammatone filters that simulate the representation of stimuli in the auditory nerve. Filter center frequencies ranged from 100 to 8000 Hz. (B) The peristimulus time histogram response of a single neuron to five repeated presentations of the stimulus in (A), 10 ms bins. This neuron shows clear responses to syllables with power near 700 Hz. (Kowalski et al. 1996; Klein et al. 2000; Theunissen et al. 2001), rðtþ ¼X, U x¼1, u¼0 hðx, uþsðx, t uþþeðtþ Each coefficient of h indicates the gain applied to frequency channel x at time lag u. Positive values indicate components of the stimulus correlated with increased firing, and negative values indicate components correlated with decreased firing. The residual, e(t), represents components of the response (nonlinearities and noise) that cannot be predicted by the linear model. To simplify the description of estimation algorithms used in this study, we rewrite the STRF in a linear algebraic form. First, we define the STRF as a vector, h, by concatenating the STRF gain at each frequency and time lag into a single vector of length Y ¼ XU, 2 3 hð1, 1Þ hð2, 1Þ.. h ¼ hð1, 2Þ ð2þ hðx, UÞ We then define a T Y stimulus matrix, S, such that each row, t, contains the stimulus coefficients at time t and the preceding u time bins, 2 3 sð1, 1Þ sð2, 1Þ sð1, 0Þ sðx, 1 UÞ sð1, 2Þ sð2, 2Þ sð1, 1Þ sðx, 2 UÞ 6 S ¼ ð3þ 5 sð1, TÞ sð2, TÞ sð1, T 1Þ sðx, T UÞ ð1þ

6 196 S. V. David et al. The response, a vector of length T, is simply the product of the stimulus and STRF, plus a vector of residuals, e, r ¼ Sh þ e ð4þ STRF estimation by boosting. STRFs were estimated from the responses to the speech stimuli by boosting (Zhang and Yu 2005). Boosting converges on an unbiased estimate of the linear mapping between stimulus and response, regardless of autocorrelation in the stimulus. Several versions of boosting algorithms exist that can be used to estimate STRFs. In this study, we used forward stagewise fitting, which employs a simple iterative algorithm (Friedman et al. 2000). Initially, the STRF is set to zero, h 0 ¼ 0. During each iteration, i, the meansquared error is calculated for the prediction after incrementing or decrementing each STRF parameter by a small amount, ". All possible increments are specified by two parameters, an index, ¼ 1,..., Y, and a sign, ¼ 1 or 1, such that, h, ðyþ ¼", y ¼ ð5þ ¼ 0, otherwise The best increment is the one that produces the largest decrease in the mean-squared error, X T 2 ðyi, ziþ ¼arg min r Sh i 1 þ h, ð6þ, t¼1 The increment is added to the STRF, h i ¼ h i 1 þ h yi, zi ð7þ and this procedure is repeated until an additional increment/decrement only adds noise to the STRF estimate. Implementing boosting requires two hyperparameters, i.e., parameters that affect the final STRF estimate but that are not contained explicitly in the stimulus/ response data. These are (1) the step size, ", and (2) the number of iterations to complete before stopping. Generally, step size can be made arbitrarily small for the best fits. Extremely small step sizes are computationally inefficient, as they require many iterations to converge. We fixed " to be a small fraction of the ratio between stimulus and response variance, sffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi " ¼ 1 varðrðtþþ ð8þ 50 varðsðx, tþþ Here, stimulus variance is averaged across all spectral channels. Despite differences in variance across spectral channels, this heuristic produced accurate estimates and required relatively little computation time. Increasing or decreasing " by a factor of two had no effect on STRFs. Of course, different values of " are likely to be optimal for different data sets. To optimize the second hyperparameter, the number of iterations to complete before stopping, we used cross validation and early stopping. We reserved a small part (5%) of the fit data from the main boosting procedure. After each iteration, we tested the ability of the STRF to predict responses in the reserved set. The optimal

7 Sparse receptive fields from natural stimuli 197 stopping point was determined as the iteration when the STRF failed to improve predictions. Note that this reserved set was distinct from the validation data used finally to evaluate each STRF (see further). STRF estimation by NRC. We also estimated STRFs from the same data using NRC, a variant of classical linear regression that has been used to estimate STRFs from natural stimuli in the visual and auditory systems (Theunissen et al. 2000; David et al. 2004). Like boosting, NRC fits a linear STRF that minimizes the meansquared error between predicted and observed neuronal response. A detailed description of a computationally efficient NRC algorithm is described in (Theunissen et al. 2001). Here, we provide a brief description of NRC for comparison to boosting. In a system with no noise, the minimum mean-squared error estimate of the STRF for response vector, r, and stimulus matrix, S, is, h ideal ¼ 1 T C 1 ss S T r ð9þ The matrix Css 1 is the inverse of the Y Y stimulus autocorrelation matrix, C ss ¼ S T S/T, and the superscript T indicates the transpose operation. This estimate is theoretically optimal, but, in practice, estimating the STRF by Equation 9 for stimuli with strong autocorrelations can amplify noise excessively (Figure 6B). In a correlated stimulus, some dimensions have higher variance than others. Low variance dimensions provide very little modulating energy and thus make it difficult to measure correlations between those dimensions and the response. The inverse autocorrelation matrix normalizes the variance along each dimension to be the same. When variance is low, normalization requires dividing by a small number and often amplifies noise in parameter estimates. To minimize these effects, NRC uses a pseudoinverse to approximate the inverse of the stimulus autocorrelation matrix. Dimensions that are below some noise threshold are forced to be zero. To compute the pseudoinverse, a singular value decomposition is applied to the autocorrelation matrix, C ss ¼ U T U The columns of U contain the unit-norm eigenvectors of C ss, and the diagonal matrix ¼ diag( 1, 2,..., Y ) contains the corresponding eigenvalues ordered from largest to smallest. A tolerance value,, specifies the fraction of stimulus variance to preserve in the pseudoinverse. The number of stimulus dimensions to preserve, m, is computed, ð10þ 1 þ 2 þþ m m ¼ arg max < 1 þ 2 þþ m þþ Y The pseudoinverse, Capprox 1, is then computed, Capprox 1 ¼ U 1 approx U T ¼ Udiag 1, 1 1 2,..., 1 m, 0,..., 0 U T ð11þ ð12þ

8 198 S. V. David et al. And the final NRC estimate of the STRF is, h NRC ¼ 1 T C 1 approx ST r ð13þ Implementing NRC requires the selection of a single hyperparameter, the tolerance value,. To choose, we used a cross-validation procedure similar to the procedure for selecting the number of iterations in the boosting algorithm. A small subset was reserved from the estimation data before using Equation 13 to compute h NRC. The STRF was estimated for a range of tolerance values, ¼ (i.e., , Figure 5B). Each STRF was then used to predict responses in the reserved data and the tolerance value that produced the best prediction was used for the final STRF estimate. As in the case of boosted STRFs, the final STRF was then validated with a separate data set that was not used at any stage of estimation (see subsequently). Data preprocessing. The same preprocessing was applied to the speech data prior to STRF estimation by boosting and NRC. Spectrograms were generated from the stimulus sound pressure waveforms using a 128-channel rectifying filter bank that simulated processing by the auditory nerve (Yang et al. 1992). Filters had center frequencies ranging from 100 to 8000 Hz, were spaced logarithmically, and had a bandwidth of approximately 1/12 octave (Q 3dB ¼ 12). The output of the filters is proportional to the amplitude (square root of power) of the stimulus spectrogram. To improve the signal-to-noise of STRF estimates, the output of the filter bank was smoothed across frequency and downsampled to 24 channels. In some cases, STRF performance can be improved by applying a compressive nonlinearity to the stimulus spectrogram, to simulate processing in the auditory nerve (Gill et al. 2006). When we applied a logarithm to the stimulus spectrogram prior to STRF estimation, we observed a small although insignificant decrease in prediction accuracy for both boosted and NRC STRFs. This transformation had no effect on the significance of findings in this study. Thus the results reported here were computed without a compressive nonlinearity. A more biologically plausible transformation that includes threshold and saturation in the auditory nerve is likely to improve performance (Gill et al. 2006). Spike rates were computed by averaging responses over the five repeated stimulus presentations. Both the stimulus spectrogram and spike rates were finally binned at 10 ms resolution. This data preprocessing requires the selection of two additional hyperparameters, the spectral and temporal sampling density. The respective values of 1/4 octave and 10 ms were chosen to match the spectral resolution of critical bands (Zwicker 1961) and temporal resolution of neurons in A1 (Schnupp et al. 2006). Increased binning resolution would change the number of fit parameters and could, in theory, affect the performance of the two estimation algorithms. Since typical NRC estimates are substantially smoothed, even at the resolution used in this study (Figure 5), the performance of NRC is likely not to be affected by increased resolution. Changing spectro-temporal resolution might improve the performance of STRFs estimated by boosting (see Discussion ). However, the resolution used in this study is sufficient for demonstrating the improved performance of boosting over NRC in these experimental conditions.

9 Sparse receptive fields from natural stimuli 199 Validation procedure. A cross-validation procedure was used to make unbiased measurements of the accuracy of the different STRF models. From the entire speech data set, 95% of the data was used to estimate the STRF (estimation data set). The STRF was then used to predict the neuronal responses in the remaining 5% (validation data set), using the same 10 ms binning. This procedure was repeated 20 times, excluding a different validation segment on each repeat. Each STRF was then used to predict the responses in its corresponding validation data set. These predictions were concatenated to produce a single prediction of the neuron s response. Prediction accuracy was determined by measuring the correlation coefficient (Pearson s r) between the predicted and observed response. This procedure avoided any danger of overfitting or of bias from differences in model parameter and hyperparameter counts (David and Gallant 2005). Studies using natural stimuli have argued that A1 encodes natural sounds with 10 ms resolution (Schnupp et al. 2006), but in some conditions A1 neurons can respond with temporal resolution on the order of 4 ms (Furukawa and Middlebrooks 2002). Measurements of prediction accuracy by correlation ignore signals at resolutions finer than the temporal window (Theunissen et al. 2001); thus the correlation values reported in this study should not be interpreted as strict lower bounds on the portion of responses predicted by STRFs. Each STRF estimated from the speech data was also used to predict responses to the TORC stimuli. Reported prediction correlations for TORCs are averaged across the 20 STRFs estimated for cross validation. Projection of boosted STRFs into the NRC subspace Boosting and NRC differ in the priors they place on estimated STRFs. Boosting assumes a sparse STRF while NRC assumes that STRF power will exist only along stimulus dimensions that have high power. Given the prior used in NRC, STRFs estimated by NRC span only part of the stimulus space. For accurate comparison of STRFs estimated using the different procedures, it is necessary to project boosted STRFs into the same subspace as the NRC STRFs (David et al. 2004; Woolley et al. 2005). For each neuron, the boosted STRF, h b, was projected into the NRC subspace by introducing the pseudoinverse projection (David et al. 2004), h b=nrc ¼ C 1 approx C ssh b ð14þ Tuning properties derived from STRFs To compare STRFs estimated using the different stimulus algorithms, we measured a similarity index between them by computing their normalized pointwise correlation. An index value of 1 indicated an exact match of excitatory and inhibitory subfields, while a value of 0 indicated no correlation. We also measured four tuning properties commonly used to describe auditory neurons (Kowalski et al. 1996; Klein et al. 2000): best frequency, peak latency, spectral scale, and preferred modulation rate. Best frequency and peak latency were measured by computing the center of mass of the STRF after setting all negative

10 200 S. V. David et al. parameters to zero. Spectral scale, measured in cycles per octave (cyc/oct), is inversely proportional to spectral bandwidth; scale describes the spectral structure of a stimulus that most strongly drives the neuron. Preferred modulation rate, measured in cycles per second (Hz), is a complementary property to scale that describes the temporal modulation rate of a stimulus that best drives the neuron. Spectral scale and preferred rate were measured by computing the modulation transfer function of the STRF (i.e., the absolute value of its two-dimensional Fourier transform, (see Klein et al and Woolley et al. 2005), averaging the first quadrant and transpose of the second quadrant, and computing the center of mass of the average. Results Spectro-temporal receptive fields of A1 neurons during stimulation by continuous speech In order to study how human speech is represented in primary auditory cortex (A1), we recorded the responses of 164 isolated A1 neurons to continuous speech stimuli (Figure 1A). The stimuli were taken from a standard speech processing library and were sampled over an assortment of speakers, balanced between male and female (Garfolo 1988). We observed reliable responses to repeated stimulus presentations, often with brief bursts associated with the onset of syllables (Figure 1B). We characterized the functional relationship between the stimulus and neuronal response by estimating the STRF for each neuron from its responses to the speech stimulus. The STRF is a linear mapping from the stimulus spectrogram to the neuron s instantaneous firing rate response (Kowalski et al. 1996; Klein et al. 2000; Theunissen et al. 2001). Often, STRFs are estimated using a linear spectrogram representation of the stimulus. In order to use a more biologically plausible model, we used a slightly different spectrogram representation that simulates the output of the auditory nerve with a bank of logarithmically spaced gammatone filters (Yang et al. 1992). In the case of white noise stimuli (i.e., stimuli that are uncorrelated in their spectrogram representation), STRFs can be estimated simply by spike-triggered averaging (Kowalski et al. 1996; Klein et al. 2000; Theunissen et al. 2001). However, natural stimuli contain strong correlations over time and between spectral channels that require more complex algorithms to achieve unbiased STRF estimates (Theunissen et al. 2001; Singh and Theunissen 2003). We compared two algorithms for finding the minimum mean-squared error estimate of the STRF that differ only in their prior assumptions about model parameters. The first, boosting, iteratively increments single STRF parameters by a small amount until further increments fail to improve response predictions (Friedman et al. 2000; Willmore et al. 2005; Zhang and Yu 2005; Willmore et al. 2006). Boosting tends to minimize the number of nonzero parameters, effectively imposing a sparse prior on the STRF (Sahani and Linden 2003). The second algorithm, NRC, is a variant of classical linear regression that estimates the STRF only in the stimulus subspace with high signal-to-noise (Theunissen et al. 2001). Speech and other natural stimuli tend to have 1/f 2 power spectra (Singh and Theunissen 2003), and spectro-temporal features at high frequencies tend to be

11 Sparse receptive fields from natural stimuli 201 (A) Frequency (Hz) (B) 8000 Frequency (Hz) Time lag (ms) Time lag (ms) Figure 2. (A) Example of a STRF estimated from the speech stimulus by boosting. Red regions indicate frequencies and time lags for which high stimulus power is correlated with an increase in neuronal spiking. Blue regions indicate stimulus components correlated with a decrease in spiking. Boosting estimates the STRF by an iterative algorithm. During each iteration, the single parameter that best improves prediction accuracy is incremented or decremented by a small amount until further increments add noise. This procedure biases STRFs to be sparse, minimizing the number of nonzero parameters. (B) STRF estimated for the same neuron using NRC. NRC biases parameters associated with low-power dimensions of the stimulus to be zero. For natural stimuli such as speech, this introduces a smooth prior. Thus the NRC STRF has broader spectral and temporal tuning than the boosted STRF in (A). (STRFs have been upsampled by a factor of two and smoothed to facilitate visualization.) 0 1 Normalized gain excluded from the STRF estimated using NRC. Projection into this subspace effectively imposes a smooth prior on the STRF (Sahani and Linden 2003; Willmore and Smyth 2003). We found that the STRFs estimated for A1 neurons from the speech data had bandpass spectro-temporal tuning, typical of measurements with other stimuli (Kowalski et al. 1996). Figure 2A shows an example of an STRF estimated by boosting. This neuron is excited by a narrow range of frequencies (best frequency 700 Hz, spectral scale 0.49 cycles per octave) with a peak latency of 13 ms and preferred modulation rate of 23.4 Hz. In order to test the accuracy of the STRF, we used it to predict the response to a validation stimulus that was also played to the

12 202 S. V. David et al. neuron but was not used for STRF estimation (David and Gallant 2005). The STRF estimated using speech predicts the validation stimulus quite accurately, with a correlation coefficient of r ¼ 0.85 (10 ms time bins). The STRF estimated by NRC from the same speech data shows substantial differences from the STRF estimated by boosting (Figure 2B). Its best frequency (725 Hz) and peak latency (17 ms) are similar, but its spectral scale (0.40 cyc/oct) and preferred rate (14.3 Hz) are much lower than the boosted STRF. The lower scale and rate are reflected in the wide spectral and temporal spread of the excitatory region of the STRF in Figure 2B. Surprisingly, despite the differences in tuning, the NRC STRF predicts with nearly the same accuracy (r ¼ 0.83) as the STRF estimated by boosting. These predictions are not significantly different (randomized paired t-test, p > 0.2). Although STRFs estimated by boosting and NRC tend to have different spectral scale and rate tuning, they generally share basic features. Figure 3 compares STRFs estimated using the two algorithms for a different neuron. The boosted STRF (Figure 3A) has an excitatory domain centered at 1160 Hz and two smaller (A) 8000 Frequency (Hz) (B) Time lag (ms) 1 0 Normalized gain Frequency (Hz) Time lag (ms) Figure 3. (A) Example STRF of a neuron with more complex tuning estimated by boosting. This STRF has an excitatory region near 1160 Hz and two inhibitory regions centered at higher frequencies. (B) The STRF estimated by NRC for the same neuron shows a similar pattern of spectral tuning as in (A), but is smoother in time.

13 Sparse receptive fields from natural stimuli 203 inhibitory domains at higher frequencies, 1800 and 2500 Hz. The NRC STRF (Figure 3B) contains the same pattern of one excitatory domain at a lower frequency than two inhibitory domains. In this example, the preferred scale of the STRFs estimated by boosting (0.81 cyc/oct) and NRC (0.80 cyc/oct) are similar. However, the STRFs have different preferred rates. As in the previous example, the boosted STRF has a higher preferred rate (19.0 Hz) than the NRC STRF (11.4 Hz). When we compared the ability of these STRFs to predict responses in a validation data set, we found, as in the previous example, that the boosted STRF (r ¼ 0.44) performed slightly, but not significantly better than the NRC STRF (r ¼ 0.42, p > 0.2). Comparison of prediction accuracy for STRFs estimated by boosting and NRC We compared the ability of STRFs estimated by boosting and NRC to predict responses to speech across the entire set of 164 A1 neurons in our sample (Figure 4A). The average prediction correlation for STRFs estimated by boosting is r ¼ 0.27, slightly but significantly greater than the average for STRFs estimated by NRC, r ¼ 0.25 (p < 0.001, randomized paired t-test, Figure 4B). Within the sample of neurons, STRFs estimated by boosting can predict validation responses with greater than random accuracy for 87% of neurons (n ¼ 143/164, p < 0.05, (A) 0.8 Boosted STRF prediction correlation n= NRC STRF prediction correlation (B) Average speech prediction correlation NRC ** Boosted ** Boosted/NRC subspace (C) Average TORC prediction correlation NRC ** * Boosted boosted/nrc subspace Figure 4. (A) Boosted and NRC STRFs were evaluated by their ability to predict responses to a validation speech data set that was not used for STRF estimation. Each point plots the correlation coefficient between the predicted and observed response for the NRC STRF (horizontal axis) and boosted STRF (vertical axis) for a single neuron. Black dots indicate neurons (133/164) for which both STRFs predict with greater than random accuracy (jackknifed t-test, p < 0.05). (B) The average prediction correlation for boosted STRFs (mean 0.27) is significantly greater than the average for NRC STRFs (mean 0.24, p < 0.001, randomized paired t-test). After the boosted STRFs are projected into the stimulus subspace used for NRC, however, prediction accuracy (mean 0.24) is significantly reduced from the original boosted STRFs (p < 0.001, randomized paired t-test) and the difference from NRC STRFs disappears. (C) Average prediction correlation for boosted and NRC STRFs predicting responses to TORCs. The average prediction correlation for boosted STRFs (mean 0.084) is significantly greater than for NRC STRFs (mean 0.072, p < 0.001, randomized paired t-test). As in the case of speech predictions, the difference in prediction accuracy disappears after the boosted STRFs are projected into the NRC stimulus subspace. (*p < 0.05; **p < 0.001).

14 204 S. V. David et al. jackknifed t-test). Nearly the same number of STRFs estimated by NRC predict validation data with greater than random accuracy (82%, n ¼ 134/164). To evaluate how well STRFs estimated using speech generalize to other stimulus domains, we also compared how well they predicted responses to TORCs (Klein et al. 2000) played to the same set of neurons (Figure 4C). (TORCs are a convenient stimulus for estimating spectro-temporal response properties, but this study is restricted only to evaluating how well STRFs estimated using speech predict responses to TORCs.) As with the speech predictions, STRFs estimated by boosting tend to predict TORC responses (mean 0.084) better than STRFs estimated by NRC (mean 0.072, p < 0.001, randomized paired t-test). The absolute prediction accuracy for both sets of STRFs is substantially lower than their speech predictions. This decrease in performance suggests that neither STRF generalizes completely to other stimulus classes. Because of nonlinear response properties, STRFs estimated using one stimulus class tend to predict responses to other stimulus classes with worse accuracy (Theunissen et al. 2001; David et al. 2004). However, the superior performance of boosted STRFs suggests that they provide a more general characterization of spectro-temporal tuning across different stimulus conditions. Effects of estimation algorithm priors on STRFs The similar performance of the two sets of STRFs, despite their substantial differences in tuning, may seem paradoxical. It is important to consider, however, that the two estimation algorithms assume very different priors about the STRFs. When signal-to-noise levels are high, priors have relatively little effect, and the stimulus response data dominates the estimates. However, when signal-to-noise is low, priors provide an informed guess for how the STRF might appear, based on existing knowledge of the system under study. Boosting employs a sparse prior. On each iteration, a single STRF parameter is incremented by a small amount until further increments add noise. The optimal number of iterations varies across neurons. For this study, the number of iterations to complete for each neuron was determined by cross-validation (see Methods ). In the presence of noise, only a few coefficients will be nonzero; thus the STRF will be biased to be sparse. This process is illustrated for an example neuron in Figure 5A. After a few iterations, only a few parameters are non-zero. With successive iterations, more parameters are filled in and predictions improve. After the optimal number of iterations (circled in gray), additional increments add noise and predictions worsen. NRC uses a different prior, fitting the STRF only in the subspace of the spectrum that contains most of the variance in the speech stimulus. Parameters that describe responses outside this subspace are fixed at zero. A neuron may be tuned in the excluded space, but tuning in that space has relatively little impact on responses to speech, as speech stimuli, by definition, contain little power in that space. The size of the subspace preserved by NRC is specified by a tolerance value, which corresponds to the fraction of stimulus variance to preserve (Theunissen et al. 2001). The optimal tolerance value depends on the tuning and responsiveness of individual neurons. In this study, tolerance values ranged from 90%

15 Sparse receptive fields from natural stimuli 205 (A) (B) Iteration count Boosted STRF Tolerance NRC STRF Overregularized Regularization strength Noisy Figure 5. (A) Illustration of the sparse prior for boosting. The iterative boosting algorithm must be stopped early to avoid overfitting to noise. Panels from top to bottom show STRFs estimated by boosting after increasing numbers of iterations (STRFs have been plotted in their raw unsmoothed state). After a few iterations, only a single parameter is nonzero. Further iterations allow more nonzero parameters but also introduce noise. The optimal number of iterations (STRF circled in gray) is determined by cross validation. (B) Illustration of the smooth prior for NRC. NRC avoids overfitting to noise by projecting the fit stimulus into the subspace that contains most of its variance. The fraction of stimulus variance preserved after the projection is specified by a tolerance value. STRFs estimated using increasingly large tolerance values are shown in the panels from top to bottom. For very low tolerances, the STRF is extremely smooth, and the smaller inhibitory regions cannot be identified. Higher tolerance values reveal more detail, but at excessively high values the STRF is dominated by noise. The optimal tolerance value (STRF circled in gray) is chosen by cross validation. (most regularized) to % (least regularized), and the value for each neuron was chosen by cross-validation (see Methods ). Speech, like most other natural sounds, contains relatively broadband components and changes relatively slowly in time. Thus speech contains relatively little power at high spectral scales (i.e., modulation of stimulus power across frequency channels) and little power at high rates (i.e., modulation of stimulus power over time). The stimulus domain typically excluded by NRC contains these high rates and high scales, and the effect of a low tolerance value is to smooth spectro-temporal

16 206 S. V. David et al. (A) (B) Frequency (Hz) Frequency (octaves) BF 1 2 Boosted STRF Time lag (ms) Time from peak latency (ms) Averaged STRF, n=133 neurons NRC STRF Time lag (ms) Time from peak latency (ms) Boosted STRF - NRC subspace projection Time lag (ms) Time from peak latency (ms) Figure 6. (A) The boosted STRF estimated for one neuron (left) shares basic spectral tuning with the NRC STRF estimated for the same neuron (center) but is more narrowly tuned in time. When the boosted STRF is projected into the subspace used by NRC (right), the tuning is nearly identical to the NRC STRF. (B) The average boosted (left) and NRC (center) STRF was calculated after shifting each STRF to have the same best frequency and peak latency and normalizing each to the same gain. The average boosted STRF has a higher spectral scale (i.e., narrower bandwidth) and a higher modulation rate (i.e., narrower latency range). When the average is computed for boosted STRFs after projecting each into the NRC subspace used for that neuron (right), it is nearly identical to the average NRC STRF. tuning. This is illustrated for an example neuron in Figure 5B. At very low tolerance values, the STRF is highly smoothed. Reducing the tolerance tightens up tuning, but for very high tolerance values, noise begins to dominate the STRF. Because of their different priors, the two algorithms can arrive at STRF estimates that appear very different (Figure 6A). In order to understand the source of these differences, it is useful to study the specific effects of the different priors in more detail. The relatively broadband STRF produced by NRC reflects the fact that the NRC algorithm has assumed that low-power stimulus dimensions have no effect on responses. In practice, the NRC STRF is estimated only in the high-power subspace of the stimulus, while the boosted STRF is estimated in the complete stimulus space. For comparison, the boosted STRF can be projected into the same subspace as the NRC STRF (Equation 14). When the boosted STRF is projected into this subspace, it appears much more similar to the NRC STRF (Figure 6A, right). We can make the same comparison across the entire set of neurons in our sample. Figure 6B shows the average STRFs estimated using each algorithm for the 133/164 neurons whose STRFs predicted with greater than random accuracy (Figure 4A) Normalized gain

17 Sparse receptive fields from natural stimuli 207 Before averaging, each STRF was shifted to have the same best frequency and peak latency and normalized to have the same overall gain. The average boosted STRF (left) clearly has narrower spectral and temporal tuning than the average NRC STRF (middle). However, when the boosted STRFs are projected into the NRC subspace, the average STRF (right) appears very similar to the average NRC STRF. To determine the effect of projecting boosted STRFs into the NRC subspace on model accuracy, we evaluated the ability of the projected STRFs to predict the same validation data used to test the boosted- and NRC STRFs. The boosted STRFs projected into the NRC subspace predict with a mean correlation of r ¼ 0.24, nearly the same accuracy as NRC STRFs and significantly lower than the mean correlation for the original boosted STRFs (randomized paired t-test, p < 0.001, Figure 4B). Thus the small enhancement in predictive power of the boosted STRFs can be attributed tuning in the subspace that is excluded by NRC. Comparison of tuning properties of STRFs estimated by boosting and NRC Our comparison of predictive power indicates that the superior performance of boosted STRFs lies in spectro-temporal tuning properties in the stimulus subspace excluded by NRC. To test for these differences directly, we measured a similarity index (the normalized correlation) between STRFs estimated for the same neuron using each of the methods. A similarity index of 1 indicates a perfect match, and an index of 0 indicates no correlation between STRFs. We restricted this analysis only to the 133/164 neurons for which boosted and NRC STRFs predicted with greater than random accuracy. When measured directly between STRFs estimated using boosting and NRC, the similarity index is relatively low (mean 0.44, Figure 7A). Across the set of neurons, 111/133 (83%) have similarity indices significantly less than 1 (p < 0.05, jackknifed t-test). However, after projecting the boosted STRFs into the NRC subspace, similarity greatly increases (mean 0.78, Figure 7B), and only 41/133 (31%) of neurons have index values significantly less than 1 (p < 0.05, jackknifed t-test). Thus the differences in tuning observed between boosted and NRC STRFs can be attributed largely to the stimulus subspace used by NRC. The differences that persist are likely to be small effects of the sparse prior used by boosting that do not significantly affect prediction accuracy. In order to identify specific differences in spectro-temporal tuning that give rise to the improved performance of the boosted STRFs, we compared the tuning properties of STRFs estimated using the two different algorithms. For each STRF we measured the best frequency, peak latency of excitation, spectral scale (i.e., the inverse of spectral bandwidth), and preferred temporal modulation rate. We first compared the tuning properties of boosted STRFs directly to those of NRC STRFs. We then projected the STRF of each boosted STRF into the subspace used by NRC for that neuron and compared the tuning of the projected STRFs to NRC STRFs. The scatter plots in Figures 8A and B, respectively, illustrate the comparison of spectral scale before and after projection. We observed only small effects of the estimation algorithm on best frequency and peak latency. The mean best frequency for boosted STRFs (2022 Hz) is slightly but not significantly greater than the mean for NRC STRFs (1986 Hz, randomized paired t-test, Figure 8C). After projecting the boosted STRFs into the NRC

18 208 S. V. David et al. (A) Number of cells Similarity < 1 (p<0.05) N.S Similarity index (Boosted vs. NRC STRFs) (B) 60 Number of cells n= Similarity index (Boosted/NRC subspace vs. NRC STRFs) Figure 7. (A) Histogram of similarity indices computed between boosted and NRC STRFs. An index of 1 indicates an exact match between STRFs and a value of 0 indicates no correlation. The mean similarity index is White bars indicate neurons for which similarity is significantly less than 1 (p < 0.05, jackknifed t-test). (B) Histogram of similarity indices computed between boosted and NRC STRFs after projecting the boosted STRFs into the stimulus subspace used by NRC. After projection, similarity increases significantly to a mean of 0.78 across the set of neurons (p < 0.001, randomized paired t-test). subspace (mean 1947 Hz), the difference becomes even smaller. Mean peak latency is also not significantly different between boosted STRFs (34.3 ms) and NRC STRFs (39.0 ms, Figure 8D). The mean peak latency of boosted STRFs projected into the NRC subspace (34.5 ms) is unchanged from before the projection. The differences in spectral scale and preferred rate determined by the estimation algorithms are more dramatic. The mean spectral scale of boosted STRFs (0.90 cyc/oct) is significantly greater than the mean for NRC STRFs (0.63 cyc/oct, p < 0.001, randomized paired t-test, Figure 8E). After projecting the boosted STRFs into the NRC subspace (mean 0.60 cyc/oct), the difference becomes much smaller and is no longer significant (p > 0.2). The mean preferred rate of boosted STRFs (22.4 Hz) is nearly twice the mean for NRC STRFs (13.4 Hz, p < 0.001, randomized paired t-test, Figure 8F). After projecting the

19 Sparse receptive fields from natural stimuli 209 (A) (B) Boosted STRF scale (cyc/oct) Boosted/NRC space scale (cyc/oct) NRC STRF scale (cyc/oct) NRC STRF scale (cyc/oct) (C) Average BF (Hz) (E) Average scale (cyc/oct) ** n=133 * p<0.05 ** p<0.001 Average peak latency (ms) Boosted STRFs NRC STRFs (D) 50 Average rate (Hz) 25 (F) ** Boosted STRF/NRC subspace Figure 8. (A) Spectral scale measured from each NRC STRF (horizontal axis) is plotted against the spectral scale measured from the boosted STRF for the same neuron (vertical axis). Neurons for which both STRFs predict with greater than random accuracy (p < 0.05, jackknifed t-test) are plotted in black. The majority of points lie above the line of unity slope, indicating that scale is generally higher in boosted STRFs. (B) Comparison of spectral scale between NRC and boosted STRFs after projection into the subspace used to estimate the NRC STRFs. The majority of points lie near the line of unity slope, indicating that scale is similar in the two sets of STRFs. (C) Comparison of average best frequency for boosted STRFs (white), NRC STRFs (black) and boosted STRFs after projection into the NRC subspace (gray). There is no significant difference between the three groups of STRFs. (D) Comparison of average peak latency, plotted as in C. (E) Comparison of average spectral scale. Boosted STRFs have significantly greater scale than NRC STRFs (p < 0.001, randomized paired t-test). This difference disappears after the boosted STRFs are projected into the NRC subspace. (F) Comparison of average preferred rate. Boosted STRFs are tuned to significantly greater rates than NRC STRFs (p < 0.001, randomized paired t-test). This difference disappears after the boosted STRFs are projected into the NRC subspace, and the average rate is actually slightly lower than the NRC STRFs (p < 0.05). *

Spectro-temporal response fields in the inferior colliculus of awake monkey

Spectro-temporal response fields in the inferior colliculus of awake monkey 3.6.QH Spectro-temporal response fields in the inferior colliculus of awake monkey Versnel, Huib; Zwiers, Marcel; Van Opstal, John Department of Biophysics University of Nijmegen Geert Grooteplein 655

More information

Analysis of in-vivo extracellular recordings. Ryan Morrill Bootcamp 9/10/2014

Analysis of in-vivo extracellular recordings. Ryan Morrill Bootcamp 9/10/2014 Analysis of in-vivo extracellular recordings Ryan Morrill Bootcamp 9/10/2014 Goals for the lecture Be able to: Conceptually understand some of the analysis and jargon encountered in a typical (sensory)

More information

USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES

USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES Varinthira Duangudom and David V Anderson School of Electrical and Computer Engineering, Georgia Institute of Technology Atlanta, GA 30332

More information

Title A generalized linear model for estimating spectrotemporal receptive fields from responses to natural sounds.

Title A generalized linear model for estimating spectrotemporal receptive fields from responses to natural sounds. 1 Title A generalized linear model for estimating spectrotemporal receptive fields from responses to natural sounds. Authors Ana Calabrese 1,2, Joseph W. Schumacher 1, David M. Schneider 1, Liam Paninski

More information

Spectrograms (revisited)

Spectrograms (revisited) Spectrograms (revisited) We begin the lecture by reviewing the units of spectrograms, which I had only glossed over when I covered spectrograms at the end of lecture 19. We then relate the blocks of a

More information

Representation of sound in the auditory nerve

Representation of sound in the auditory nerve Representation of sound in the auditory nerve Eric D. Young Department of Biomedical Engineering Johns Hopkins University Young, ED. Neural representation of spectral and temporal information in speech.

More information

How is the stimulus represented in the nervous system?

How is the stimulus represented in the nervous system? How is the stimulus represented in the nervous system? Eric Young F Rieke et al Spikes MIT Press (1997) Especially chapter 2 I Nelken et al Encoding stimulus information by spike numbers and mean response

More information

HCS 7367 Speech Perception

HCS 7367 Speech Perception Long-term spectrum of speech HCS 7367 Speech Perception Connected speech Absolute threshold Males Dr. Peter Assmann Fall 212 Females Long-term spectrum of speech Vowels Males Females 2) Absolute threshold

More information

AUDL GS08/GAV1 Signals, systems, acoustics and the ear. Pitch & Binaural listening

AUDL GS08/GAV1 Signals, systems, acoustics and the ear. Pitch & Binaural listening AUDL GS08/GAV1 Signals, systems, acoustics and the ear Pitch & Binaural listening Review 25 20 15 10 5 0-5 100 1000 10000 25 20 15 10 5 0-5 100 1000 10000 Part I: Auditory frequency selectivity Tuning

More information

Stimulus-Dependent Auditory Tuning Results in Synchronous Population Coding of Vocalizations in the Songbird Midbrain

Stimulus-Dependent Auditory Tuning Results in Synchronous Population Coding of Vocalizations in the Songbird Midbrain The Journal of Neuroscience, March 1, 2006 26(9):2499 2512 2499 Behavioral/Systems/Cognitive Stimulus-Dependent Auditory Tuning Results in Synchronous Population Coding of Vocalizations in the Songbird

More information

Robust Neural Encoding of Speech in Human Auditory Cortex

Robust Neural Encoding of Speech in Human Auditory Cortex Robust Neural Encoding of Speech in Human Auditory Cortex Nai Ding, Jonathan Z. Simon Electrical Engineering / Biology University of Maryland, College Park Auditory Processing in Natural Scenes How is

More information

Behavioral generalization

Behavioral generalization Supplementary Figure 1 Behavioral generalization. a. Behavioral generalization curves in four Individual sessions. Shown is the conditioned response (CR, mean ± SEM), as a function of absolute (main) or

More information

Spectral-Temporal Receptive Fields of Nonlinear Auditory Neurons Obtained Using Natural Sounds

Spectral-Temporal Receptive Fields of Nonlinear Auditory Neurons Obtained Using Natural Sounds The Journal of Neuroscience, March 15, 2000, 20(6):2315 2331 Spectral-Temporal Receptive Fields of Nonlinear Auditory Neurons Obtained Using Natural Sounds Frédéric E. Theunissen, 1 Kamal Sen, 4 and Allison

More information

What you re in for. Who are cochlear implants for? The bottom line. Speech processing schemes for

What you re in for. Who are cochlear implants for? The bottom line. Speech processing schemes for What you re in for Speech processing schemes for cochlear implants Stuart Rosen Professor of Speech and Hearing Science Speech, Hearing and Phonetic Sciences Division of Psychology & Language Sciences

More information

Analysis of spectro-temporal receptive fields in an auditory neural network

Analysis of spectro-temporal receptive fields in an auditory neural network Analysis of spectro-temporal receptive fields in an auditory neural network Madhav Nandipati Abstract Neural networks have been utilized for a vast range of applications, including computational biology.

More information

Nature Neuroscience: doi: /nn Supplementary Figure 1. Trial structure for go/no-go behavior

Nature Neuroscience: doi: /nn Supplementary Figure 1. Trial structure for go/no-go behavior Supplementary Figure 1 Trial structure for go/no-go behavior a, Overall timeline of experiments. Day 1: A1 mapping, injection of AAV1-SYN-GCAMP6s, cranial window and headpost implantation. Water restriction

More information

The development of a modified spectral ripple test

The development of a modified spectral ripple test The development of a modified spectral ripple test Justin M. Aronoff a) and David M. Landsberger Communication and Neuroscience Division, House Research Institute, 2100 West 3rd Street, Los Angeles, California

More information

Dynamics of Precise Spike Timing in Primary Auditory Cortex

Dynamics of Precise Spike Timing in Primary Auditory Cortex The Journal of Neuroscience, February 4, 2004 24(5):1159 1172 1159 Behavioral/Systems/Cognitive Dynamics of Precise Spike Timing in Primary Auditory Cortex Mounya Elhilali, Jonathan B. Fritz, David J.

More information

Task Difficulty and Performance Induce Diverse Adaptive Patterns in Gain and Shape of Primary Auditory Cortical Receptive Fields

Task Difficulty and Performance Induce Diverse Adaptive Patterns in Gain and Shape of Primary Auditory Cortical Receptive Fields Article Task Difficulty and Performance Induce Diverse Adaptive Patterns in Gain and Shape of Primary Auditory Cortical Receptive Fields Serin Atiani, 1 Mounya Elhilali, 3 Stephen V. David, 2 Jonathan

More information

EEL 6586, Project - Hearing Aids algorithms

EEL 6586, Project - Hearing Aids algorithms EEL 6586, Project - Hearing Aids algorithms 1 Yan Yang, Jiang Lu, and Ming Xue I. PROBLEM STATEMENT We studied hearing loss algorithms in this project. As the conductive hearing loss is due to sound conducting

More information

Extra-Classical Tuning Predicts Stimulus-Dependent Receptive Fields in Auditory Neurons

Extra-Classical Tuning Predicts Stimulus-Dependent Receptive Fields in Auditory Neurons The Journal of Neuroscience, August 17, 2011 31(33):11867 11878 11867 Behavioral/Systems/Cognitive Extra-Classical Tuning Predicts Stimulus-Dependent Receptive Fields in Auditory Neurons David M. Schneider

More information

Linguistic Phonetics. Basic Audition. Diagram of the inner ear removed due to copyright restrictions.

Linguistic Phonetics. Basic Audition. Diagram of the inner ear removed due to copyright restrictions. 24.963 Linguistic Phonetics Basic Audition Diagram of the inner ear removed due to copyright restrictions. 1 Reading: Keating 1985 24.963 also read Flemming 2001 Assignment 1 - basic acoustics. Due 9/22.

More information

Frequency Tracking: LMS and RLS Applied to Speech Formant Estimation

Frequency Tracking: LMS and RLS Applied to Speech Formant Estimation Aldebaro Klautau - http://speech.ucsd.edu/aldebaro - 2/3/. Page. Frequency Tracking: LMS and RLS Applied to Speech Formant Estimation ) Introduction Several speech processing algorithms assume the signal

More information

Pattern Playback in the '90s

Pattern Playback in the '90s Pattern Playback in the '90s Malcolm Slaney Interval Research Corporation 180 l-c Page Mill Road, Palo Alto, CA 94304 malcolm@interval.com Abstract Deciding the appropriate representation to use for modeling

More information

Responses of Neurons in Cat Primary Auditory Cortex to Bird Chirps: Effects of Temporal and Spectral Context

Responses of Neurons in Cat Primary Auditory Cortex to Bird Chirps: Effects of Temporal and Spectral Context The Journal of Neuroscience, October 1, 2002, 22(19):8619 8632 Responses of Neurons in Cat Primary Auditory Cortex to Bird Chirps: Effects of Temporal and Spectral Context Omer Bar-Yosef, 1 Yaron Rotman,

More information

Information Processing During Transient Responses in the Crayfish Visual System

Information Processing During Transient Responses in the Crayfish Visual System Information Processing During Transient Responses in the Crayfish Visual System Christopher J. Rozell, Don. H. Johnson and Raymon M. Glantz Department of Electrical & Computer Engineering Department of

More information

LATERAL INHIBITION MECHANISM IN COMPUTATIONAL AUDITORY MODEL AND IT'S APPLICATION IN ROBUST SPEECH RECOGNITION

LATERAL INHIBITION MECHANISM IN COMPUTATIONAL AUDITORY MODEL AND IT'S APPLICATION IN ROBUST SPEECH RECOGNITION LATERAL INHIBITION MECHANISM IN COMPUTATIONAL AUDITORY MODEL AND IT'S APPLICATION IN ROBUST SPEECH RECOGNITION Lu Xugang Li Gang Wang Lip0 Nanyang Technological University, School of EEE, Workstation Resource

More information

Complete Functional Characterization of Sensory Neurons by System Identification

Complete Functional Characterization of Sensory Neurons by System Identification Annu. Rev. Neurosci. 2006. 29:477 505 The Annual Review of Neuroscience is online at neuro.annualreviews.org doi: 10.1146/ annurev.neuro.29.051605.113024 Copyright c 2006 by Annual Reviews. All rights

More information

Who are cochlear implants for?

Who are cochlear implants for? Who are cochlear implants for? People with little or no hearing and little conductive component to the loss who receive little or no benefit from a hearing aid. Implants seem to work best in adults who

More information

Essential feature. Who are cochlear implants for? People with little or no hearing. substitute for faulty or missing inner hair

Essential feature. Who are cochlear implants for? People with little or no hearing. substitute for faulty or missing inner hair Who are cochlear implants for? Essential feature People with little or no hearing and little conductive component to the loss who receive little or no benefit from a hearing aid. Implants seem to work

More information

SUPPLEMENTARY INFORMATION. Table 1 Patient characteristics Preoperative. language testing

SUPPLEMENTARY INFORMATION. Table 1 Patient characteristics Preoperative. language testing Categorical Speech Representation in the Human Superior Temporal Gyrus Edward F. Chang, Jochem W. Rieger, Keith D. Johnson, Mitchel S. Berger, Nicholas M. Barbaro, Robert T. Knight SUPPLEMENTARY INFORMATION

More information

HEARING AND PSYCHOACOUSTICS

HEARING AND PSYCHOACOUSTICS CHAPTER 2 HEARING AND PSYCHOACOUSTICS WITH LIDIA LEE I would like to lead off the specific audio discussions with a description of the audio receptor the ear. I believe it is always a good idea to understand

More information

FREQUENCY COMPRESSION AND FREQUENCY SHIFTING FOR THE HEARING IMPAIRED

FREQUENCY COMPRESSION AND FREQUENCY SHIFTING FOR THE HEARING IMPAIRED FREQUENCY COMPRESSION AND FREQUENCY SHIFTING FOR THE HEARING IMPAIRED Francisco J. Fraga, Alan M. Marotta National Institute of Telecommunications, Santa Rita do Sapucaí - MG, Brazil Abstract A considerable

More information

Hearing Lectures. Acoustics of Speech and Hearing. Auditory Lighthouse. Facts about Timbre. Analysis of Complex Sounds

Hearing Lectures. Acoustics of Speech and Hearing. Auditory Lighthouse. Facts about Timbre. Analysis of Complex Sounds Hearing Lectures Acoustics of Speech and Hearing Week 2-10 Hearing 3: Auditory Filtering 1. Loudness of sinusoids mainly (see Web tutorial for more) 2. Pitch of sinusoids mainly (see Web tutorial for more)

More information

G5)H/C8-)72)78)2I-,8/52& ()*+,-./,-0))12-345)6/3/782 9:-8;<;4.= J-3/ J-3/ "#&' "#% "#"% "#%$

G5)H/C8-)72)78)2I-,8/52& ()*+,-./,-0))12-345)6/3/782 9:-8;<;4.= J-3/ J-3/ #&' #% #% #%$ # G5)H/C8-)72)78)2I-,8/52& #% #$ # # &# G5)H/C8-)72)78)2I-,8/52' @5/AB/7CD J-3/ /,?8-6/2@5/AB/7CD #&' #% #$ # # '#E ()*+,-./,-0))12-345)6/3/782 9:-8;;4. @5/AB/7CD J-3/ #' /,?8-6/2@5/AB/7CD #&F #&' #% #$

More information

Comment by Delgutte and Anna. A. Dreyer (Eaton-Peabody Laboratory, Massachusetts Eye and Ear Infirmary, Boston, MA)

Comment by Delgutte and Anna. A. Dreyer (Eaton-Peabody Laboratory, Massachusetts Eye and Ear Infirmary, Boston, MA) Comments Comment by Delgutte and Anna. A. Dreyer (Eaton-Peabody Laboratory, Massachusetts Eye and Ear Infirmary, Boston, MA) Is phase locking to transposed stimuli as good as phase locking to low-frequency

More information

Linguistic Phonetics Fall 2005

Linguistic Phonetics Fall 2005 MIT OpenCourseWare http://ocw.mit.edu 24.963 Linguistic Phonetics Fall 2005 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. 24.963 Linguistic Phonetics

More information

Coincidence Detection in Pitch Perception. S. Shamma, D. Klein and D. Depireux

Coincidence Detection in Pitch Perception. S. Shamma, D. Klein and D. Depireux Coincidence Detection in Pitch Perception S. Shamma, D. Klein and D. Depireux Electrical and Computer Engineering Department, Institute for Systems Research University of Maryland at College Park, College

More information

Chapter 5. Summary and Conclusions! 131

Chapter 5. Summary and Conclusions! 131 ! Chapter 5 Summary and Conclusions! 131 Chapter 5!!!! Summary of the main findings The present thesis investigated the sensory representation of natural sounds in the human auditory cortex. Specifically,

More information

Supplementary Information for Correlated input reveals coexisting coding schemes in a sensory cortex

Supplementary Information for Correlated input reveals coexisting coding schemes in a sensory cortex Supplementary Information for Correlated input reveals coexisting coding schemes in a sensory cortex Luc Estebanez 1,2 *, Sami El Boustani 1 *, Alain Destexhe 1, Daniel E. Shulz 1 1 Unité de Neurosciences,

More information

Chapter 40 Effects of Peripheral Tuning on the Auditory Nerve s Representation of Speech Envelope and Temporal Fine Structure Cues

Chapter 40 Effects of Peripheral Tuning on the Auditory Nerve s Representation of Speech Envelope and Temporal Fine Structure Cues Chapter 40 Effects of Peripheral Tuning on the Auditory Nerve s Representation of Speech Envelope and Temporal Fine Structure Cues Rasha A. Ibrahim and Ian C. Bruce Abstract A number of studies have explored

More information

Essential feature. Who are cochlear implants for? People with little or no hearing. substitute for faulty or missing inner hair

Essential feature. Who are cochlear implants for? People with little or no hearing. substitute for faulty or missing inner hair Who are cochlear implants for? Essential feature People with little or no hearing and little conductive component to the loss who receive little or no benefit from a hearing aid. Implants seem to work

More information

Computational Perception /785. Auditory Scene Analysis

Computational Perception /785. Auditory Scene Analysis Computational Perception 15-485/785 Auditory Scene Analysis A framework for auditory scene analysis Auditory scene analysis involves low and high level cues Low level acoustic cues are often result in

More information

Neural correlates of the perception of sound source separation

Neural correlates of the perception of sound source separation Neural correlates of the perception of sound source separation Mitchell L. Day 1,2 * and Bertrand Delgutte 1,2,3 1 Department of Otology and Laryngology, Harvard Medical School, Boston, MA 02115, USA.

More information

Acoustics, signals & systems for audiology. Psychoacoustics of hearing impairment

Acoustics, signals & systems for audiology. Psychoacoustics of hearing impairment Acoustics, signals & systems for audiology Psychoacoustics of hearing impairment Three main types of hearing impairment Conductive Sound is not properly transmitted from the outer to the inner ear Sensorineural

More information

Information-theoretic stimulus design for neurophysiology & psychophysics

Information-theoretic stimulus design for neurophysiology & psychophysics Information-theoretic stimulus design for neurophysiology & psychophysics Christopher DiMattina, PhD Assistant Professor of Psychology Florida Gulf Coast University 2 Optimal experimental design Part 1

More information

Binaural Hearing. Steve Colburn Boston University

Binaural Hearing. Steve Colburn Boston University Binaural Hearing Steve Colburn Boston University Outline Why do we (and many other animals) have two ears? What are the major advantages? What is the observed behavior? How do we accomplish this physiologically?

More information

The Effect of Analysis Methods and Input Signal Characteristics on Hearing Aid Measurements

The Effect of Analysis Methods and Input Signal Characteristics on Hearing Aid Measurements The Effect of Analysis Methods and Input Signal Characteristics on Hearing Aid Measurements By: Kristina Frye Section 1: Common Source Types FONIX analyzers contain two main signal types: Puretone and

More information

Sound localization psychophysics

Sound localization psychophysics Sound localization psychophysics Eric Young A good reference: B.C.J. Moore An Introduction to the Psychology of Hearing Chapter 7, Space Perception. Elsevier, Amsterdam, pp. 233-267 (2004). Sound localization:

More information

Supplementary Figure 1. Recording sites.

Supplementary Figure 1. Recording sites. Supplementary Figure 1 Recording sites. (a, b) Schematic of recording locations for mice used in the variable-reward task (a, n = 5) and the variable-expectation task (b, n = 5). RN, red nucleus. SNc,

More information

PCA Enhanced Kalman Filter for ECG Denoising

PCA Enhanced Kalman Filter for ECG Denoising IOSR Journal of Electronics & Communication Engineering (IOSR-JECE) ISSN(e) : 2278-1684 ISSN(p) : 2320-334X, PP 06-13 www.iosrjournals.org PCA Enhanced Kalman Filter for ECG Denoising Febina Ikbal 1, Prof.M.Mathurakani

More information

arxiv:submit/ [q-bio.nc] 15 Oct 2017

arxiv:submit/ [q-bio.nc] 15 Oct 2017 Learning Mid-Level Auditory Codes from Natural Sound Statistics Wiktor Młynarski*, Josh H. McDermott mlynar@mit.edu arxiv:submit/2374 [q-bio.nc] 5 Oct 27 Abstract Interaction with the world requires an

More information

Over-representation of speech in older adults originates from early response in higher order auditory cortex

Over-representation of speech in older adults originates from early response in higher order auditory cortex Over-representation of speech in older adults originates from early response in higher order auditory cortex Christian Brodbeck, Alessandro Presacco, Samira Anderson & Jonathan Z. Simon Overview 2 Puzzle

More information

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 THE DUPLEX-THEORY OF LOCALIZATION INVESTIGATED UNDER NATURAL CONDITIONS

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 THE DUPLEX-THEORY OF LOCALIZATION INVESTIGATED UNDER NATURAL CONDITIONS 19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 27 THE DUPLEX-THEORY OF LOCALIZATION INVESTIGATED UNDER NATURAL CONDITIONS PACS: 43.66.Pn Seeber, Bernhard U. Auditory Perception Lab, Dept.

More information

EFFECTS OF TEMPORAL FINE STRUCTURE ON THE LOCALIZATION OF BROADBAND SOUNDS: POTENTIAL IMPLICATIONS FOR THE DESIGN OF SPATIAL AUDIO DISPLAYS

EFFECTS OF TEMPORAL FINE STRUCTURE ON THE LOCALIZATION OF BROADBAND SOUNDS: POTENTIAL IMPLICATIONS FOR THE DESIGN OF SPATIAL AUDIO DISPLAYS Proceedings of the 14 International Conference on Auditory Display, Paris, France June 24-27, 28 EFFECTS OF TEMPORAL FINE STRUCTURE ON THE LOCALIZATION OF BROADBAND SOUNDS: POTENTIAL IMPLICATIONS FOR THE

More information

Juan Carlos Tejero-Calado 1, Janet C. Rutledge 2, and Peggy B. Nelson 3

Juan Carlos Tejero-Calado 1, Janet C. Rutledge 2, and Peggy B. Nelson 3 PRESERVING SPECTRAL CONTRAST IN AMPLITUDE COMPRESSION FOR HEARING AIDS Juan Carlos Tejero-Calado 1, Janet C. Rutledge 2, and Peggy B. Nelson 3 1 University of Malaga, Campus de Teatinos-Complejo Tecnol

More information

J Jeffress model, 3, 66ff

J Jeffress model, 3, 66ff Index A Absolute pitch, 102 Afferent projections, inferior colliculus, 131 132 Amplitude modulation, coincidence detector, 152ff inferior colliculus, 152ff inhibition models, 156ff models, 152ff Anatomy,

More information

Acoustics Research Institute

Acoustics Research Institute Austrian Academy of Sciences Acoustics Research Institute Modeling Modelingof ofauditory AuditoryPerception Perception Bernhard BernhardLaback Labackand andpiotr PiotrMajdak Majdak http://www.kfs.oeaw.ac.at

More information

THRESHOLD PREDICTION USING THE ASSR AND THE TONE BURST CONFIGURATIONS

THRESHOLD PREDICTION USING THE ASSR AND THE TONE BURST CONFIGURATIONS THRESHOLD PREDICTION USING THE ASSR AND THE TONE BURST ABR IN DIFFERENT AUDIOMETRIC CONFIGURATIONS INTRODUCTION INTRODUCTION Evoked potential testing is critical in the determination of audiologic thresholds

More information

Neural Encoding. Naureen Ghani. February 10, 2018

Neural Encoding. Naureen Ghani. February 10, 2018 Neural Encoding Naureen Ghani February 10, 2018 Introduction Neuroscientists are interested in knowing what neurons are doing. More specifically, researchers want to understand how neurons represent stimuli

More information

Supporting Information

Supporting Information Supporting Information ten Oever and Sack 10.1073/pnas.1517519112 SI Materials and Methods Experiment 1. Participants. A total of 20 participants (9 male; age range 18 32 y; mean age 25 y) participated

More information

Theme 2: Cellular mechanisms in the Cochlear Nucleus

Theme 2: Cellular mechanisms in the Cochlear Nucleus Theme 2: Cellular mechanisms in the Cochlear Nucleus The Cochlear Nucleus (CN) presents a unique opportunity for quantitatively studying input-output transformations by neurons because it gives rise to

More information

Adaptive Changes in Cortical Receptive Fields Induced by Attention to Complex Sounds

Adaptive Changes in Cortical Receptive Fields Induced by Attention to Complex Sounds J Neurophysiol 9: 337 36, 7. First published August 5, 7; doi:/jn.55.7. Adaptive Changes in Cortical Receptive Fields Induced by Attention to Complex Sounds Jonathan B. Fritz, Mounya Elhilali, and Shihab

More information

CS/NEUR125 Brains, Minds, and Machines. Due: Friday, April 14

CS/NEUR125 Brains, Minds, and Machines. Due: Friday, April 14 CS/NEUR125 Brains, Minds, and Machines Assignment 5: Neural mechanisms of object-based attention Due: Friday, April 14 This Assignment is a guided reading of the 2014 paper, Neural Mechanisms of Object-Based

More information

Implant Subjects. Jill M. Desmond. Department of Electrical and Computer Engineering Duke University. Approved: Leslie M. Collins, Supervisor

Implant Subjects. Jill M. Desmond. Department of Electrical and Computer Engineering Duke University. Approved: Leslie M. Collins, Supervisor Using Forward Masking Patterns to Predict Imperceptible Information in Speech for Cochlear Implant Subjects by Jill M. Desmond Department of Electrical and Computer Engineering Duke University Date: Approved:

More information

Modulation and Top-Down Processing in Audition

Modulation and Top-Down Processing in Audition Modulation and Top-Down Processing in Audition Malcolm Slaney 1,2 and Greg Sell 2 1 Yahoo! Research 2 Stanford CCRMA Outline The Non-Linear Cochlea Correlogram Pitch Modulation and Demodulation Information

More information

Can components in distortion-product otoacoustic emissions be separated?

Can components in distortion-product otoacoustic emissions be separated? Can components in distortion-product otoacoustic emissions be separated? Anders Tornvig Section of Acoustics, Aalborg University, Fredrik Bajers Vej 7 B5, DK-922 Aalborg Ø, Denmark, tornvig@es.aau.dk David

More information

Sum of Neurally Distinct Stimulus- and Task-Related Components.

Sum of Neurally Distinct Stimulus- and Task-Related Components. SUPPLEMENTARY MATERIAL for Cardoso et al. 22 The Neuroimaging Signal is a Linear Sum of Neurally Distinct Stimulus- and Task-Related Components. : Appendix: Homogeneous Linear ( Null ) and Modified Linear

More information

Brad May, PhD Johns Hopkins University

Brad May, PhD Johns Hopkins University Brad May, PhD Johns Hopkins University When the ear cannot function normally, the brain changes. Brain deafness contributes to poor speech comprehension, problems listening in noise, abnormal loudness

More information

Neurobiology of Hearing (Salamanca, 2012) Auditory Cortex (2) Prof. Xiaoqin Wang

Neurobiology of Hearing (Salamanca, 2012) Auditory Cortex (2) Prof. Xiaoqin Wang Neurobiology of Hearing (Salamanca, 2012) Auditory Cortex (2) Prof. Xiaoqin Wang Laboratory of Auditory Neurophysiology Department of Biomedical Engineering Johns Hopkins University web1.johnshopkins.edu/xwang

More information

Thresholds for different mammals

Thresholds for different mammals Loudness Thresholds for different mammals 8 7 What s the first thing you d want to know? threshold (db SPL) 6 5 4 3 2 1 hum an poodle m ouse threshold Note bowl shape -1 1 1 1 1 frequency (Hz) Sivian &

More information

Error Detection based on neural signals

Error Detection based on neural signals Error Detection based on neural signals Nir Even- Chen and Igor Berman, Electrical Engineering, Stanford Introduction Brain computer interface (BCI) is a direct communication pathway between the brain

More information

The role of low frequency components in median plane localization

The role of low frequency components in median plane localization Acoust. Sci. & Tech. 24, 2 (23) PAPER The role of low components in median plane localization Masayuki Morimoto 1;, Motoki Yairi 1, Kazuhiro Iida 2 and Motokuni Itoh 1 1 Environmental Acoustics Laboratory,

More information

IEEE SIGNAL PROCESSING LETTERS, VOL. 13, NO. 3, MARCH A Self-Structured Adaptive Decision Feedback Equalizer

IEEE SIGNAL PROCESSING LETTERS, VOL. 13, NO. 3, MARCH A Self-Structured Adaptive Decision Feedback Equalizer SIGNAL PROCESSING LETTERS, VOL 13, NO 3, MARCH 2006 1 A Self-Structured Adaptive Decision Feedback Equalizer Yu Gong and Colin F N Cowan, Senior Member, Abstract In a decision feedback equalizer (DFE),

More information

Signals, systems, acoustics and the ear. Week 5. The peripheral auditory system: The ear as a signal processor

Signals, systems, acoustics and the ear. Week 5. The peripheral auditory system: The ear as a signal processor Signals, systems, acoustics and the ear Week 5 The peripheral auditory system: The ear as a signal processor Think of this set of organs 2 as a collection of systems, transforming sounds to be sent to

More information

HST-582J/6.555J/16.456J-Biomedical Signal and Image Processing-Spring Laboratory Project 1 The Electrocardiogram

HST-582J/6.555J/16.456J-Biomedical Signal and Image Processing-Spring Laboratory Project 1 The Electrocardiogram HST-582J/6.555J/16.456J-Biomedical Signal and Image Processing-Spring 2007 DUE: 3/8/07 Laboratory Project 1 The Electrocardiogram 1 Introduction The electrocardiogram (ECG) is a recording of body surface

More information

Characterizing Neurons in the Primary Auditory Cortex of the Awake Primate U sing Reverse Correlation

Characterizing Neurons in the Primary Auditory Cortex of the Awake Primate U sing Reverse Correlation Characterizing Neurons in the Primary Auditory Cortex of the Awake Primate U sing Reverse Correlation R. Christopher dec harms decharms@phy.ucsf.edu Michael M. Merzenich merz@phy.ucsf.edu w. M. Keck Center

More information

OPTO 5320 VISION SCIENCE I

OPTO 5320 VISION SCIENCE I OPTO 5320 VISION SCIENCE I Monocular Sensory Processes of Vision: Color Vision Mechanisms of Color Processing . Neural Mechanisms of Color Processing A. Parallel processing - M- & P- pathways B. Second

More information

Variation in spectral-shape discrimination weighting functions at different stimulus levels and signal strengths

Variation in spectral-shape discrimination weighting functions at different stimulus levels and signal strengths Variation in spectral-shape discrimination weighting functions at different stimulus levels and signal strengths Jennifer J. Lentz a Department of Speech and Hearing Sciences, Indiana University, Bloomington,

More information

Progressive Recovery of Cortical and Midbrain Sound Feature Encoding Following Profound Cochlear Neuropathy.

Progressive Recovery of Cortical and Midbrain Sound Feature Encoding Following Profound Cochlear Neuropathy. Progressive Recovery of Cortical and Midbrain Sound Feature Encoding Following Profound Cochlear Neuropathy. The Harvard community has made this article openly available. Please share how this access benefits

More information

functions grow at a higher rate than in normal{hearing subjects. In this chapter, the correlation

functions grow at a higher rate than in normal{hearing subjects. In this chapter, the correlation Chapter Categorical loudness scaling in hearing{impaired listeners Abstract Most sensorineural hearing{impaired subjects show the recruitment phenomenon, i.e., loudness functions grow at a higher rate

More information

Neuro-Audio Version 2010

Neuro-Audio Version 2010 ABR PTA ASSR Multi-ASSR OAE TEOAE DPOAE SOAE ECochG MLR P300 Neuro-Audio Version 2010 one device for all audiological tests Auditory brainstem response (ABR)/Brainstem evoked response audiometry (BERA)

More information

Analyzing Auditory Neurons by Learning Distance Functions

Analyzing Auditory Neurons by Learning Distance Functions In Advances in Neural Information Processing Systems (NIPS), MIT Press, Dec 25. Analyzing Auditory Neurons by Learning Distance Functions Inna Weiner 1 Tomer Hertz 1,2 Israel Nelken 2,3 Daphna Weinshall

More information

Supplementary materials for: Executive control processes underlying multi- item working memory

Supplementary materials for: Executive control processes underlying multi- item working memory Supplementary materials for: Executive control processes underlying multi- item working memory Antonio H. Lara & Jonathan D. Wallis Supplementary Figure 1 Supplementary Figure 1. Behavioral measures of

More information

DATA MANAGEMENT & TYPES OF ANALYSES OFTEN USED. Dennis L. Molfese University of Nebraska - Lincoln

DATA MANAGEMENT & TYPES OF ANALYSES OFTEN USED. Dennis L. Molfese University of Nebraska - Lincoln DATA MANAGEMENT & TYPES OF ANALYSES OFTEN USED Dennis L. Molfese University of Nebraska - Lincoln 1 DATA MANAGEMENT Backups Storage Identification Analyses 2 Data Analysis Pre-processing Statistical Analysis

More information

Binaural Hearing. Why two ears? Definitions

Binaural Hearing. Why two ears? Definitions Binaural Hearing Why two ears? Locating sounds in space: acuity is poorer than in vision by up to two orders of magnitude, but extends in all directions. Role in alerting and orienting? Separating sound

More information

9/29/14. Amanda M. Lauer, Dept. of Otolaryngology- HNS. From Signal Detection Theory and Psychophysics, Green & Swets (1966)

9/29/14. Amanda M. Lauer, Dept. of Otolaryngology- HNS. From Signal Detection Theory and Psychophysics, Green & Swets (1966) Amanda M. Lauer, Dept. of Otolaryngology- HNS From Signal Detection Theory and Psychophysics, Green & Swets (1966) SIGNAL D sensitivity index d =Z hit - Z fa Present Absent RESPONSE Yes HIT FALSE ALARM

More information

Twenty subjects (11 females) participated in this study. None of the subjects had

Twenty subjects (11 females) participated in this study. None of the subjects had SUPPLEMENTARY METHODS Subjects Twenty subjects (11 females) participated in this study. None of the subjects had previous exposure to a tone language. Subjects were divided into two groups based on musical

More information

Quick Guide - eabr with Eclipse

Quick Guide - eabr with Eclipse What is eabr? Quick Guide - eabr with Eclipse An electrical Auditory Brainstem Response (eabr) is a measurement of the ABR using an electrical stimulus. Instead of a traditional acoustic stimulus the cochlear

More information

64 IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, VOL. 61, NO. 1, JANUARY 2014

64 IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, VOL. 61, NO. 1, JANUARY 2014 64 IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, VOL. 61, NO. 1, JANUARY 2014 Signal-Processing Strategy for Restoration of Cross-Channel Suppression in Hearing-Impaired Listeners Daniel M. Rasetshwane,

More information

Development of a new loudness model in consideration of audio-visual interaction

Development of a new loudness model in consideration of audio-visual interaction Development of a new loudness model in consideration of audio-visual interaction Kai AIZAWA ; Takashi KAMOGAWA ; Akihiko ARIMITSU 3 ; Takeshi TOI 4 Graduate school of Chuo University, Japan, 3, 4 Chuo

More information

Digital hearing aids are still

Digital hearing aids are still Testing Digital Hearing Instruments: The Basics Tips and advice for testing and fitting DSP hearing instruments Unfortunately, the conception that DSP instruments cannot be properly tested has been projected

More information

Implementation of Spectral Maxima Sound processing for cochlear. implants by using Bark scale Frequency band partition

Implementation of Spectral Maxima Sound processing for cochlear. implants by using Bark scale Frequency band partition Implementation of Spectral Maxima Sound processing for cochlear implants by using Bark scale Frequency band partition Han xianhua 1 Nie Kaibao 1 1 Department of Information Science and Engineering, Shandong

More information

Auditory Phase Opponency: A Temporal Model for Masked Detection at Low Frequencies

Auditory Phase Opponency: A Temporal Model for Masked Detection at Low Frequencies ACTA ACUSTICA UNITED WITH ACUSTICA Vol. 88 (22) 334 347 Scientific Papers Auditory Phase Opponency: A Temporal Model for Masked Detection at Low Frequencies Laurel H. Carney, Michael G. Heinz, Mary E.

More information

ABSTRACT 1. INTRODUCTION 2. ARTIFACT REJECTION ON RAW DATA

ABSTRACT 1. INTRODUCTION 2. ARTIFACT REJECTION ON RAW DATA AUTOMATIC ARTIFACT REJECTION FOR EEG DATA USING HIGH-ORDER STATISTICS AND INDEPENDENT COMPONENT ANALYSIS A. Delorme, S. Makeig, T. Sejnowski CNL, Salk Institute 11 N. Torrey Pines Road La Jolla, CA 917,

More information

Supplementary Figure 1. Localization of face patches (a) Sagittal slice showing the location of fmri-identified face patches in one monkey targeted

Supplementary Figure 1. Localization of face patches (a) Sagittal slice showing the location of fmri-identified face patches in one monkey targeted Supplementary Figure 1. Localization of face patches (a) Sagittal slice showing the location of fmri-identified face patches in one monkey targeted for recording; dark black line indicates electrode. Stereotactic

More information

The Structure and Function of the Auditory Nerve

The Structure and Function of the Auditory Nerve The Structure and Function of the Auditory Nerve Brad May Structure and Function of the Auditory and Vestibular Systems (BME 580.626) September 21, 2010 1 Objectives Anatomy Basic response patterns Frequency

More information

Power-Based Connectivity. JL Sanguinetti

Power-Based Connectivity. JL Sanguinetti Power-Based Connectivity JL Sanguinetti Power-based connectivity Correlating time-frequency power between two electrodes across time or over trials Gives you flexibility for analysis: Test specific hypotheses

More information

Cortical Encoding of Auditory Objects at the Cocktail Party. Jonathan Z. Simon University of Maryland

Cortical Encoding of Auditory Objects at the Cocktail Party. Jonathan Z. Simon University of Maryland Cortical Encoding of Auditory Objects at the Cocktail Party Jonathan Z. Simon University of Maryland ARO Presidential Symposium, February 2013 Introduction Auditory Objects Magnetoencephalography (MEG)

More information