IMPROVING CHANNEL SELECTION OF SOUND CODING ALGORITHMS IN COCHLEAR IMPLANTS. Hussnain Ali, Feng Hong, John H. L. Hansen, and Emily Tobey

Similar documents
A neural network model for optimizing vowel recognition by cochlear implant listeners

International Journal of Scientific & Engineering Research, Volume 5, Issue 3, March ISSN

University of Arkansas at Little Rock. remaining auditory neurons directly.

Implementation of Spectral Maxima Sound processing for cochlear. implants by using Bark scale Frequency band partition

The role of periodicity in the perception of masked speech with simulated and real cochlear implants

Effects of Cochlear Hearing Loss on the Benefits of Ideal Binary Masking

Differential-Rate Sound Processing for Cochlear Implants

REVISED. The effect of reduced dynamic range on speech understanding: Implications for patients with cochlear implants

ADVANCES in NATURAL and APPLIED SCIENCES

Modern cochlear implants provide two strategies for coding speech

CHAPTER 1 INTRODUCTION

Juan Carlos Tejero-Calado 1, Janet C. Rutledge 2, and Peggy B. Nelson 3

Speech enhancement based on harmonic estimation combined with MMSE to improve speech intelligibility for cochlear implant recipients

Fig. 1 High level block diagram of the binary mask algorithm.[1]

HCS 7367 Speech Perception

Masking release and the contribution of obstruent consonants on speech recognition in noise by cochlear implant users

PLEASE SCROLL DOWN FOR ARTICLE

Role of F0 differences in source segregation

Simulations of high-frequency vocoder on Mandarin speech recognition for acoustic hearing preserved cochlear implant

BINAURAL DICHOTIC PRESENTATION FOR MODERATE BILATERAL SENSORINEURAL HEARING-IMPAIRED

Who are cochlear implants for?

What you re in for. Who are cochlear implants for? The bottom line. Speech processing schemes for

An Auditory-Model-Based Electrical Stimulation Strategy Incorporating Tonal Information for Cochlear Implant

Essential feature. Who are cochlear implants for? People with little or no hearing. substitute for faulty or missing inner hair

Hearing the Universal Language: Music and Cochlear Implants

Frequency Tracking: LMS and RLS Applied to Speech Formant Estimation

Adaptive dynamic range compression for improving envelope-based speech perception: Implications for cochlear implants

We are IntechOpen, the world s leading publisher of Open Access books Built by scientists, for scientists. International authors and editors

SOLUTIONS Homework #3. Introduction to Engineering in Medicine and Biology ECEN 1001 Due Tues. 9/30/03

A NOVEL METHOD FOR OBTAINING A BETTER QUALITY SPEECH SIGNAL FOR COCHLEAR IMPLANTS USING KALMAN WITH DRNL AND SSB TECHNIQUE

Noise Susceptibility of Cochlear Implant Users: The Role of Spectral Resolution and Smearing

Essential feature. Who are cochlear implants for? People with little or no hearing. substitute for faulty or missing inner hair

752 IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, VOL. 51, NO. 5, MAY N. Lan*, K. B. Nie, S. K. Gao, and F. G. Zeng

A. SEK, E. SKRODZKA, E. OZIMEK and A. WICHER

Coding Strategies for Cochlear Implants Under Adverse Environments

Prelude Envelope and temporal fine. What's all the fuss? Modulating a wave. Decomposing waveforms. The psychophysics of cochlear

Adaptation of Classification Model for Improving Speech Intelligibility in Noise

NOISE REDUCTION ALGORITHMS FOR BETTER SPEECH UNDERSTANDING IN COCHLEAR IMPLANTATION- A SURVEY

Sonic Spotlight. SmartCompress. Advancing compression technology into the future

Effects of Presentation Level on Phoneme and Sentence Recognition in Quiet by Cochlear Implant Listeners

Hybrid Masking Algorithm for Universal Hearing Aid System

International Journal of Informative & Futuristic Research ISSN:

A channel-selection criterion for suppressing reverberation in cochlear implants

RESEARCH ON SPOKEN LANGUAGE PROCESSING Progress Report No. 22 (1998) Indiana University

Perceptual Effects of Nasal Cue Modification

UvA-DARE (Digital Academic Repository) Perceptual evaluation of noise reduction in hearing aids Brons, I. Link to publication

Computational Perception /785. Auditory Scene Analysis

2/25/2013. Context Effect on Suprasegmental Cues. Supresegmental Cues. Pitch Contour Identification (PCI) Context Effect with Cochlear Implants

HybridMaskingAlgorithmforUniversalHearingAidSystem. Hybrid Masking Algorithm for Universal Hearing Aid System

EEL 6586, Project - Hearing Aids algorithms

Investigating the effects of noise-estimation errors in simulated cochlear implant speech intelligibility

AUDL GS08/GAV1 Signals, systems, acoustics and the ear. Pitch & Binaural listening

Issues faced by people with a Sensorineural Hearing Loss

ReSound NoiseTracker II

Acoustics, signals & systems for audiology. Psychoacoustics of hearing impairment

Representation of sound in the auditory nerve

USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES

Investigating the effects of noise-estimation errors in simulated cochlear implant speech intelligibility

Simulation of an electro-acoustic implant (EAS) with a hybrid vocoder

Challenges in microphone array processing for hearing aids. Volkmar Hamacher Siemens Audiological Engineering Group Erlangen, Germany

ADVANCES in NATURAL and APPLIED SCIENCES

Sonic Spotlight. Binaural Coordination: Making the Connection

LANGUAGE IN INDIA Strength for Today and Bright Hope for Tomorrow Volume 10 : 6 June 2010 ISSN

Exploring the parameter space of Cochlear Implant Processors for consonant and vowel recognition rates using normal hearing listeners

Jitter, Shimmer, and Noise in Pathological Voice Quality Perception

The development of a modified spectral ripple test

Hearing Lectures. Acoustics of Speech and Hearing. Auditory Lighthouse. Facts about Timbre. Analysis of Complex Sounds

FREQUENCY COMPRESSION AND FREQUENCY SHIFTING FOR THE HEARING IMPAIRED

Advanced Audio Interface for Phonetic Speech. Recognition in a High Noise Environment

CONTACTLESS HEARING AID DESIGNED FOR INFANTS

Research Article The Acoustic and Peceptual Effects of Series and Parallel Processing

Spectrograms (revisited)

64 IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, VOL. 61, NO. 1, JANUARY 2014

Time Varying Comb Filters to Reduce Spectral and Temporal Masking in Sensorineural Hearing Impairment

HCS 7367 Speech Perception

SPEECH PERCEPTION IN CHILDREN USING THE ADVANCED SPEAK SPEECH-PROCESSING STRATEGY

1. Human auditory pathway.

Combined spectral and temporal enhancement to improve cochlear-implant speech perception

Speech Enhancement Based on Spectral Subtraction Involving Magnitude and Phase Components

Speech conveys not only linguistic content but. Vocal Emotion Recognition by Normal-Hearing Listeners and Cochlear Implant Users

Sound localization psychophysics

ABSTRACT. Møller A (ed): Cochlear and Brainstem Implants. Adv Otorhinolaryngol. Basel, Karger, 2006, vol 64, pp

Frequency refers to how often something happens. Period refers to the time it takes something to happen.

Hearing Research 241 (2008) Contents lists available at ScienceDirect. Hearing Research. journal homepage:

BORDERLINE PATIENTS AND THE BRIDGE BETWEEN HEARING AIDS AND COCHLEAR IMPLANTS

The effect of wearing conventional and level-dependent hearing protectors on speech production in noise and quiet

SPEECH PERCEPTION IN A 3-D WORLD

THE ROLE OF VISUAL SPEECH CUES IN THE AUDITORY PERCEPTION OF SYNTHETIC STIMULI BY CHILDREN USING A COCHLEAR IMPLANT AND CHILDREN WITH NORMAL HEARING

Proceedings of Meetings on Acoustics

The REAL Story on Spectral Resolution How Does Spectral Resolution Impact Everyday Hearing?

DO NOT DUPLICATE. Copyrighted Material

Effects of speaker's and listener's environments on speech intelligibili annoyance. Author(s)Kubo, Rieko; Morikawa, Daisuke; Akag

Combination of Bone-Conducted Speech with Air-Conducted Speech Changing Cut-Off Frequency

A. Faulkner, S. Rosen, and L. Wilkinson

Chapter 40 Effects of Peripheral Tuning on the Auditory Nerve s Representation of Speech Envelope and Temporal Fine Structure Cues

An evaluation of speech perception when introducing ADRO to adults who use both a cochlear implant and a contralateral hearing aid

University of California

Noise-Robust Speech Recognition in a Car Environment Based on the Acoustic Features of Car Interior Noise

Michael Dorman Department of Speech and Hearing Science, Arizona State University, Tempe, Arizona 85287

Transcription:

2014 IEEE International Conference on Acoustic, Speech and Signal Processing (ICASSP) IMPROVING CHANNEL SELECTION OF SOUND CODING ALGORITHMS IN COCHLEAR IMPLANTS Hussnain Ali, Feng Hong, John H. L. Hansen, and Emily Tobey Center for Robust Speech Systems, Cochlear Implant Laboratory Erik Jonsson School of Engineering and Computer Science The University of Texas at Dallas, Richardson TX hussnain.ali, feng.hong, john.hansen, etobey @utdallas.edu ABSTRACT Spectral maxima sound coding algorithms, for example n-ofm strategies, used in commercial cochlear implant devices rely on selecting channels with the highest energy in each frequency band. This technique works well in quiet, but is inherently problematic in noisy conditions when noise dominates the target, and noise-dominant channels are mistakenly selected for stimulation. A new channel selection criterion is proposed to addresses this shortcoming which adaptively assigns weights to each time-frequency unit based on the formant location of speech and instantaneous signal to noise ratio. The performance of the proposed technique is evaluated acutely with three cochlear implant users in different noise scenarios. Results indicate that the proposed technique improves speech intelligibility and perception quality, particularly at low signal-to-noise ratio. Significance of the proposed technique lies in its ability to be integrated with the existing sound coding framework employed within commercial cochlear implant processors, making it easier to adapt for resource-limited and time critical devices. Index Terms Cochlear implants, sound coding, algorithms. 1. INTRODUCTION Multi-channel cochlear implant (CI) devices operate on the principle of frequency coding based on the tonotopic mapping of the cochlea, and provide stimulation to the auditory neurons via an electrode array implanted along the basilar membrane. In principle, finer frequency resolution can be achieved by increasing the number of electrodes but factors such as current spread (caused by electric stimulation) and the number of surviving auditory neurons constrain the design of CI devices. Electrode arrays used in current CI devices, such as Freedom implant by Cochlear Corporation, are able to provide up to 22 active sites for electric stimulation along the length of the cochlea. However, not all sites are used for stimulation in each stimulation cycle. This is to avoid channel interaction and control the current spread within the cochlea. Continuous Interleaved Sampling (CIS) [1] sound coding algorithm, for example, decomposes the acoustic signal into 8 10 frequency bands/channels. n-of-m strategies, such as the Advanced Combinations Encoder (ACE) [2], decompose the signal in m (22) channels and pick n (8 12) channels in each stimulation cycle. The goal of CI sound coding strategies is to represent the meaningful sound features in a limited number of channels. ACE strategy achieves this by selecting 8 12 channels with the highest spectral energy, which works well in quiet; however, listening performance of CI users drop significantly in noisy environments. This limitation is inherent in the channel selection criterion of the ACE strategy which can mistakenly select channels that are dominated by noise. This is potentially the main reason that CI users are unable to tease apart meaningful features of the target speech from noise because target-dominant channels may never be activated. Therefore, an intelligent channelselection strategy is needed which is able to classify and select channels with the highest amount of target-dominant speech, and not necessarily just energy. A number of noise reduction algorithms for cochlear implants have been proposed over the years which are either based on signal pre-processing [3] [5] or signal conditioning integrated with sound coding [6], [7]. While the former approach can work well in hearing aids, it is potentially susceptible to unwanted signal distortion which can easily be enhanced by CI processing (e.g., compression function emphasizes low energy sounds logarithmically) or it can be computationally intensive, thus making it unfavorable for CI processors. The later approach generally relies on spectral modification or modifying channel-selection based on the signal to noise ratio (SNR). Hu et al. [6], for example, used a sigmoidal-shaped function that applies attenuation to the noisy envelopes (computed by CIS strategy) inversely proportional to the estimated SNR in each channel. By discarding channels with an SNR < 0 db (binary masking) and varying the number of active channels, [7] reported restoration of speech intelligibility to the level attained in quiet when the channel SNR was known (ideal condition). While this approach can work well in ideal conditions, one 978-1-4799-2893-4/14/$31.00 2014 IEEE 905

clear disadvantage is that binary masking would completely discard channels containing speech components that are essential for speech intelligibility, but are either unfortunately dominated by noise or wrongly classified by the noise estimation algorithm. The technique proposed in this paper takes its inspiration from the former two approaches and shapes the weighting functions used in the ACE processing based on the instantaneous SNR of each time frequency (TF) unit. In addition, priority is assigned to channels containing the three speech formants, F1, F2, and F3. This paper is organized as follows. Sec. 2 describes the proposed technique which improves the channel selection in n-of-m strategy. Sec. 3 presents evaluation of the proposed technique with 3 CI human subjects followed by discussion and conclusion in Sec. 4. 2. METHOD In the clinical/standard ACE (STD_ACE) strategy (Fig. 1, inside the dotted block), the acoustic signal is sampled at 16 khz, pre-emphasized, and buffered using a Blackman window into 8 ms (128 samples) analysis frames. Frame overlapping (or analysis rate) typically depends on the channel stimulation rate. For each analysis frame, 128 point FFT and magnitude squared spectrum is computed, thereby giving 64 frequency bins, each bin has a frequency resolution of 125 Hz. These bins are passed through 22 weighting filters (Fig. 2), which essentially computes the envelope of each channel. Next, 8-12 channels with the highest amplitudes are selected and compressed to the current levels using a loudness growth function (LGF) and the patient s clinical MAP, which maps the acoustic amplitudes to the patient s electrical dynamic range. Fig. 1 shows the block diagram of the proposed technique (in conjunction with the STD_ACE routine). The proposed technique operates based on two principles, 1) by assigning priority to formant bands and 2) by assigning weights to each TF unit based on the instantaneous SNR. 2.1 Assigning priority to the formant bands The frequencies of the first three speech formant (F1 F3) peaks as well as their trajectory over time provide valuable cues to listeners for vowels, glides and stop-consonant perception [8]. This is the reason that feature-extraction strategies [9] [13] have been popular in earlier generation CI processors of 1980s and 1990s. F0/F2 [9], [10], and F0/F1/F2 [11] strategies extract formant locations (F1 and F2) and stimulate the corresponding electrodes at a rate of F0 pulses/sec (pps) for voiced segments and an average rate of 100 pps for unvoiced segments. The MULTIPEAK (MPEAK) strategy [12], [13] stimulates four electrodes at a time and always activates electrode numbers 4 and 7 for F1 and F2 respectively, and then selects the remaining two based on the spectral content of the speech signal. Spectral maxima based sound strategies were later adapted to encode the entire spectrum of the speech signal of whom ACE is the prime example. The shortcoming in spectral maxima algorithms, as noted earlier, comes from the fact that channel selection is based on the largest filter amplitudes which are not necessarily the spectral peaks, and hence may not encode the major formant frequencies. Several maxima may come from a single spectral peak [14]. This can be problematic in noise, which tends to reduce the dynamic range of the spectrum as well as the spectral contrast (peak-to-valley ratio on LPC spectrum). Thus, preference will be given to noise dominant channels irrespective of the presence/absence of spectral peaks. In fairness, the F1 spectral peak is preserved to a certain degree in noise which works to the advantage of ACE. Although, the location of peaks of higher formants may not be affected as much in noise, spectral smearing and reduced spectral contrast would give preference to the noise dominant Fig.1. Signal flow in the standard ACE strategy (shown inside the dotted block). Processing blocks for the proposed technique are shown in the darker tone. Numbers on connecting arrows represent the frame size in number of samples at each step. 906

Fig. 2. Weighting filters used in the ACE strategy plotted in solid line. Spectrum shaping by soft-masking technique shapes the spectrum based on SNR of each TF unit as shown by the color-filled plots. channels. The proposed technique continuously computes the first three formant peaks (F1 F3) in each analysis cycle and assigns priority to the channels corresponding to the formant frequencies during the channel selection process. Formant frequencies are computed by solving for the roots of the linear prediction coefficients (LPC). Formant continuity constraints are imposed to avoid unwanted distortion. 2.2 Assigning attenuation factor based on the SNR Both [6] and [7] apply binary and soft masking techniques respectively to channels with low SNR. Given that a channel can comprise of as many as 8 frequency bins, for example, for higher frequencies (more if the number of stimulations sites is less than 20), channel classification would be compromised for bin-widths of 2 or more. The proposed technique estimates the SNR for each TF unit, X(i,j), where X is magnitude squared spectrum of the i th analysis frame and j th frequency bin. This yields a total of 64 SNR values for each analysis frame (stimulation cycle). Based on the computed SNR, an attenuation factor is generated. In the present study, we analyzed both binary and soft masking techniques. In binary masking, a weight of 0 was assigned to the SNR(i,j) < 0 db, for the rest (i.e., SNR(i,j) 0 db), a binary value of 1 was assigned. In the soft masking approach, a sigmoidal-shaped function was considered which plateaus for SNRs > 15 db and floors to 0 for SNRs < -15dB. Both weighting functions are shown in Fig. 3. The 64 weighting values generated for each TF unit are then used to shape the gain of weighting functions. This is illustrated in Fig. 2 for the soft masking technique. In order to evaluate the effectiveness of the proposed technique, tests were first performed with the SNR of each TF unit known a priori. The results from this experiment would validate if the proposed technique is Fig. 3. Binary weighting (red) and soft-masking (blue) attenuation functions. effective. In the second phase, the instantaneous SNR of each TF unit was estimated using improved minimum controlled recursive average (IMCRA) algorithm [15]. While any SNR estimation algorithm could be used, IMCRA utilizes the spectrum components of each TF unit, X(i,j), which are already computed by ACE. Furthermore, the advantages of the IMCRA method are particularly notable in adverse environments involving nonstationary noise, weak speech components, and low SNR conditions. 3. PROCEDURE, EVALUATION, AND RESULTS A total of 3 CI users participated in this acute study. All participants were native speakers of American English and fitted with Nucleus 24 multichannel device manufactured by Cochlear Corp. All participants used ACE as their speech processing strategy. IEEE sentences [16] were used as the speech stimuli for testing. 10 sentences for each test condition were used. The algorithms were implemented offline in MATLAB and stimuli were presented via UT Dallas s PDA-based research platform [17]. Two sets of experiments were conducted. In the first set, the effectiveness of assigning priority to formant channels (F1, F2, and F3) was evaluated both in terms of speech intelligibility and perception quality. All words were marked for correctness. A total of 6 conditions were tested in experiment 1, namely speech in quiet, speech in 10 db SNR speech shaped noise (SSN), speech in 5 db SNR SSN, speech in 10 db SNR white Gaussian noise (WGN), speech in reverberation with reverberation time T 60 = 600 ms, and finally, speech in reverberation (T 60 = 600 ms) and 10dB noise. Fig. 4 shows the mean intelligibility scores for STD_ACE and formant-based ACE (FRMNTS_ACE) technique. While there is very little to no significant improvement in intelligibility at high SNRs, there was an improvement of 17% for speech at 5 db SNR SSN, and 20% when noise was added to the reverberant signal (Rev600+n). Modest improvement was also observed for speech in WGN at 10 db SNR. For the speech quality tests, the same sentences processed with STD_ACE and FRMNTS_ACE strategies 907

Percent Correct 100% 90% 80% 70% 60% 50% 40% 30% 20% 10% 0% STD_ACE FRMNTS_ACE Quiet 10dB SSN 5dB SSN Rev600 Rev600+n WGN10dB Fig. 4. Mean speech intelligibility scores of 3 CI users in Exp. 1. Error bars represent standard error of the mean. was streamed back to back and CI users were asked to rate the quality of the second sentence as compared to the first in terms of being pleasant, clear and free of sound distortions on a scale of -3 to 3, with 0 being about the same, -3 being much worse, and +3 indicating much better. On average, the user response was between +1 and 0, indicating slightly better to no difference in all test conditions. The subjects reported that words popped out more in the FRMNTS_ACE strategy. In the second experiment, four techniques were assessed separately, namely ideal-binary (IdBinary), idealsoft (IdSoft), estimated-binary (EsBinary), and estimatedsoft (EsSoft). Ideal conditions represent when the SNR was known a priori, while the estimated conditions represent when the SNR was estimated using IMCRA. Binary conditions correspond to the output of the attenuation factor to binary values, while the soft conditions represent output from the sigmoidal attenuation function, as described earlier. Speech intelligibility and quality measures were assessed for each technique with speech in 10 db SNR SSN, and speech in 5 db SNR SSN. The results from STD_ACE were used as baseline scores for comparison. The mean intelligibility scores for experiment 2 are presented in Fig. 5. The results show that both IdBinary and IdSoft techniques are able to restore speech intelligibility to the level equivalent to speech in quiet. This establishes the effectiveness of the proposed technique in masking noise if cues to the SNR are available. Fig. 5 also presents intelligibility scores when the noise is estimated. There was no significant improvement at 10 db SNR level. However, at 5 db SNR, improvement is seen with both EsBinary and EsSoft, but later results in significant gains in intelligibility (>20 percent). The results indicate that the proposed technique can potentially improve speech intelligibility in low-snr conditions. Quality tests for Experiment 2 indicate an average score of Much Better as compared to the STD_ACE (unprocessed) for both ideal conditions. For the estimated SNR at 10 db and 5 db SNR levels, the average score was (+2) corresponding to Better as compared to the Percent Correct 100% 90% 80% 70% 60% 50% 40% 30% 20% 10% 0% 10dB SSN 5dB SSN IdBinary IdSoft EsBinary EsSoft STD_ACE Fig. 5. Speech intelligibility scores for Experiment 2. Comparison of the proposed technique using IdBinary, IdSoftm, EsBinary, and EsSoft approaches with STD_ACE strategy. Error bars represent standard error of the mean. STD_ACE. The quality scores between binary and soft masking approaches indicate slight preference for the soft masking approach, with an average score of +0.5. 4. CONCLUSION This study considered two approaches for improving the channel selection process of spectral maxima sound coding algorithms for cochlear implant systems. The first approach assigns priority to the channels containing the formant frequencies, while the second approach adaptively assigns weights to each time-frequency (TF) unit based on the estimated SNR in each stimulation cycle. Two types of maskers, binary and soft, were used to assign attenuation factor to each TF unit based on SNR. Both approaches were evaluated independently and synergistically on 3 CI users. The intelligibility scores indicate significant improvement at low SNR levels, with speech at 5 db SNR and speech masked by reverberation and noise. The quality scores revealed very little preference to the formants-based approach over the standard ACE, and soft-masking over the binary masking. However, the noise masking approach was greatly preferred over standard ACE. The intelligibility scores for the two approaches revealed some benefit of the formants-based technique over the standard ACE, and soft masking over binary masking at low SNR levels. The proposed technique is inherently limited by the accuracy of formant locations and SNR estimation. Speech intelligibility reached to the level obtained in quiet when SNR was known a priori. The results from this study indicate the potential of the proposed technique to improve speech perception of CI users in adverse listening environments. ACKNOWLEDGEMENT This work was supported by the National Institutes of Health under contract N01-DC-6-0002 from the National Institute on Deafness and Other Communication Disorders and R01 DC010494-01A. 908

12. REFERENCES [1] B. Wilson, C. Finley, D. T. Lawson, R. D. Wolford, D. K. Eddington, W. M. Rabinowitz, Better speech recognition with cochlear implants, Nature, vol. 352, pp. 236-238, July 1991. [2] A. E. Vandali, L. A. Whitford, K. L. Plant, and G. M. Clark,, Speech perception as a function of electrical stimulation rate using the Nucleus 24 cochlear implant system, Ear and Hearing, vol. 21 (6), pp. 608 624, 2000. [3] J. Wouters, and J. B. Vanden., Speech recognition in noise for cochlear implantees with a two microphone monaural adaptive noise reduction system, Ear and Hearing, vol. 22, issue 5, pp. 420 430, 2001. [4] P. C. Loizou, Speech processing in vocoder-centric cochlear implants, Adv. Oto-Rhino-Laryngol. vol. 64, pp. 109 143. [5] R. V. Hoesel, and G. Clark, Evaluation of a portable twomicrophone adaptive beamforming speech processor with cochlear implant patients, J. Acoust. Soc. Am. vol. 97, issue 4, pp. 2498 2503. [6] Y. Hu, P. C. Loizou, N. Li, K Kasturi, Use of a sigmoidalshaped function for noise attenuation in cochlear implants, J. Acoust. Soc. Am. vol. 122, issue 4, 2007. [7] Y. Hu, P. C. Loizou, A new sound coding strategy for suppressing noise in cochlear implants, J. Acoust. Soc. Am. vol. 124, issue 1, pp. 498 509, 2008. [8] P. C. Loizou, Speech Enhancement Theory and Practice, CRC Press, 2012. [9] G. Clark, The University of Melbourne-Nucleus multielectrode cochlear implant, Advances in Oto-Rhino- Laryngology, vol. 38, pp. 1 189, 1987. [10] P. Seligman, J. Patrick, Y. Tong, G. Clark, R. Dowell, and P.Crosby, A signal processor for a multiple-electrode hearing prosthesis, Acta Otolaryngologica, pp. 135 139. (Suppl. 411), 1984. [11] P. Blamey, R. Dowell, and G. Clark, Speech recognition for 40 patients receiving multichannel cochlear implants, Archives of Otolaryngology, Head and Neck Surgery, vol. 112, pp. 1054 1059, 1986. [12] J. Patrick, P. Seligman, D. Money, and J. Kuzma, Engineering, in Cochlear prosthesis (G. Clark, Y. Tong, and J. Patrick, eds.), Edinburgh: Churchill Livingstone, 1990, pp. 99-124. [13] J. Patrick and G. Clark, The Nucleus 22-channel cochlear implant system, Ear and Hearing, vol. 12, pp. 3 9, 1991. [14] P. C. Loizou, Mimicking the human ear, IEEE Signal Processing Magazine, vol. 15, no. 5, pp. 101 130, Sep. 1998. [15] I. Cohen, Noise spectrum estimation in adverse environments: Improved minima controlled recursive averaging, IEEE Transactions on Speech and Audio Processing, vol. 11, issue 5, pp. 466-475, 2003. [16] IEEE Subcommittee, "IEEE Recommended Practice for Speech Quality Measurements," IEEE Trans. Audio and Electroacoustics, vol. AU-17, no. 3, pp. 225-246, 1969. [17] H. Ali, A. P. Lobo, P. C. Loizou, "Design and Evaluation of a Personal Digital Assistant-based Research Platform for Cochlear Implants," IEEE Transactions on Biomedical Engineering, vol.60, no.11, pp.3060,3073, Nov. 2013. 909