Channel selection systems and methods that employ temporal modification
An exemplary sound processor may 1) divide an audio signal into M analysis channels, 2) select only N analysis channels included in the M analysis channels for presentation to a patient during a stimulation frame, wherein N is less than M, 3) increase a probability that a particular analysis channel will be included in the N analysis channels selected for presentation to the patient during the stimulation frame if the particular analysis channel was not selected for presentation to the patient during one or more stimulation frames that temporally precede the stimulation frame, and 4) decrease the probability that the particular analysis channel will be included in the N analysis channels selected for presentation to the patient during the stimulation frame if the particular analysis channel was selected for presentation to the patient during the one or more stimulation frames that temporally precede the stimulation frame.
Latest Advanced Bionics AG Patents:
- Systems and methods for measuring evoked responses from a brain of a patient
- Intraoperative detection of electrode lead misplacement during a lead insertion procedure
- Apparatus and methods for opposing torque imparted onto cochlear implant magnets by magnetic fields
- Systems and methods for remote loading of a sound processing program onto a sound processor included within a cochlear implant system
- Microphone assembly for use with an off-the-ear sound processor
Various N-of-M channel selection strategies for cochlear implant systems have been described in which a cochlear implant system only stimulates a subset of the available electrodes in each given stimulation frame. In these N-of-M channel selection strategies, a cochlear implant system divides an incoming audio signal into M analysis channels (or simply “channels”) and then presents only N analysis channels to the patient (e.g., by applying electrical stimulation representative of the signals contained within the N analysis channels by way of a plurality of intracochlear electrodes).
In one conventional N-of-M channel selection strategy, a cochlear implant system selects only the analysis channels with the highest amplitude signals for presentation to a patient during a particular stimulation frame. This means that all information in the lower amplitude channels is lost during that frame. This could be disadvantageous in situations where the overall frequency distribution remains relatively constant for a period of time, such as when the patient is listening in certain noisy environments or detecting background sounds during vowels. One example of this would be someone honking a horn while someone is talking. If the horn is loud enough, its spectral content would overwhelm the talker, and this N-of-M channel selection strategy would only deliver envelope information to the pulse generator for those channels which contain “horn content”. All of the other channels would be effectively muted.
Other N-of-M channel selection strategies use temporal masking characteristics to select the most relevant channels for presentation to a patient during a particular stimulation frame. These temporal masking approaches take into account the refractory phenomena associated with auditory nerve fibers by decreasing the probability that a particular channel that has been selected for presentation during a stimulation frame will be again selected for presentation during one or more stimulation frames that immediately follow the stimulation frame. However, these temporal masking approaches have heretofore been computationally intensive, requiring computation of masking models during each stimulation frame. Moreover, these temporal masking approaches do not specifically increase the probability that a particular channel that has not been selected for a relatively long time will be again selected, thus causing the patient to miss out on the information contained within that channel.
The accompanying drawings illustrate various embodiments and are a part of the specification. The illustrated embodiments are merely examples and do not limit the scope of the disclosure. Throughout the drawings, identical or similar reference numbers designate identical or similar elements.
Channel selection systems and methods that employ temporal modification for use in cochlear implant systems are described herein. As will be described below, an exemplary sound processor included in a cochlear implant system associated with a patient may 1) divide an audio signal into M analysis channels, 2) select only N analysis channels included in the M analysis channels for presentation to the patient during a stimulation frame, wherein N is less than M, 3) increase a probability that a particular analysis channel included in the M analysis channels will be included in the N analysis channels selected for presentation to the patient during the stimulation frame if the particular analysis channel was not selected for presentation to the patient during one or more stimulation frames that temporally precede the stimulation frame, and 4) decrease the probability that the particular analysis channel will be included in the N analysis channels selected for presentation to the patient during the stimulation frame if the particular analysis channel was selected for presentation to the patient during the one or more stimulation frames that temporally precede the stimulation frame.
By employing temporal modification (i.e., by increasing a probability that a particular analysis channel will be selected for presentation during a stimulation frame if it was not selected during one or more stimulation frames that temporally precede the stimulation frame and decreasing the probability that the particular analysis channel will be selected for presentation during the stimulation frame if it was selected during the one or more stimulation frames that temporally precede the stimulation frame), the systems and methods described herein may achieve complete spectral coverage while taking into account the temporal masking phenomena described above. Moreover, the systems and methods described herein are relatively computationally efficient, thereby saving power and providing other processing benefits.
Microphone 102 may be configured to detect audio signals presented to the patient. Microphone 102 may be implemented in any suitable manner. For example, microphone 102 may include a microphone that is configured to be placed within the concha of the ear near the entrance to the ear canal, such as a T-MIC™ microphone from Advanced Bionics. Such a microphone may be held within the concha of the ear near the entrance of the ear canal by a boom or stalk that is attached to an ear hook configured to be selectively attached to sound processor 104. Additionally or alternatively, microphone 102 may be implemented by one or more microphones disposed within headpiece 106, one or more microphones disposed within sound processor 104, one or more beam-forming microphones, and/or any other suitable microphone as may serve a particular implementation.
Sound processor 104 (i.e., one or more components included within sound processor 104) may be configured to direct cochlear implant 108 to generate and apply electrical stimulation (also referred to herein as “stimulation current”) representative of one or more audio signals (e.g., one or more audio signals detected by microphone 102, input by way of an auxiliary audio input port, etc.) to one or more stimulation sites associated with an auditory pathway (e.g., the auditory nerve) of the patient. Exemplary stimulation sites include, but are not limited to, one or more locations within the cochlea, the cochlear nucleus, the inferior colliculus, and/or any other nuclei in the auditory pathway. To this end, sound processor 104 may process the one or more audio signals in accordance with a selected sound processing strategy or program to generate appropriate stimulation parameters for controlling cochlear implant 108. Sound processor 104 may include or be implemented by a behind-the-ear (“BTE”) unit, a body worn device, and/or any other sound processing unit as may serve a particular implementation. For example, sound processor 104 may be implemented by an electro-acoustic stimulation (“EAS”) sound processor included in an EAS system configured to provide electrical and acoustic stimulation to a patient.
In some examples, sound processor 104 may wirelessly transmit stimulation parameters (e.g., in the form of data words included in a forward telemetry sequence) and/or power signals to cochlear implant 108 by way of a wireless communication link 114 between headpiece 106 and cochlear implant 108. It will be understood that communication link 114 may include a bi-directional communication link and/or one or more dedicated uni-directional communication links.
Headpiece 106 may be communicatively coupled to sound processor 104 and may include an external antenna (e.g., a coil and/or one or more wireless communication components) configured to facilitate selective wireless coupling of sound processor 104 to cochlear implant 108. Headpiece 106 may additionally or alternatively be used to selectively and wirelessly couple any other external device to cochlear implant 108. To this end, headpiece 106 may be configured to be affixed to the patient's head and positioned such that the external antenna housed within headpiece 106 is communicatively coupled to a corresponding implantable antenna (which may also be implemented by a coil and/or one or more wireless communication components) included within or otherwise associated with cochlear implant 108. In this manner, stimulation parameters and/or power signals may be wirelessly transmitted between sound processor 104 and cochlear implant 108 via a communication link 114 (which may include a bi-directional communication link and/or one or more dedicated uni-directional communication links as may serve a particular implementation).
Cochlear implant 108 may include any type of implantable stimulator that may be used in association with the systems and methods described herein. For example, cochlear implant 108 may be implemented by an implantable cochlear stimulator. In some alternative implementations, cochlear implant 108 may include a brainstem implant and/or any other type of active implant or auditory prosthesis that may be implanted within a patient and configured to apply stimulation to one or more stimulation sites located along an auditory pathway of a patient.
In some examples, cochlear implant 108 may be configured to generate electrical stimulation representative of an audio signal processed by sound processor 104 (e.g., an audio signal detected by microphone 102) in accordance with one or more stimulation parameters transmitted thereto by sound processor 104. Cochlear implant 108 may be further configured to apply the electrical stimulation to one or more stimulation sites within the patient via one or more electrodes 112 disposed along lead 110 (e.g., by way of one or more stimulation channels formed by electrodes 112). In some examples, cochlear implant 108 may include a plurality of independent current sources each associated with a channel defined by one or more of electrodes 112. In this manner, different stimulation current levels may be applied to multiple stimulation sites simultaneously (also referred to as “concurrently”) by way of multiple electrodes 112.
Spectral analysis facility 302 may divide an audio signal (e.g., an audio signal presented to the patient and representative of speech and/or other types of sound) into M analysis channels. Each analysis channel includes a signal representative of a distinct portion of the audio signal. As used herein, “M” is an integer and refers to a total number of analysis channels into which an audio signal may be divided. For example, M may be any integer greater than three (e.g., equal to or greater than eight). In some examples, M is equal to the total number of stimulation channels formed by electrodes 112 through which electrical stimulation representative of the audio signal may be applied to the patient.
Spectral analysis facility 302 may divide the audio signal into M analysis channels in any suitable manner. For example, spectral analysis facility 306 may be implemented by a plurality of band-pass filters configured to divide the audio signal into a plurality of frequency channels or bands. Additionally or alternatively, spectral analysis facility 306 may be configured to convert the audio signal from a time domain into a frequency domain and then divide the resulting frequency bins into the plurality of analysis channels. To this end, spectral analysis facility 206 may include one or more components configured to apply a Discrete Fourier Transform (e.g., a Fast Fourier Transform (“FFT”)) to the audio signal.
Each analysis channel 402 may correspond to a distinct frequency band, and, as shown, includes a signal (e.g., signal 404) representative of a distinct portion of the audio signal. The vertical axis in
Returning to
As used herein, an analysis channel that has been “selected for presentation” to a patient during a stimulation frame refers to an analysis channel whose signal contained therein will be represented by electrical stimulation applied to the patient during the stimulation frame. In other words, once an analysis channel has been selected for presentation to the patient during the stimulation frame, sound processor 104 (e.g., channel selection facility 304) may direct cochlear implant 108 to apply electrical stimulation representative of the signal contained within the analysis channel to the patient. For example, with reference again to
As used herein, a “stimulation frame” refers to a period of time during which electrical stimulation representative of the N selected analysis channels is applied to one or more stimulation channels defined by one or more of electrodes 112. During a stimulation session, a plurality of stimulation frames may temporally follow one another. For example, a first stimulation frame, a second stimulation frame, a third stimulation frame, etc. may follow sequentially right after one another as an audio signal is represented to a cochlear implant patient.
Channel selection facility 304 may select the N analysis channels for presentation to the patient in accordance with any suitable N-of-M channel selection strategy. For example, channel selection facility 304 may simply select the N analysis channels that contain the highest average energy levels for presentation to the patient. To illustrate, if N is equal to 4, channel selection facility 304 may select the four analysis channels with the highest average energy levels for presentation to a patient. In the example of
Hence, in some examples, channel selection facility 304 may select the N analysis channels for presentation to the patient in accordance with an N-of-M channel selection strategy that employs temporal modification. For example, channel selection facility 304 may increase a probability that a particular analysis channel (e.g., analysis channel 402-2) included in the M analysis channels will be included in the N analysis channels selected for presentation to the patient during a stimulation frame if the particular analysis channel was not selected for presentation to the patient during one or more stimulation frames that temporally precede the stimulation frame. Likewise, channel selection facility 304 may decrease the probability that a particular analysis channel (e.g., analysis channel 402-4) will be included in the N analysis channels selected for presentation to the patient during the stimulation frame if the particular analysis channel was selected for presentation to the patient during one or more stimulation frames that temporally precede the stimulation frame.
To illustrate,
As shown in
As used herein, a “selection signal” refers to a signal that has not been modified by temporal modification block 504 and that represents a particular characteristic of the signal contained within the analysis channel that corresponds to the selection signal. For example, each selection signal may represent an energy level (e.g., an average energy level, a total energy level, and/or a peak energy level), a spectrally enhanced signal, and/or any other characteristic of the signals contained within the analysis channels. The values (i.e., E1 through E8) of the selection signals represent the values of the particular signal characteristic represented by the selection signals. For example, if the selection signals represent an energy level, the values contained within selection signal array 502 may represent the average energy contained within each of the M analysis channels. As shown, each selection signal may be represented mathematically by E[J,T], where J represents the analysis channel number and T represents the stimulation frame.
In some conventional N-of-M channel selection strategies, the selection signals represented by E[J,T] are used to select N analysis channels for presentation to the patient. However, in accordance with the systems and methods described herein, the selection signals represented by E[J,T] are not used to select N analysis channels for presentation to the patient. Rather, the selection signals represented by E[J,T] are input into temporal modification block 504, which, as described below, modifies the selection signals based on whether their corresponding analysis channels have been selected for presentation during one or more stimulation frames that temporally precede the stimulation frame. The modified selection signals are represented by E′[J,T] and, like the selection signals represented by E[J,T], each have an individual value. These individual values may quantify a characteristic (e.g., an amount of energy) represented by each of the modified selection signals and may be used by N-of-M channel selection block 506 to select the N analysis channels for presentation to the patient.
For example, N-of-M channel selection block 506 may identify the N highest values out of the individual values of the modified selection signals. The N-of-M channel selection block 506 may then designate the analysis channels that correspond to the N highest values as the N analysis channels that are selected for presentation to the patient during the stimulation frame. This may be performed in any suitable manner. For example, as will be described below, N-of-M channel selection block 506 may assign a binary value to each analysis channel that indicates whether each analysis channel has been selected for presentation during the stimulation frame.
Temporal modification block 504 may generate the modified selection signals represented by E′[J,T] in any suitable manner. For example, if temporal modification block 504 determines that a particular analysis channel was not selected for presentation to the patient during one or more stimulation frames that temporally precede the stimulation frame, temporal modification block 504 may increase a value of the modified selection signal corresponding to the particular analysis channel (e.g., by adding gain to the selection signal that corresponds to the particular analysis channel). As will be described below, this may increase the probability that the particular analysis channel will be included in the N analysis channels selected by N-of-M channel selection block 506 for presentation to the patient during the stimulation frame.
As another example, if temporal modification block 504 determines that a particular analysis channel was selected for presentation to the patient during the one or more stimulation frames that temporally precede the stimulation frame, temporal modification block 504 may decrease a value of the modified selection signal corresponding to the particular analysis channel (e.g., by subtracting gain from the selection signal that corresponds to the particular analysis channel). As will be described below, this may decrease the probability that the particular analysis channel will be included in the N analysis channels selected by N-of-M channel selection block 506 for presentation to the patient during the stimulation frame.
As shown in
For example, as shown, the output of N-of-M channel selection block 506 is represented by the binary values contained within selection state array 508. These values are collectively represented by D[J,T]. As shown, selection state array 508 includes a cell that corresponds to each analysis channel. Hence, the binary value included in each cell indicates whether the analysis channel has been selected for presentation to the patient during a particular stimulation frame. In the particular example of
The contents of selection state array 508 may be output to smoothing function block 510, which may output a smoothed signal represented by D′[J,T] based on the contents of selection state array 508. One of the purposes of smoothing function block 510 is to modify the number of stimulation frames that are taken into account by temporal modification block 504. In some examples, temporal modification block 504 will only take into account a stimulation frame that immediately and temporally precedes the stimulation frame represented by T (i.e., a stimulation frame represented by T-1). In alternative examples, temporal modification block 504 will take into account a plurality of stimulation frames that immediately and temporally precede the stimulation frame. As used herein, if a stimulation frame “immediately and temporally” precedes another stimulation frame, no other stimulation frames are temporally interposed in between the two stimulation frames.
The output of smoothing function block 510 may be represented mathematically by D′[J,T]=D[J,T-1]+α*(D′[J,T-1]−D[J,T-1]), where a represents a user definable constant that controls an integration time associated with the smoothed signal.
It will be recognized that smoothing function block 510 may alternatively not be included within implementation 500. However, assuming that smoothing function block 510 is included in implementation 500, the modified selection signals output by temporal modification block 504 may be represented by E′[J,T]=E[J,T]+(D′[J,T]−0.5) * G, where G represents a user definable amount of gain.
Hence, in accordance with this equation, if, for a particular analysis channel, the signal D′[J,T] is greater than 0.5 (i.e., the analysis channel is dropped for one or more stimulation frames), E′[J,T] will be greater than E[J,T], thus increasing the probability that the particular analysis channel will be selected for presentation during the stimulation frame represented by T. Alternatively, if the signal D′[J,T] is less than 0.5 (i.e., the analysis channel is selected for one or more stimulation frames), E′[J,T] will be less than E[J,T], thus decreasing the probability that the particular analysis channel will be selected for presentation during the stimulation frame represented by T.
In step 602, a sound processor divides an audio signal into M analysis channels. Step 602 may be performed in any of the ways described herein.
In step 604, the sound processor selects only N analysis channels included in the M analysis channels for presentation to a patient during a stimulation frame, wherein N is less than M. Step 604 may be performed in any of the ways described herein.
In step 606, the sound processor increases a probability that a particular analysis channel included in the M analysis channels will be included in the N analysis channels selected for presentation to the patient during the stimulation frame if the particular analysis channel was not selected for presentation to the patient during one or more stimulation frames that temporally precede the stimulation frame. Step 606 may be performed in any of the ways described herein.
In step 608, the sound processor decreases the probability that the particular analysis channel will be included in the N analysis channels selected for presentation to the patient during the stimulation frame if the particular analysis channel was selected for presentation to the patient during the one or more stimulation frames that temporally precede the stimulation frame. Step 608 may be performed in any of the ways described herein.
In the preceding description, various exemplary embodiments have been described with reference to the accompanying drawings. It will, however, be evident that various modifications and changes may be made thereto, and additional embodiments may be implemented, without departing from the scope of the invention as set forth in the claims that follow. For example, certain features of one embodiment described herein may be combined with or substituted for features of another embodiment described herein. The description and drawings are accordingly to be regarded in an illustrative rather than a restrictive sense.
Claims
1. A sound processor included in a cochlear implant system associated with a patient and comprising:
- a spectral analysis facility that divides an audio signal into M analysis channels; and
- a channel selection facility communicatively coupled to the spectral analysis facility and that selects only N analysis channels included in the M analysis channels for presentation to the patient during a stimulation frame, wherein N is less than M, increases a probability that a particular analysis channel included in the M analysis channels will be included in the N analysis channels selected for presentation to the patient during the stimulation frame if the particular analysis channel was not selected for presentation to the patient during one or more stimulation frames that temporally precede the stimulation frame, and decreases the probability that the particular analysis channel will be included in the N analysis channels selected for presentation to the patient during the stimulation frame if the particular analysis channel was selected for presentation to the patient during the one or more stimulation frames that temporally precede the stimulation frame.
2. The sound processor of claim 1, wherein the channel selection facility selects the N analysis channels for presentation to the patient during the stimulation frame by:
- generating M modified selection signals having individual values and each corresponding to a different analysis channel included in the M analysis channels, the M modified selection signals including a modified selection signal corresponding to the particular analysis channel;
- identifying N highest values out of the individual values, the N highest values corresponding to a subset of modified selection signals included in the M modified selection signals, the subset of modified selection signals corresponding to a subset of analysis channels included in the M analysis channels; and
- designating the subset of analysis channels that correspond to the subset of modified selection signals as the N analysis channels that are selected for presentation to the patient during the stimulation frame.
3. The sound processor of claim 2, wherein the channel selection facility:
- determines that the particular analysis channel was not selected for presentation to the patient during the one or more stimulation frames that temporally precede the stimulation frame; and
- increases, in response to the determination that the particular analysis channel was not selected for presentation to the patient during the one or more stimulation frames that temporally precede the stimulation frame, the probability that the particular analysis channel will be included in the N analysis channels selected for presentation to the patient during the stimulation frame by increasing a value of the modified selection signal corresponding to the particular analysis channel prior to the identifying of the N highest values.
4. The sound processor of claim 3, wherein the increasing the value of the modified selection signal corresponding to the particular analysis channel is performed by adding gain to a selection signal corresponding to the particular analysis channel.
5. The sound processor of claim 2, wherein the channel selection facility:
- determines that the particular analysis channel was selected for presentation to the patient during the one or more stimulation frames that temporally precede the stimulation frame; and
- decreases, in response to the determination that the particular analysis channel was selected for presentation to the patient during the one or more stimulation frames that temporally precede the stimulation frame, the probability that the particular analysis channel will be included in the N analysis channels selected for presentation to the patient during the stimulation frame by decreasing a value of the modified selection signal corresponding to the particular analysis channel prior to the identifying of the N highest values.
6. The sound processor of claim 5, wherein the decreasing the value of the modified selection signal corresponding to the particular analysis channel is performed by subtracting gain from a selection signal corresponding to the particular analysis channel.
7. The sound processor of claim 2, wherein the individual values quantify an amount of energy represented by the M modified selection signals.
8. The sound processor of claim 2, wherein the generating of the M modified selection signals is performed in accordance with E′[J,T]=E[J,T]+(D′[J,T]−0.5) * G, wherein D′[J,T]=D[J,T-1]+α*(D′[J,T-1]−D[J,T-1]), and wherein:
- J represents the particular analysis channel;
- T represents the stimulation frame;
- T-1 represents a stimulation frame that temporally and immediately precedes the stimulation frame;
- G represents gain;
- E′[J,T] represents the modified selection signal corresponding to the particular analysis channel;
- E[J,T] represents a selection signal corresponding to the particular analysis channel;
- D[J,T-1] is a binary value indicative of whether the particular analysis channel was selected for presentation during the stimulation frame represented by T-1;
- D′[J,T] represents a smoothed signal corresponding to D[J,T-1]; and
- α represents a user definable constant that controls an integration time associated with the smoothed signal.
9. The sound processor of claim 1, wherein the one or more stimulation frames comprise a single stimulation frame that immediately and temporally precedes the stimulation frame.
10. The sound processor of claim 1, wherein the one or more stimulation frames immediately precede the stimulation frame.
11. The sound processor of claim 1, wherein M is an integer equal to or greater than eight, and wherein N is an integer that is at least one less than M.
12. The sound processor of claim 1, wherein the channel selection facility directs a cochlear implant included in the cochlear implant system to apply electrical stimulation representative of signals included in the N analysis channels to the patient during the stimulation frame.
13. The sound processor of claim 1, wherein the audio signal is representative of sound presented to the patient.
14. A system comprising:
- a cochlear implant implanted within a patient;
- a plurality of electrodes communicatively coupled to the cochlear implant and implanted within a cochlea of the patient;
- a sound processor communicatively coupled to the cochlear implant and that divides an audio signal into M analysis channels, selects only N analysis channels included in the M analysis channels for presentation to a patient during a stimulation frame, wherein N is less than M, increases a probability that a particular analysis channel included in the M analysis channels will be included in the N analysis channels selected for presentation to the patient during the stimulation frame if the particular analysis channel was not selected for presentation to the patient during one or more stimulation frames that temporally precede the stimulation frame, and decreases the probability that the particular analysis channel will be included in the N analysis channels selected for presentation to the patient during the stimulation frame if the particular analysis channel was selected for presentation to the patient during the one or more stimulation frames that temporally precede the stimulation frame.
15. The system of claim 14, wherein the sound processor directs the cochlear implant to apply electrical stimulation representative of signals included in the N analysis channels by way of the plurality of electrodes to the patient during the stimulation frame.
16. A method comprising:
- dividing, by a sound processor included in a cochlear implant system associated with a patient, an audio signal into M analysis channels;
- selecting, by the sound processor, only N analysis channels included in the M analysis channels for presentation to the patient during a stimulation frame, wherein N is less than M;
- increasing, by the sound processor, a probability that a particular analysis channel included in the M analysis channels will be included in the N analysis channels selected for presentation to the patient during the stimulation frame if the particular analysis channel was not selected for presentation to the patient during one or more stimulation frames that temporally precede the stimulation frame; and
- decreasing, by the sound processor, the probability that the particular analysis channel will be included in the N analysis channels selected for presentation to the patient during the stimulation frame if the particular analysis channel was selected for presentation to the patient during the one or more stimulation frames that temporally precede the stimulation frame.
17. The method of claim 16, wherein the selecting of the N analysis channels for presentation to the patient during the stimulation frame comprises:
- generating M modified selection signals having individual values and each corresponding to a different analysis channel included in the M analysis channels, the M modified selection signals including a modified selection signal corresponding to the particular analysis channel;
- identifying N highest values out of the individual values, the N highest values corresponding to a subset of modified selection signals included in the M modified selection signals, the subset of modified selection signals corresponding to a subset of analysis channels included in the M analysis channels; and
- designating the subset of analysis channels that correspond to the subset of modified selection signals as the N analysis channels that are selected for presentation to the patient during the stimulation frame.
18. The method of claim 17, further comprising:
- determining, by the sound processor, that the particular analysis channel was not selected for presentation to the patient during the one or more stimulation frames that temporally precede the stimulation frame; and
- increasing, by the sound processor in response to the determining that the particular analysis channel was not selected for presentation to the patient during the one or more stimulation frames that temporally precede the stimulation frame, the probability that the particular analysis channel will be included in the N analysis channels selected for presentation to the patient during the stimulation frame by increasing a value of the modified selection signal corresponding to the particular analysis channel prior to the identifying of the N highest values.
19. The method of claim 17, further comprising:
- determining, by the sound processor, that the particular analysis channel was selected for presentation to the patient during the one or more stimulation frames that temporally precede the stimulation frame; and
- decreasing, by the sound processor in response to the determining that the particular analysis channel was selected for presentation to the patient during the one or more stimulation frames that temporally precede the stimulation frame, the probability that the particular analysis channel will be included in the N analysis channels selected for presentation to the patient during the stimulation frame by decreasing a value of the modified selection signal corresponding to the particular analysis channel prior to the identifying of the N highest values.
20. The method of claim 16, further comprising directing, by the sound processor, a cochlear implant included in the cochlear implant system to apply electrical stimulation representative of signals included in the N analysis channels to the patient during the stimulation frame.
7130694 | October 31, 2006 | Voelkel |
7751900 | July 6, 2010 | Voelkel |
8000797 | August 16, 2011 | Sarpeshkar et al. |
20100198301 | August 5, 2010 | Smith |
20110106211 | May 5, 2011 | Litvak |
20110286618 | November 24, 2011 | Vandali et al. |
20140079226 | March 20, 2014 | Kludt et al. |
2542302 | April 2014 | EP |
- Vandali, et al., “Speech Perception as a Function of Electrical Stimulation Rate: Using the Nucleus 24 Cochlear Implant System”, Ear & Hearing, Dec. 2000, vol. 21, No. 6, 608-624.
- International Search Report and Written Opinion received in International Application No. PCT/US14/059339, dated Jan. 9, 2015.
Type: Grant
Filed: Oct 6, 2014
Date of Patent: Jul 24, 2018
Patent Publication Number: 20170246457
Assignee: Advanced Bionics AG (Staefa)
Inventor: Leonid M. Litvak (Los Angeles, CA)
Primary Examiner: Brian T Gedeon
Application Number: 15/516,020
International Classification: A61N 1/36 (20060101); A61N 1/05 (20060101);