Anda di halaman 1dari 27

Accessibility Information and TipsRevised Date: 07/2011

Back
1 article(s) will be saved.
To continue, in Internet Explorer, select FILE then SAVE AS from your browser's toolbar above.
Be sure to save as a plain text file (.txt) or a 'Web Page, HTML only' file (.html). In FireFox,
select FILE then SAVE FILE AS from your browser's toolbar above. In Chrome, select right click
(with your mouse) on this page and select SAVE AS

EBSCO Publishing

Citation Format: APA (American Psychological Assoc.):

Review

NOTE:

the

instructions

at

http://support.ebsco.com/help/?

int=ehost&lang=&feature_id=APA and make any necessary corrections before using. Pay


special attention to personal names, capitalization, and dates.Always consult your library
resources

for

the

exact

formatting

and

punctuation

guidelines.

References
Dillon, H., & Keidser, G. (2012). Siemens Expert Series: NAL-NL2 - Principles, Background
Data, and Comparison to Other Procedures. Audiology Online, 1.
<!--Additional
Persistent

link

Information:
to

this

record

(Permalink):

http://search.ebscohost.com/login.aspx?

direct=true&db=a9h&AN=85096349&site=ehost-live
End of citation-->

Siemens Expert Series: NAL-NL2 - Principles, Background Data, and Comparison to


Other

Procedures

October 22, 2012


Editor's Note: This text course is an edited transcript of a live seminar on AudiologyOnline,
presented in partnership with Siemens Hearing Instruments, Inc. Download supplemental
course materials here.

Dr. Harvey Dillon: It is a pleasure to be talking to you all, especially without the 20-hour flight to
get from Australia to the U.S. Let's get right into our topic, the NAL-NL2 prescription.
Before getting into the intricacies of the prescription, I want to reassure you that using NAL-NL2
is as simple as using any of the previous National Acoustic Laboratories (NAL) prescriptions.
For an adult hearing aid fitting, the clinical process won't change with this new prescription. That
process involves measuring hearing thresholds and entering those thresholds into whatever
software you are using. Usually the hearing aid manufacturer's software has the NAL-NL2
algorithm incorporated, and then we verify that with a real-ear measurement and adjust the
amplification to match the prescription.
For children, we recommend a variation to the procedure where the thresholds are measured or
estimated on the basis of electrophysiology. Individual real-ear-to-coupler differences (RECD)
are either measured or estimated on the basis of age. Those two sets of information are entered
into the manufacturer's software; or alternatively into standalone software incorporating NALNL2, also. The result of that is a coupler prescription, to which a hearing aid is adjusted to
match. The real-ear effects have already been taken into account via the RECD.
The goals of the NAL-NL2 are to make speech as intelligible as possible and to make loudness
comfortable. We know very well that people want additional things from their hearing aids - such
as localization, a nice tonal quality, the ability to detect sounds around them, and sound
naturalness - and there are many variables that affect these qualities. The trouble is, we do not
really know how to quantify them, and so we focused on intelligibility and loudness, which we
suspect may be more important than the other factors.
The main question posed is, "How much amplification?" Hearing aids are very sophisticated,
and they make very complex decisions about how much gain to apply when and where. The end
result of all of that is amplification at each frequency. That is what the NAL-NL2 prescription is
aimed at specifying.
I want to make an analogy with a lolly shop, where a child goes into the lolly shop with an
allotted amount of money and wants to come out with the best possible taste experience. The
child may have a favorite lolly, but why spend all his money on that one lolly when he can get a
broader experience by having a range of sweets? Consider that not all the sweets are equally
good, so the child may well spend more of his money on the favorite lolly than on any other lolly.
It is the same with gain and frequency response. We have a budget, which in our case is a
loudness budget. If we make the hearing aid too loud, people will not like the hearing aid. We
have to spend that loudness budget by applying gain at different frequencies with different
amounts. We want to spend the budget most wisely, namely concentrating on those frequencies
where we can make the biggest contribution to intelligibility.

Developing NAL-NL2
The goal in developing NAL-NL2 was specifically to maximize speech intelligibility, but keep the
loudness less than or equal to normal loudness. That is the loudness that would be perceived by
a normal hearing person listening to the same sound. NAL-NL1 was actually based on the same
goals, but in the decade or more since then, we have had a number of empirical studies, mostly
led by Dr. Keidser, which have given us information about how well NAL-NL1 does or does not
achieve its aims. Believe me, we know more about the deficiencies in NAL-NL1 than anybody
else, and that has provided useful data into guiding the derivation of NAL-NL2. We also had
some specially-designed experiments in order to find out more about how much information
hearing-impaired people can extract from speech at different levels for different degrees of
hearing loss.
Here is how the basic derivation works. Start with a speech spectrum and level. Just pick one.
Let us choose 65 dB speech, for instance, and the spectrum that normally goes with it. Assume
some arbitrary gain frequency response, and that will then tell us what the amplified speech
spectrum is. If we put that into a model for predicting intelligibility, the result is the intelligibility
achieved for that particular gain-frequency response. If we then make an iterative adjustment to
the gain frequency response, we can keep adjusting it in the direction that increases the
predicted intelligibility according to the model. In other words, we have a calculation loop. So we
are traveling around this loop, changing the gain each time at each frequency in order to
maximize the intelligibility.
You can guess what would happen if this is all we did. We would end up with a lot of gain at all
frequencies, and everything would come out sounding loud, no matter what the original speech
signal was. Doing that by itself is not enough. We also need to calculate the loudness that is a
consequence of each gain-frequency response. That predicts what the loudness would be for
the hearing-impaired person.
So how does it compare to the loudness that a normal-hearing person would perceive? To
answer this, we also apply the loudness model to the unamplified speech spectrum to predict
what the normal loudness would be. We can then compare the normal-hearing loudness to the
hearing-impaired loudness. If it is too loud, then we simply turn the gain down at all frequencies.
We have two loops, one where the individual gains and individual frequencies are being turned
up or down to maximize intelligibility, and one which is turning the overall gain down so that the
overall loudness is equal to, or no more than, normal-hearing loudness. Both loops work
simultaneously. Eventually, this process converges on a solution, and what we then have is the
predicted gain frequency response for a particular audiogram and for a particular speech input
level and spectrum.

We repeated that process for a range of many audiograms. It was meant to cover a wide range
of possible audiograms and a wide range of input levels. We end up with a prescribed optimum
gain for each of those combinations (Figure 1).

: Figure 1.
Process for deriving gain frequency responses for NAL-NL2.
That process, of course, is of no use to an individual clinician. That is just the background
process that we went through, which produced these set of optimum gain frequency responses.
What we then have to do is to bring it into a formula that is easy to implement in software.
That is not all we did. The derivation I have so far described is all very theoretical. The
intelligibility model is based on real data from real people, but it is still a theoretical prediction
based on psychoacoustics and speech science, along with the assumptions of maximizing
intelligibility and controlling loudness. We need to be able to build in the empirical observations
that we have collected over the last decade, and even the ones going back before then.
If there is a disagreement, we need to make an adjustment to the assumptions or the rationale
so that we do end up with a prescription that agrees with what we have empirically found people
prefer and do best with. That is the process we have used here. We will talk more about the
adjustments later in this talk.
The result of all that is a final formula. In this case, we have produced the final formula with
something called a neural network. A neural network is where you put inputs in. In this case, the
inputs were the audiogram values and the overall sound pressure level of speech. We know
what the right answers are for our examples, because we chose the audiograms and the sound
pressure levels, so we know what gains came out of the optimizer.
Because we have the right answers for many audiograms, each combined with several sound
levels, that means we can train the neural network. Training the neural network will then produce
the gains for any other audiogram sound pressure level that is within the broad range of those
used for training. The end result of our derivation is actually the coefficients of a neural network,

which actually just represents a complicated formula. In other words, we end up with a formula
that summarizes that derivation process that I have talked about. The two key ingredients of the
derivation process are the loudness model and the speech intelligibility model.
The Loudness Model
We have not changed anything in the loudness model. We have entirely adopted the loudness
model of Moore and Glasberg (2004). It is a sophisticated model, and we are happy to be able
to use it. The intelligibility model is our own variation of the standard speech intelligibility index
(SII), also known as articulation index, which I will also talk about.
The basic rationale for predicting intelligibility is based on the speech spectrum. Figure 2 is an
illustration, characterized by a 30 dB dynamic range on a graph of 1/3 octave SPL versus
frequency. We then compare that to the person's threshold. Anything above the threshold is
audible, and anything below the threshold is not audible. If there is noise present, then we also
need to only take into account the parts of the speech spectrum that are above the noise and
the threshold. Only the shaded region is audible, so that is the proportion of the information that
is coming through (Figure 2). We can quantify that by the number of decibels that are audible at
each 1/3 octave frequency.

: Figure 2.
Predicting intelligibility based on audibility (shaded region) above hearing threshold (yellow) and
noise (pink).
Of course, not all frequencies are equally important. Just like in a lolly shop, not all the lollies are
equally desirable. We know that the very low and the very high frequencies do not contribute as
much to intelligibility as the mid frequencies. So what we can do is multiply each of those
audibilities times the importance of each frequency, and that gives a contribution to intelligibility
of each frequency. These can then be summed. In this particular example (Figure 2), the
calculation comes to 0.3, or 30%. Roughly speaking, ignoring the variation of importance with
frequency, we can say that about 30% of that speech spectrum is audible to the person in this
example. That is the standard SII calculation method, summarized by this formula:

SII = Ai Ii at each frequency, which is characterized by i.


There is a third twist to that, and that is that if you make speech too loud, intelligibility, even for
normal hearers, actually goes down because there is more spread of masking in the cochlea.
There is a level-distortion factor that is not represented in that formula. Here is how that
happens. As the speech level rises above 73 dB, the SII model tells us that the level-distortion
factor decreases below 1. In other words, you get smaller contributions to intelligibility the louder
you make it.
That is a bit of a problem with hearing-impaired people, because if we do not make it louder,
then we do not get audibility. So right in there is the tradeoff that we are making with hearingimpaired people. We want gain for audibility, but not too much gain, or we start to degrade
intelligibility, as well as making the sound unnatural to them. If a soft sound is amplified to be
something that sounds very loud, then it will sound out of place to people.
So what do we do with that SII number? We can turn that into percent correct using something
called the transfer function. Depending on how difficult the speech material is, there are different
transfer functions. If we are dealing with sentences, an SII of 0.3, for instance, might give us
80% correct, whereas with nonsense syllables, an SII of 0.3 might only give us 35% correct.
How well does this calculation method work? For people with normal hearing, it works pretty
well. For people with mild hearing loss, it also works well. By the time we get to severe hearing
loss, it does not work that well.
Ching, Dillon & Byrne (1998) measured speech intelligibility at different sensation levels and
predicted intelligibility from the model (Figure 3). For mild flat losses and mild sloping losses, the
prediction is reasonably close to the measured data. Once we get into the severe range, the
predictions always over-predict the amount of intelligibility listeners will receive. Because of
distortions in a damaged cochlea, people with hearing impairment experience reduced
frequency selectivity and reduced temporal selectivity. They cannot extract as much information
as a normal-hearing person can, even when it is audible.

: Figure 3. Sensation levels for


individuals with mild flat (MF), mild sloping (MS), severe flat (SF) and severe sloping (SS)
hearing losses. The blue dots correspond to percent-correct as the sensation level of the stimuli
increases.
We experimented with 20 normal-hearing subjects and 55 with a wide range of hearing losses to
find out how much information people can extract from speech at different frequencies when
amplified by different amounts. We ended up with several measured results and some predicted
curves, which I will not go into. There is large scatter, and there are reasons for that, but what
we are interested in is how deficient a person is compared with the expected result.
Figure 4 shows the transfer function, the yellow curve showing the relationship between
expected percent correct and SII. We can get this transfer function for normal listeners. The
expected result for the hearing-impaired population will be worse than the normal-hearing
function. The question is, "How much worse?"

: Figure 4.
Transfer function indicating the intelligibility deficit faced by listeners with hearing impairment.
This example shows that an individual scores 60% correct with an SII of 0.6, whereas a normalhearing individual would score 60% correct at an SII of 0.4.
Suppose this is a particular result for a particularly hearing-impaired person listening to speech
filtered and amplified in a particular way. There is a deficit, because the point is below the line
(Figure 4). We can quantify that deficit not by a low percent-correct score, but by how much
information it seems like the person is getting. They are getting 60% correct, and a normalhearing person would get 60% correct with an SII of only 0.4, or with only 40% of the information
available. This hearing-impaired person has physically had available 60% of the information,
and they also got 60% correct (which is a coincidence). However, there is a deficit in SII of 0.2
that we need to somehow explain on the basis of their hearing loss.
We calculated those deficits using nonsense syllables and sentences. There is a reasonable
correlation between the two, so we are measuring something intrinsic about the person's ability
to extract information from speech, whether that speech is nonsense syllables or sentences.
What do we do with that information? We have to somehow modify the SII model to incorporate
this reduced ability to extract information. We call this desensitization, which is not a great word,
but that is the word that seems to have been adopted.
As you increase the sensation level from 0 to 30 dB, the amount of information extracted from
speech goes from 0 to 1 (i.e. 100%) for a person with normal hearing. It appears that hearingimpaired people cannot extract the same amount of information, so we need a different curve
that rises to a different maximum that is less than 1. Our data have indicated for low-sensation

levels just above 0 that hearing-impaired people are pretty good at extracting information, almost
the same as normal hearers. As you go on increasing the sensation level, normal hearers keep
on getting more and more information out; whereas, hearing-impaired people are less able to
extract more information.
How much less depends on how much hearing loss they have. We fitted a model to our speech
intelligibility data that lets us predict this parameter m, which is the maximum information that we
could extract, and also the parameter p, which controls how quickly they go from normal to nonnormal as the sensation level rises (Figure 5).

: Figure 5.
Intelligibility and audibility; the yellow line indicates a normal-hearing person's ability to
understand as sensation level increases; the white line indicates a hearing-impaired person's
ability to understand as sensation level increases, which is on a different trajectory than the
normal-hearing path.
We ended up with a very convincing story. Figure 6 shows the graph of m, or the amount of
information people could extract relative to a normal hearer for different hearing losses. What we
can see is that for normal and mild losses, people extract nearly all the information available to
them. By the time you get to a 66 dB hearing loss at any frequency, people with hearing loss can
only extract about half the information. In the profound hearing loss region, the proportion of
information extracted is quite low, at only 10 to 20%.

: Figure 6. Maximum
achievable information extracted as a function of hearing threshold. The same curve applies at
all frequencies, not just the frequencies shown in the legend.
What we found quite reassuring was when we applied this same analysis to sentence material
based on the Bamford-Kowal-Bench (BKB) and City University of New York (CUNY) tests, and
then to consonants in the form of the vowel-consonant-vowels (VCVs), we got very much the
same curves. We think this is, at least on average, a reasonable indication of how much
information hearing-impaired people can extract from speech. It is why we fit cochlear implants
to people with severe and profound hearing loss, because the ability to extract information, even
when we make speech as audible as possible, is so low when presented acoustically.
We ended up with a series of curves showing the effective audibility (Figure 7), meaning that as
the sensation level increases, the amount of sound that people can actually use increases.
However, the more hearing loss you have, the less it increases.

: Figure 7. Effective audibility


as sensation level increases. The colored lines indicate hearing level from 0dB (red) to 120dB
(green). Audibility decreases as hearing loss increases.
Psychoacoustics
It is a bit disappointing in a way that, after all these years, we are still prescribing hearing aids on
the basis of pure-tone thresholds. What aboutthe many psychoacoustic parameters that we
know are affected by hearing loss?
We have also measured outer hair cell function through click-evoked otoacoustic emissions. We
estimated frequency resolution from psychoacoustic tuning curves and from a test of dead
regions known as the TEN test that I will say more about soon. We have also measured
cognitive ability, and of course, age.
Psychoacoustic tuning curves give you an idea of how sharply tuned the cochlea is to signals at
a range of frequencies. If the signal is 1,000 Hz, we can see in Figure 8 that it is best masked by
a narrow band masker also tuned to 1,000 Hz. That is what we expect for a normal functioning
cochlea. As the masker moves away from the frequency, either up or down, then the masker
becomes less effective.

: Figure 8. Tuning curve for a


signal of 1,000 Hz in a normal cochlea.
For somebody who has a non-functioning cochlea at 4,000 Hz, the curve might look wide and
flattened (Figure 9). The signal is at 4,000 Hz, but the most effective masking frequency is at
3,000 Hz. In other words, it seems that the person is detecting this 4,000 Hz signal, yet not at
the 4,000 Hz place in the cochlea, but at the 3,000 Hz place in the cochlea. That is why a 3,000
Hz masker is actually slightly more effective than a 4,000 Hz masker. So the psychoacoustic
tuning curves tell us about the sharpness of the tuning, and they also tell us about the likelihood
of the person listening at places in the cochlea other than where the signal is, presumably
because the inner hair cells are not sufficiently functioning in order to send a good detection
signal to the brain at the original frequency in the cochlea.

: Figure 9. Tuning curve for a


4,000 Hz signal in a damaged cochlea.

There is, of course, a very specific test of dead regions from Moore and colleagues (2004)
called the TEN Test. There is some evidence from the same research group in Cambridge that
one needs to take this knowledge of dead regions into account in determining whether it is worth
amplifying at each frequency. There is also some conflicting evidence suggesting that hearing
thresholds are sufficient, and we wanted to investigate that.
The basilar membrane has an excitation pattern as a function of position along the cochlea.
What happens if we have a dead region of the cochlea (i.e. no functioning inner hair cells)? Let's
suppose we put a signal in at a frequency that produces a specific excitation pattern with a peak
at one place in the cochlea. Normally, that would get picked up most sensitively by nerves
connected to this area in the cochlea and the adjacent frequencies. But if this high-frequency
region of the cochlea is all dead, then the person is actually listening to parts of the cochlea that
are tuned to lower frequencies.
If we were to put a masking sound in for the person with normal hearing, the masking sound
would have no effect if the excitation from the masker is lower than the peak excitation from the
signal. For the person who is actually listening in a frequency range outside the target frequency
due to dead regions, the excitation from the signal where it is being detected is lower, so the
masking sound may now mask the tone, so the input level would have to be increased when the
broadband masking sound is applied. That is the basic concept: Apply a broadband masking
sound and then look at what the threshold of the tone is. If it is higher than you would normally
expect, it suggests that the person was listening elsewhere in the cochlea. The test is called the
'Threshold Equalizing Noise Test' or TEN Test, which is a very clever acronym, because the
criterion is that the threshold rises 10 dB above the masking level of the noise.
We have looked at a comparison between the TEN Test and the tuning curve test, and in many
cases, both tests agreed that the region of the cochlea was alive. In several cases, they agreed
that that region of the cochlea was dead, that is no functioning inner hair cells. But there were
some regions of disagreement, where one test said the region was alive and the other said it
was dead. Sometimes you will find that you cannot complete one or other of the tests. This is
because the masking sound gets too loud for the patient to stand it before it gets intense enough
to mask the signal. So both tests will produce some uncertain results from time to time. Maybe
we did not do the psychoacoustic tuning curves with the best possible stimuli. We are still
scratching our heads as to why we did not get a good agreement there.
One of the problems in this work is that everything is correlated with everything else. We found
significant correlations between the hearing thresholds, the sharpness of the tuning curves, the
outer hair cell functioning as evidence by otoacoustic emissions strength, the results of the TEN
Test, people's cognitive ability and their age. This applied both at 2 kHz and 4 kHz. Correlations
were also present at lower frequencies, but were less strong.

When all these factors are correlated, how do we know what factor is causing what? More
specifically, can we make a better prediction of intelligibility if we make use of the
psychoacoustic results?
The answer is, yes, we can predict intelligibility just a little bit better if we know about frequency
selectivity. In our judgment, the advantage was not sufficiently great to cause audiologists
around the world to be spending time measuring frequency selectivity; we did not think that it
contributed very much to the ability to prescribe a hearing aid. If the aim of the hearing aid is to
maximize intelligibility, and the psychoacoustics do not let us significantly increase our ability to
predict speech intelligibility, then we think it does not add much to the hearing aid fitting process.
Why are thresholds so important and related to everything else? We think the mechanism is like
this: as people get older, they develop hearing loss, and also their cognitive abilities decline (on
average). Reduced functioning of the outer hair cells reduces the sharpness of the tuning curve,
reduces the level of the otoacoustic emissions, and also cause elevation of the TEN Test results.
Degradation of the inner hair cells causes further, sometimes marked, elevation of the TEN
masked thresholds.
Hearing loss may or may not also produce a cognitive decline, as people are increasingly shut
off from those around them. That is a very interesting question that we cannot answer yet. Of
course, age by itself does not cause anything. Age causes cardiovascular problems and other
physiologic problems. They, in turn, cause variations inside the cochlea, such as the
degradation of outer hair cells and the stria vascularis, mechanical changes, a stiffer cochlea
and perhaps reduced cognition as well. These things then cause the observed changes in
tuning curve sharpness, otoacoustic emissions, TEN strength, and the straightforward
audiogram. We think that is the mechanism underlying it. The end result is that, when you know
about the hearing loss, you can then make predictions about these other parameters to a
sufficient degree to do a reasonable job of fitting a hearing aid.
Intervening factors are many: frequency selectivity, temporal resolution, central auditory
processing abilities, and other things that we probably have not even thought of yet. But at least
if we know the hearing thresholds, we can make some predictions about speech perception. Of
course, age is associated with degradation of all these other quantities.
I am now going to hand it over to Dr. Gitte Keidser to talk about some empirical findings.
Factors Affecting Prescription
Dr. Gitte Keidser: Hello, everyone. I am going to talk about the factors that will affect the NALNL2 prescription.

After the release of NAL-NL1, we received comments from clinicians around the world
suggesting that many clients would not accept the overall gain prescribed by NAL-NL1. To study
this further, we looked at the gain preferences by 187 adults who had taken part in various
research projects that we had conducted here in Australia (Keidser, Dillon, Carter & O'Brien, in
review). For each of these individuals, we had a real-ear insertion gain measurement of their
preferred or fine-tuned fitting for listening to normal conversational input levels.
Figure 10 shows the average selected overall gain relative to that prescribed by NAL-NL1. The
gain variations are shown separately for females and males, for new hearing aid users and
experienced hearing aid users with a mild, moderate, or severe hearing loss. Negative values
indicate that the subjects preferred less gain than prescribed. Across the board, our subjects
preferred 3 dB less gain than prescribed by NAL-NL1, which means that NAL-NL2 prescribes 3
dB less overall gain than NAL-NL1.

:
Figure 10. Preferred gain deviation from NAL-NL1 by new and experienced hearing aid users
with mild, moderate and severe to profound hearing losses.
You may have noticed that whether the hearing aid users were new or experienced, and
independent of what the degree of hearing loss was, the females tend to prefer less gain than
the male hearing aid users. The difference was, on average, 2 dB, and that difference is

statistically significant. Based on this finding, NAL-NL2 prescribes 2 dB higher gain for males
than for females if you enter the gender of your client into the fitting software.
Finally, we found that there was no difference in the gain preferences between new and
experienced hearing aid users with a mild hearing loss. However, new hearing aid users with a
moderate hearing loss did prefer significantly less gain than did experienced hearing aid users
with a moderate hearing loss. Therefore, we have introduced an effect of experience in NALNL2.
Experience was specifically investigated in one study that we did (Keidser, O'Brien, Carter,
McLelland & Yeend, 2008). We looked at the gain variation from the prescription by new hearing
aid users, and by experienced hearing aid users. We found that increasingly less gain was
preferred by new users as the degree of hearing loss increased. For the experienced hearing
aid users, the gain variation was relatively constant around minus 3 dB. So on this basis, NALNL2 does prescribe different gain adjustments for new hearing aid users, depending on the
degree of hearing loss.
In Figure 11, the green broken line shows the net difference in gain that is prescribed to new and
experienced hearing aid users as the degree of hearing loss increases. No difference in gain is
prescribed while the loss is very mild. We see an increased difference in prescribed gain across
the moderate hearing loss range, and then we settle on a constant gain difference as the
hearing loss becomes severe to profound. The reason for this is that we have no data to suggest
that we should vary the gain even more for those with profound hearing loss. It also seems
counterproductive to reduce gain further for those who need it the most.

: Figure 11. Adjustments to


prescription to allow for experience level.
So what does that mean? That must mean that new users adapt to gain over time. What is that
time period? We were able to follow 11 subjects with moderate hearing loss longitudinally

(Keidser et al., submitted). We saw them for a period of three years, roughly every six months.
At each appointment, we let them select their preferred program and the volume setting that
they commonly used, and then we measured the overall gain. On average, these subjects
increased their gain setting steadily over the first two years of hearing aid usage. After two
years, their preferred gain setting was within the 95% confidence intervals of the gain preferred
by experienced hearing aid users with a similar degree of hearing loss.
So far, we have talked about gain preferences for a medium input level, but what happens at low
and at high input levels? Two studies conducted independently of each other both used a
research device that logged information about the input level of the subjects' environment, as
well as the volume control settings that they selected in these environments (Smeds, Keidser,
Zakis, Dillon, Leijon, et al., 2006; Zakis, Dillon, McDermott, 2007). Both studies showed that the
subjects selected further gain reduction for an 80 dB input level, relative to the gain variation
from the prescription they selected at a 65-dB input, but less gain reduction for a 50-dB input
level. What that means is that they actually preferred higher compression ratios than those
prescribed by NAL-NL1.
We have established that adults prefer 3 dB less gain for a 65-dB input level, and that they
prefer a higher compression ratio than prescribed by NAL-NL1, which is achieved by reducing
gain further for high input levels, but not for low input levels.
Other studies (Scollie et al., 2010) have suggested that children tend to prefer higher gain levels
than adults for medium input levels. An increase in gain is more likely to lead to greater speech
intelligibility at low input levels where speech may be limited by audibility. It is also less likely to
cause a noise-induced hearing loss if introduced at low than at high input levels. Consequently,
we are prescribing children with the same higher compression ratio that is prescribed to adults.
For children, the higher compression ratio is achieved by increasing gain relative to the NAL-NL1
prescription, with more increase at low input levels, and not at all for high input levels.
An effect of age on prescribed overall gain has been introduced in NAL-NL2. The result did raise
the question whether the difference in the preferred amount of gain between children and adults
was due to age or whether it was because the hearing loss was congenital or acquired. We
addressed this question by looking at data we had collected on 43 adults who were all fitted with
the same type of hearing aids. Of these adults, 15 had a congenital hearing loss and 28 had an
acquired hearing loss. There was no significant difference in the gain that they selected relative
to the NAL-NL2 prescription across low or across high frequencies (Figure 12). So for now, we
assume that age is determining whether children or adults prefer different amounts of gain.

: Figure 12. Preferred gain


from subjects with congenital or acquired hearing loss for low and high frequencies. LFA= lowfrequency average; HFA= high-frequency average.
Compression
I would like to go back to compression for a moment. The optimization process that Harvey
talked about before does suggest that we should increase the compression ratio as the degree
of hearing loss increases. We think that that will work fine, as long as we fit our clients with slowacting compressors. However, many studies have demonstrated that high compression ratios
combined with fast time constants can have a deleterious effect on speech understanding. In
one study, we did find that people with severe and profound hearing loss, when fitted with fastacting compressors, actually preferred closer to linear amplification, more so at low than at high
frequencies (Keidser, Dillon, Dyrlund, Carter & Hartley, 2007). So in the derivation of NAL-NL2,
we did develop two different formulas, one for slow-acting and one for fast-acting compressors.
When we derived the formula for fast-acting compression, the gain from the optimization
process was limited according to a mathematic function.
To sum it up, NAL-NL2 does prescribe different compression for slow and fast compressors for
those with high degrees of hearing loss. There is no difference in the compression ratios
prescribed for those with mild or moderate hearing loss with the two different compressor
speeds.
Bilateral Loudness Correction
Bilateral loudness correction was introduced in NAL-NL1 to take bilateral loudness summation
into account when we fit two hearing aids as opposed to one. Going back to the data on the 187
adults (see Figure 10), we had 136 subjects who were bilaterally fitted and 51 who were
unilaterally fitted. We found that those who were unilaterally fitted selected significantly lower

gain than did those who were bilaterally fitted. This indicates that the bilateral loudness
correction in NAL-NL1 is too large. That is also supported by some newer data (Whilby,
Florentine, Wagner & Marozeau, 2006; Epstein & Florentine, 2009).
In NAL-NL1, people who were unilaterally fitted were fitted with 3 dB more gain across the lower
input levels, with an increase in gain up to about 8 dB for higher input levels. In NAL-NL2, this
bilateral correction has been reduced to just 2 dB across the low input levels, increasing up to 6
dB for high input levels.
Effect of Language
The optimized gain at each frequency depends on the importance of that frequency to
understanding speech. In tonal languages that are commonly found across the Asian and
African countries, low frequencies are more important than they are in the English language.
Therefore, two versions of NAL-NL2 have been derived: one for tonal and one for non-tonal
languages. This was done by running the optimization process twice, one with an important
function that was applicable to tonal languages, and one with an important function that was
applicable to non-tonal languages.
The graph of Figure 13 shows an example of the difference in prescriptions for tonal languages,
the blue line, and non-tonal languages, shown by the red line. Immediately, the difference does
not seem to be great, but the effect of the difference is still to be verified in a clinical study.

: Figure 13. Comparison of


gain for tonal (blue line) and non-tonal (red line) languages using male talkers.
In summary, NAL-NL2 differs from NAL-NL1 by prescribing a different gain frequency response
and higher compression ratios, with the exception that we do prescribe lower compression ratios
for those with severe and profound hearing loss if they are fitted with fast-acting compressors.
Finally, the overall gain in NAL-NL2 depends on gender, age, experience with hearing aids, and
language as well.
Examples of Prescriptions

In Figure 14, the colored full lines show the NAL-NL2 prescription for 50 dB, 65 dB, and 80 dB
input levels; the broken line shows the NAL-NL1 prescription for a 65 dB input level. The
prescribed curves are for a moderate, gently sloping hearing loss, shown on the left of the
figure. As you can see, NAL-NL2 generally does prescribe a more smooth insertion gain curve.
Relative to NAL-NL1, it prescribes sightly more gain at low and high frequencies, and less gain
at mid frequencies.

: Figure 14.
Example audiogram of a moderate sloping hearing loss (left) and the NAL-NL2 prescriptions at
various inputs (color lines) compared with previous NAL-NL1 prescription (dashed line).
Similarly, for a flat hearing loss, NAL-NL2 has a more smooth prescription target and more gain
is prescribed at low and at high frequencies than at mid frequencies, relative to NAL-NL1 (Figure
15). Looking at a steeply sloping audiogram (Figure 16), the main difference is at the high
frequencies, but again, slightly less gain is prescribed at the mid frequencies. And for a ski-slope
hearing loss (Figure 17), we have a smoother target across the high frequencies with more highfrequency gain prescribed than NAL-NL1, while less gain is prescribed at the mid frequencies.
Finally, for a reverse sloping hearing loss (Figure 18), the main difference is in the low
frequencies, where NAL-NL2 prescribes more gain than NAL-NL1 and in the high frequencies
where NAL-NL2 prescribes less gain than NAL-NL1. Again, we have a more smooth insertion
gain target.

: Figure 15.
Comparison of NAL-NL2 to NAL-NL1 prescriptions for a flat moderate hearing loss.

: Figure 16.
Comparison of NAL-NL2 to NAL-NL1 prescriptions for a steeply-sloping hearing loss.

: Figure 17. Comparison of NAL-NL2 to NAL-NL1 prescriptions for a ski-slope hearing loss.

: Figure 18.
Comparison of NAL-NL2 to NAL-NL1 prescriptions for a reverse-slope hearing loss.
Dr. Harvey Dillon: Earl Johnson and I published a paper last year (Johnson & Dillon, 2011)
showing a comparison of NAL-NL2 relative to NAL-NL1, DSL m[i/o], and CAMEQ2-HF. We
applied a loudness model and also the speech intelligibility model to each of those prescriptions
and to seven different audiograms.
For a flat, moderate hearing loss, the different prescriptions look pretty similar in terms of how
much loudness they would prescribe. We start to see some differences, however, when we get
to more steeply sloping losses. Where hearing loss is most severe in the high frequencies, NALNL2 gave up more easily on providing gain than any of the other prescriptions. CAMEQ-HF
always tries to provide audibility at all frequencies. If it could talk, NAL-NL2 would have said that
high frequency gain produces diminishing returns for intelligibility while increasing loudness, so
it gave more emphasis to the mid and low frequencies instead.
That is the essential difference. In the reverse slope loss where the most severe loss is in the
low frequencies, NAL-NL2 tends to more give up than the other prescriptions. That is as a
consequence of that hearing loss desensitization being built into the derivation. In terms of
loudness, CAMEQ2-HF was pretty similar in loudness to NAL-NL1, and DSL and NAL-NL2 gave
lower overall loudness. We do not see very large differences between the four procedures for
the SII, or prediction of intelligibility.
Pediatrics
We have not said much about children, but we have a longitudinal study under way at the
moment led by Dr. Teresa Ching. We have collected data on about 500 children, looking at
everything that happens to them as far as amplification is concerned, as well as their outcomes
for language. The data is now complete up to age three, and almost complete at the age-five
data measurement point.

Included in this experiment was a randomized allocation of half the children to get the NAL
prescription, which, when we started, was NAL-NL1. We since have changed over to NAL-NL2,
and then the other half of the subjects got the DSL prescription, which again, when we started
was 4.1, and now is DSL m[i/o]. So far, we have found no difference in the language outcomes
between the NAL-NL2 and the DSL m[i/o] groups. There are still some differences in the
prescriptions, but as our groups have gone through successive iterations, the formulas have
come somewhat closer together. It appears that there are no consequences of the differences in
prescription for language outcomes of children, at least up to the age of five years.
I would like to mention just two things to finish. The first one is that insertion gain is the heart of
the NAL-NL2 prescription. The audiogram gives rise to the insertion gain. The insertion gain, as
Gitte said, is also affected by experience, the type of language, gender, age of the person,
speed of the compression, and whether the hearing loss is bilateral or unilateral. All of those
parameters let us calculate the insertion gain. From the insertion gain, we can then calculate the
real-ear aided gain, the coupler gain, the gain in an ear simulator, and it also influences the
maximum output of the hearing aid
Trainable Hearing Aids
There is a development which we think will take a little bit of pressure off measuring whether
gain-frequency responses have achieved the prescribed responses. If the hearing aid software
can do a reasonable job of getting close to the prescription, then the trainable hearing aid offers
the potential for the user to do the fine-tuning rather than the audiologist.
The hearing aid takes notice of the adjustments the wearer makes to the hearing aid,
remembers them and uses them to fine-tune the hearing aid. The most sophisticated trainable
hearing aids also measure and remember the acoustics of the situations in which those
adjustments were made and use those measurements to help interpret the adjustments. Over
time, the hearing aid can build up a picture of what the person likes in different environments,
and it can then make predictions about new environments.
For example, suppose a person adjusted the volume control and the hearing aid measured the
input level. After the person had been in six environments, we would have six observations of
how much gain the person liked for each of those input levels. The hearing aid could then do
some curve fitting (Figure 19). What does that curve tell us?

: Figure 19. Training gain,


compression ratio, and compression threshold for trainable hearing aids. The diamonds show
input levels of six different environments.
Well, the first thing it tells us is the compression threshold, or the point at which the gain starts to
decrease. In this case (Figure 19), that is 57 dB SPL. The second thing that it tells us is how
much gain the person wants below compression threshold. The third thing is how much the gain
reduces as the input level rises above compression threshold. That, of course, tells us what the
compression ratio is.
Just on the basis of the person adjusting the volume control, they have actually fine-tuned the
hearing aid's gain below compression threshold, the compression threshold itself, and
compression ratio. We think this is a good development for hearing aids, which lets people finetune hearing aids in their own environment, by doing nothing other than adjusting their hearing
aid in the usual way.
It is our challenge to develop the best prescriptive strategies and fit them the best way we know
how. So choose carefully, and thanks for listening.
Question & Answer
When might we see NAL-NL2 targets incorporated in probe mic systems?
The NAL-NL2 prescription has been licensed to most of the manufacturers of probe mic
systems. So very soon, if it is not in them already, is my answer.
How did maximum power output (MPO) change with the new prescription?
We have not changed our prescription much for MPO or saturation sound pressure level (SSPL)
from what it was with NAL-NL1. We prescribe maximum output of the hearing aid from hearing
thresholds. We did change some of the details of it, however. We think we allowed too
aggressively for multichannel hearing aids, so there is now a smaller difference between the

prescription for multichannel hearing aids and the prescription for when limiting is done on the
whole signal rather than separately within each channel. That was a small change. We are still
prescribing MPO based on threshold, not on measurement of discomfort level.
We did some careful research 12 or so years ago to try and answer the question, "Should we be
measuring loudness discomfort level when prescribing maximum output?" We came to a
convincing answer that, no, it did not add enough to the prescription in order to justify the time
taken clinically.
Should we be using RECDs for adults just like we use for infants?
There is no absolute right or wrong answer to this. You can prescribe a hearing aid on the basis
of insertion gain and verify it with the insertion gain. Or, you can prescribe a hearing aid on the
basis of real-ear aided gain, and then verify it with a real-ear aided gain measurement. Or, you
can measure the individual's RECD and infer what the real-ear aided gain should be. Both for
adults and infants, you can do it either way. We think just the balance of practicalities drives you
in different directions for the two.
In the case of infants, it is pretty hard to get the infant to sit still while you have a probe tube in
the ear without a hearing aid to hold the probe tube still. There is also a philosophical question
to consider. We know that infants will have a very high-frequency canal resonance at 6 kHz
when they are born. That progressively moves down to 2.7 kHz by the time they are five or six
years of age. Is that something that we need to preserve in a hearing aid fitting, or do babies
just have a high-frequency resonance in their ear canal simply because they have small ear
canals, which they have because they have small heads, which in turn they have because they
need to be born?
Well, I think the latter is probably right; in which case, there is no real point in recreating that
high-frequency canal resonance. If we use a real-ear aided gain approach rather than an
insertion gain approach, we will actually give them the same pressure at the eardrum as we
would for an adult. So the practicalities of difficulty in measuring insertion gain and the
philosophical argument of, "Do we really want to recreate that super-high frequency canal
resonance?" push us toward real-ear aided gain for the infant.
Then it becomes a logistics question, "Is it easier to measure with a probe tube after you have
fitted the hearing aid or do it before and measure the real-ear-to-coupler difference?" We think
measuring the real-ear-to-coupler difference is easier. When it is not possible, it is actually not
too unreasonable, I think, to predict the real-ear-to-coupler difference based on the age of the
child.

For adults, I think insertion gain is easier, and it is less prone to errors. With insertion gain, it is a
different measure. You only need one microphone, the probe microphone. You are not
comparing a probe to a coupler microphone. There is less potential for a calibration error in your
equipment to cause a problem.
So there is the long answer, long probably because there is no real right answer. The short
answer is that we recommend real-ear aided gain, achieved by a RECD measurement for
infants and insertion gain for adults.
References
Ching, T. Y. C., Dillon, H., & Byrne, D. (1998). Speech recognition of hearing-impaired listeners:
Predictions from audibility and the limited role of high-frequency amplification. Journal of the
Acoustical Society of America, 103(2), 1128-1140.
Epstein, M., & Florentine, M. (2009). Binaural loudness summation for speech and tones
presented via earphones and loudspeakers. Ear and Hearing, 30(2), 234-237.
Johnson, E., & Dillon, H. (2011). A comparison of gain for adults from generic hearing aid
prescriptive methods: Impacts on predicted loudness, frequency bandwidth, and speech
intelligibility. Journal of the American Academy of Audiology, 22, 441-459.
Keidser, G., Dillon, H., O'Brien, A., & Carter, L. (In review). NAL-NL2 empirical adjustments.
Trends in Amplification.
Keidser, G., O'Brien, A., Carter, L., McLelland, M., & Yeend, I. (2008). Variation in preferred gain
with experience for hearing aid users. International Journal of Audiology, 47(10), 621-635.
Keidser, G., Dillon, H., Dyrlund, O., Carter, L., & Hartley, D. (2007). Preferred low and high
frequency compression ratios among hearing aid users with moderately severe to profound
hearing loss. Journal of the American Academy of Audiology, 18(1), 17-33.
Moore, B.C.J. & Glasberg, B.R. (2004). A revised model of loudness perception applied to
cochlear hearing loss. Hearing Research, 188, 70-88.
Moore, B.C.J., Glasberg, B.R., & Stone, M.A. (2004). New version of the TEN test with
calibrations in dB HL. Ear and Hearing, 25, 478-487.
Scollie, S.D., Ching, T., Seewald, R., Dillon, H., Britton, L., Steinberg, J., & Corcoran, J. (2010).
Evaluation of the NAL-NL1 and DSL v4.1 prescriptions for children: Preference in real world use.
International Journal of Audiology, 49(S1), S49-S63.

Smeds, K., Keidser, G., Zakis, J., Dillon, H., Leijon, A., Grant, F., Convery, E., & Brew, C. (2006).
Preferred overall loudness II: Listening through hearing aids in field and laboratory tests.
International Journal of Audiology, 45(1), 12-25.
Whilby, S., Florentine, M., Wagner, E., & Marozeau, J. (2006). Monaural and binaural loudness
of 5- and 200-ms tones in normal and impaired hearing. Journal of the Acoustical Society of
America, 119, 3931-3939.
Zakis, J.A., Dillon, H., & McDermott, H.J. (2007). The design and evaluation of a hearing aid
with trainable amplification parameters. Ear and Hearing, 28(6), 812-830.
~~~~~~~~
By Harvey Dillon, PhD, Director of the National Acoustic Laboratories and Gitte Keidser, PhD,
Senior Research Scientist
Harvey Dillon is the Director of the National Acoustic Laboratories. He is involved in the
supervision of research into hearing assessment, procedures for rehabilitating hearing loss,
improved devices for assisting hearing loss, and hearing loss prevention.
Gitte Keidser is a Senior Research Scientist with the National Acoustic Laboratories where she
manages projects involving hearing devices and rehabilitation.
Copyright of Audiology Online is the property of Audiology Holdings, LLC and its content may
not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's
express written permission. However, users may print, download, or email articles for individual
use.

Back

Anda mungkin juga menyukai