- the DiaSketch
Jakob Halskov
Dept. of Computational Linguistics
Copenhagen Business School
e-mail: jh.id@cbs.dk
Abstract
Identifying recurrent usage patterns of terms in non-specialized contexts may act as a filtering device
and thus help increase the precision of web-based term extraction algorithms. This article presents a
corpus-driven approach to the study of determinologization and investigates the claim in Melby [16]
that terms are always characterized by special reference irrespective of the context in which they are
used. An implementation of a system called the DiaSketch (Diachronic wordSketch) is outlined. The
DiaSketch can detect changing co-occurrence patterns of mother terms in diachronic corpora and thus
indirectly assess their termhood. Analyzing the usage of a term from the domain of Information
Technology (IT) in specialized and non-specialized contexts by means of DiaSketches, it is shown that
termhood seems to be a gradable property.
Introduction
39
Determinologization
This section discusses what determinologization is, why it should be studied and how one
might proceed to study it with statistical methods from corpus linguistics.
2.1
What is determinologization?
It is not surprising that conceptual fuzziness tends to occur when non-specialists use
terminology in non-specialized communicative contexts. It seems intuitive that what is a term
(representing a clear-cut concept) to one person may be a (possibly unknown) word
representing a fuzzy category to another person who lacks the required specialist knowledge
to decode the term fully and correctly. It also seems probable that traces of this conceptual
fuzziness can be registered in linguistic usage. While one-off cases of creative or fuzzy usage
of terms pose no problem to web-based ATR software, it is a different story when large
numbers of non-specialists use terms from a domain forming strong collocations which,
formally speaking, may resemble terminological neologisms, while not functioning as such.
Although determinologization has received little attention and has yet to be studied in a
quantitative framework, it has been defined as "the ways in which terminological usage and
meaning can 'loosen' when a term captures the interest of the general public" [17]p.12 and
"det at ei eksisterende terminologisk ordform gr over i allmennsprket" 1 [19]p.112. The
40
definition in Meyer and Mackintosh is the more specific of the two and groups the
semantic/pragmatic changes caused by determinologization into two types:
1) Maintien des aspects fondamentaux du sens terminologique
2) Dilution du sens terminologique d'origine [18]pp.202, 205
To illustrate the difference between 1) preservation and 2) dilution of a terminological sense
we can consider the two phrases from The New York Times (1999) below:
A large server
When the term server is modified by the conceptually fuzzy adjective large, it seems that the
reference of the combined phrase has somehow become less accurate. Are we talking about a
server, which is physically large, or about a server which is equipped with large amounts of
RAM? In spite of the superficial fuzziness, the reference to the domain specific concept2
seems to be largely intact. However, this is not the case with the figurative use of the term
reboot. When taking the context into consideration it becomes obvious that reboot no longer
refers to the original domain specific concept of shutting down and restarting an operating
system, but is being used in the more general sense of starting something afresh.
Clearly, there are a lot of intermediate stages in-between determinologization of type 1)
and 2), also known as sense modulation and sense selection in lexical semantics [6], and it is
not at all clear whether terms which are exposed to this phenomenon will come to represent
increasingly fuzzy categories or not. The rest of the article will outline an approach, which
may eventually answer this and other questions regarding determinologization.
2.2
a computer that controls or performs a particular job for all the computers in a network (definition from
MacMillan English Dictionary for Advanced Learners, 2002)
41
42
2.3
While the definition in section 2.1 provided a starting point for an empirical investigation of
the linguistic properties of determinologization, it will be necessary to explicate the theory of
meaning adopted in this study in order to arrive at an operational definition of conceptual
fuzziness and determinologization.
From a statistical NLP perspective
it is [...] natural to think of meaning as residing in the distribution of contexts over
which words and utterances are used [25]p.16
In a contextual theory of meaning syntax and semantics are intimately related and
interdependent. Every word has a certain semantic potential, but it is the context, ie.
neighbouring words (and sometimes extra-linguistic context), which activates a particular
sense in a particular case. This theory of meaning is highly pragmatic and descriptive and can
be summarized by the famous Wittgenstein aphorism: "The meaning of a word is its use in the
language" [27]. Based on a contextual theory of meaning, determinologization can be
described as the process by which the combinatory potential, ie. relational co-occurrence
patterns, of a term starts to resemble that of a comparable lexical unit from general language.
While conceptual fuzziness can be measured synchronically, we should not forget that
determinologization is controlled by an extra-linguistic factor, namely the degree of diffusion
of domain concepts into the consciousness of the general public. The spread of domainspecific concepts into non-specialized discourse is a diachronic phenomenon, and a
description of the linguistic properties of determinologization can thus only be attempted by
comparing a series of synchronic assessments of conceptual fuzziness. Such assessments are
performed by computing lexical profiles of terms in diachronic corpora.
Lexical profiling has primarily been used within lexicography, and an example of an
implementation for English is described in [14] and [15]. The basic technique involves the
calculation of the strength of association of key relational collocates of a word in a part of
speech tagged corpus and the subsequent listing of these collocates ordered by relation and
statistical salience. Figure 1 shows such a lexical profile, also known as a "wordsketch", of
the term server in the British National Corpus (BNC)3.
43
So far wordsketching has only been applied to synchronic studies of lexical units, but the
technique seems very promising for diachronic studies of how terminological units behave in
general language corpora. Retrieving significant collocational changes from wordsketch to
wordsketch in successive time slices of a general language corpus will yield what we might
call a DiaSketch (Diachronic wordSketch). Classifying the speed and manner in which
changes might register in such a DiaSketch will then help us gain a greater knowledge of the
linguistic properties of determinologization and perhaps allow us to refine the definition
proposed in section 2.1. Before proceeding to an actual test run of the DiaSketch
implementation in section 6, sections 3 and 4 will summarize the theoretical debate on the
issue of termhood.
This section discusses the notion of termhood as a prelude to the summary of theoretical
approaches to terminology in section 4.
44
superficial ambiguity [16]p.55, does not impede efficient and unambiguous specialized
communication between domain experts.
Ideally, a distinctive feature of terms, as opposed to words, is that the 1:1
correspondence between a term and the concept it labels is impervious to linguistic context,
and the meaning of a term can thus be fully decoded irrespective of the context in which it
was used. Words, on the other hand, represent fuzzy categories, which partly overlap with
adjacent categories, and for their meaning to be fully decoded, one normally needs to consider
both the linguistic and communicative contexts. In fact, statistical analysis of large amounts
of actual usage is needed to cluster the meanings of a polysemous word into a list of
predominant word senses, and such a list will necessarily be infinite and highly dynamic due
to the fundamental ambiguity [16]p.55 of general language. When we move beyond neat,
synchronic samples of expert-expert communication within a highly specialized subdomain,
the clear-cut line between terms and words starts to blur, however.
45
a lexical unit is by itself neither terminological nor general but [...] it is general by
default and acquires special or terminological meaning when this is activated by the
pragmatic characteristics of the discourse [my emphasis] [5]pp. 189-190
Sager has been arguing that the key to distinguishing terms from words is a theory of
reference [21] and a multidimensional model of knowledge space. Words, he argued, map
onto notions through general reference, while terms map onto concepts (more restricted
segments of knowledge space) through special reference. Twenty years later, in [23], he
reemphasizes that terms are basically the end products of a double evolutionary process of
abstraction and subsequent specification in natural language. The process starts with unique
reference (proper names), the abstraction of which is general reference (nouns), and ends with
special reference (terms).
To avoid misunderstanding and make it possible to enhance the knowledge of
mankind, Sager argues that notions (or general representations) must be refined into concepts
or bundles of judgment. From this perspective a term is then
the name given to a set of judgements considered pro tem as a unit representing a
scientifically defined concept [23]p.53
By pro tem Sager acknowledges the dynamic nature of natural language and the fact that
concepts
may
degenerate
into
notions
through
usage-governed
process
of
46
47
Theories of terminology
and
finally
approach
in
position
this
the
theoretical
framework.
4.1
importance that can be said about terminology is more appropriately said in the
context of linguistics, or information science or computational linguistics [22]p.1
other scholars criticize GTT for ignoring actual usage:
Wster developed a theory about what terminology should be [my emphasis] in order
to ensure unambiguous plurilingual communication and not about what terminology
actually is in its great variety and plurality [5]p.167
Harsher critics claim that:
Traditional Terminology confuses principles, i.e. objectives to be aimed at, with facts
which are the foundation of a science. By raising principles to the level of facts, it
converts wishes into reality [26]p.15
As long as one recognizes that the primary objectives of GTT are knowledge structuring and
standardization, rather than a descriptive account of terminological usage in various contexts,
I find this bias perfectly legitimate, however. The following sections will briefly review
alternative paradigms in terminology and finally explicate the theoretical foundations of the
present study on determinologization.
4.2
SocioCognitive Terminology
While the GTT approach must be counted among the positivist or objectivist theories of
science, Sociocognitive terminology, as advocated in Temmerman [26], is a hermeneutic or
experientialist theory. The premise in experientialism is that reality does not exist
independently of the perceiving subject. All knowledge comes from experience, and meaning
cannot be completely objectified because it always involves a subject and is perceived and
expressed through an inescapable filter (natural language). Inspired by recent findings in
Cognitive Science which suggest that there is no clear separation between general and
specialized knowledge, [26] thus claims that terms, more often than not, represent categories
(notions in Sagers terminology) which are as fuzzy and dynamic as those represented by
words.
49
Temmerman argues that clear-cut concepts, which are not prototypical to some extent,
are extremely rare outside of exact sciences like Mathematics and Chemistry [26]p.223. The
analytical (intensional) definitions used in GTT are thus often inadequate because
prototypical categories with gradable membership cannot be understood in a logical or
ontological structure. The core of Temmermans criticism of GTT is that it rejects the unity of
the linguistic sign, by dissociating form (the term) from content (the concept) and thus
reducing terms to context-independent labels for things.
In her Sociocognitive terminology Temmerman speaks of Units of Understanding
(UU), rather than of concepts. These UUs typically have prototype structure and are in
constant evolution. UUs can rarely be intensionally defined but should be interpreted by
means of templates of understanding which are composed of different modules of
information depending on the receiver and the context.
On the whole I agree with Temmerman that GTT needs to be extended in various
directions to achieve descriptive and explanatory adequacy, and I believe it is correct that
terms, in certain contexts, represent categories as fuzzy as those represented by words, but I
disagree with her that this is the general case. I think it requires a process of
determinologization, and this process is only initiated when the domain to which the term
belongs catches the interest of non-specialists. While the SocioCognitive approach to
terminology shakes the very foundations of classical terminology the Communicative Theory
of Terminology outlined in the next section is much more inclusive of GTT.
4.3
Cabr ([4],[5]) claims that the research object of terminology is not concepts, nor units of
understanding, but rather Terminological Units (TU).
At the core of the knowledge field of terminology we, therefore, find the terminological
unit seen as a polyhedron with three viewpoints: the cognitive (the concept), the
linguistic (the term) and the communicative (the situation) [5]p.187
While GTT accounts for one dimension of the terminological polyhedron, namely the
conceptual one, it fails to consider the other dimensions. This does not mean that GTT is
flawed, because TUs are such complex and multidimensional phenomena that they can hardly
be accessed on all fronts at once. It does mean, however, that GTT can only be an ancillary
50
component in a more comprehensive theory the outline of which has only recently manifested
itself. Cabr argues that:
it is impossible to account for the complexity of terminology within a single theory [...]
a number of integrated and complementary theories are required which deal with the
different facets of terms [3]pp.12-13
Although Sager already discussed the communicative dimension of terms in [22], Cabr
iterates his arguments and calls for a Communicative Theory of Terminology (CTT) in which
"each one of the three dimensions [the cognitive, linguistic and communicative], while being
inseparable in the terminological unit, permits a direct access to the object" [5]p.187
4.4 Socioterminology
Like Sociocognitive Terminology and CTT, Socioterminology, as outlined in Gambier ([9],
[10]) and Gaudin [11], also argues that GTT needs to be extended. Socioterminology is
basically a functionalist approach to terminology, which stipulates that we should include
contextual factors like language change and social practices in the study of terminology:
Un terme ne peut pas tre vu seulement par rapport un systme (adquation de la
dsignation, rattachement un rseau de notions): il est aussi voir dans son
fonctionemment, sur le terrain des contradictions socials. (Qui utilise quoi? Qui
innove? Comment et par qui les termes se diffusent-ils? Comment soprent les
rajustements terminologiques, les reformulations? Etc.)5 [10]p.320
Its focal point is thus linguistic reality, or terminological performance in Chomskian terms,
rather than terminological standardization or knowledge structuring as such:
En rupture avec les usages traditionnels: consultation dexperts, travaux sur les corpus
limits, ignorance de la dimension orale, une attitude plus linguistique la linguistique
A term cannot be viewed exclusively with respect to a system (the adequacy of the designation, inclusion in a
network of concepts): it should also be viewed wrt. its function in the field of social contradictions. (Who uses
what? Who innovates? How and by whom are terms spread? How are terms subjected to language change? Etc.)
51
tant essentiellement une science descriptive suppose que les termes soient tudis
dans leur dimension interactive et discursive6. [11]p.295
This admittedly simplified survey of GTT and three newer theoretical schools shows how the
frameworks of sociolinguistics, cognitive science and communication theory are being
applied to terminology to increase the explanatory adequacy of classical terminological
theory. The following section will describe the theoretical foundations of the DiaSketch
approach to determinologization.
4.5
A corpus-based description of how domain specific terms are used in general language is
faced with two obvious problems:
1. since terms are specific to a domain we must expect their frequency of occurrence to
be relatively low outside the domain in question.
Breaking with traditional approaches (consulting experts, working with limited corpora, ignoring the oral
dimension), a more linguistic approach (linguistics being essentially a descriptive science) presupposes that the
terms are studied in their interactive and discursive dimension.
52
2. when lexical units, which function as terms in specialized discourse (e.g. bus, server,
driver), occur in non-specialized contexts, the most frequent senses are likely to be the
non-specialized ones.
While the context of the domain allows us to presume monosemy, language outside Melby's
wall is rife with polysemy. Mother terms thus need to be disambiguated (sense tagged) before
any reliable DiaSketching can take place.
Table 1 lists twenty terms from the ANSDIT8 terminology compilation, which have
the highest average relative frequency in a 1.8M word fragment of a specialized corpus
(PcPlus) and a 64M word fragment of a newspaper corpus (New York Times). Not
surprisingly, the lemmas seem to
represent
very
superordinate
sense.
These
general-
5.1
tagging
can
be
53
his landmark paper from 1995, Yarowsky [29] proposes and evaluates a semi-unsupervised
WSD algorithm, which achieves accuracy rates rivaling those of the best supervised methods.
It does so by making two very simple, but powerful assumptions, namely that polysemy is
restricted by the fact that polysemous lexical units typically have one sense per discourse and
one sense per collocation. Thanks to these two assumptions, the algorithm only needs a few
seed collocates (sense indicators), and can then use the assumptions as "bridges" to new
contexts from which more (or better) sense indicators can be retrieved in an iterative fashion.
The basic steps in the Yarowsky algorithm are as follows:
1. identify all occurrences of the polysemous word and store the contexts
2. for each possible sense identify a seed word (fx. through vs. pop-up for window)
3. using these sense indicators, extract a seed set of manually disambiguated
occurrences (table 2)
4. for each collocation type in the seed set compute log(P(sensea|collocationi)/
P(senseb|collocationi)) <=> log((f(sensea)+0.5)/(f(senseb)+0.5))9
5. order the collocations by numerical log-likelihood ratio to get a decision list (table 3)
6. Apply the decision list on all contexts from step 1 to classify more occurrences
7. Optionally apply one-sense-per-discourse assumption to classify more occurrences
8. Iterate steps 4 - 8
9. stop when the decision list is unchanged. New data can now be sense tagged.
54
phys
phys
phys
phys
comp
comp
comp
...
This kind of (virtually) unsupervised approach will be used to sense tag all occurrences of
candidate mother terms in the final implementation of the DiaSketch, but even DiaSketches of
candidate mother terms which have not been semantically disambiguated may yield
interesting results as can be seen in the case study of section 6.
5.2
subsequently
phrase
chunked
with
10
55
5.3
measures implemented by Evert himself in the Utilities for Co-occurrence Statistics (UCS)14
toolkit. He favours relational co-occurrence over simple positional co-occurrence because the
former reduces noise from grammatically unrelated n-grams and leads to more meaningful
results [7]. The UCS tables generated by Evert's perl scripts are called frequency signatures
and contain the four values, joint frequency (O11), marginal frequencies (O12, O21) and sample
size (N). This corresponds to a classical four-celled contingency table like table 6, where O11
is the number of times partition and server co-occur in the sample (of size N) of all noun
bigrams in the corpus, O12 is the number of times
partition occurs with another noun in this sample
and O21 the number of times server occurs with
another noun.
Comparing the observed with the expected
frequencies (E11-E22) provides a measure of the
strength of association of the two lemmas. This
association measure can be based on a number of
statistical models, but in the case study in section 6
14
PcPlus
6M
2000-2004
low
Newsgroups
100Bn16
1981-present
low/moderate
LDC Gigaword
1.7Bn
1994-2002
high
56
5.4
The
original
wordsketches
as
vserver
O12
O22
E12 = (R1*C2)/N
E22 = (R2*C2)/N
will retrieve (virtually) all these relations in the given corpus slice.
By specifying that the PoS of the rightmost verb must be simple present (VVZ), simple past
(VVD), gerund, (VVG) or infinitive (VVP), we filter out passives (which have PoS=VVN)
from this set of SUBJ_OF relations. Moreover, by setting the matching strategy to "longest",
we make sure that we get the main verb of complex VPs like to help convince (cf. table 7). If
longer relative clauses intervene, however, the pattern will simply match the first main verb.
This can only be avoided by carrying out a computationally expensive full parsing.
The lemma pairs (ie. agency/convince, model/make etc.) are then extracted and piped
into UCS, yielding the N value mentioned in section 5.3. In case we want a sketch of the term
server, the general query is simply transformed to a specific query by substituting
[lemma="server"] for [pos="NN.*"]. Such a query will then provide all the O11, O12 and O21
values needed to complete the contingency table and compute the co-occurrence statistics.
17
A noun in plural or singular possibly followed by a relative pronoun, any number of VP chunk elements and
finally a mandatory full verb.
57
According to MacMillan English Dictionary for Advanced Learners (2002) the lexical unit
server has five senses:
1. a computer
2. player who starts to play
3. large spoon/fork/etc
4. sb who helps in church
5. sb who brings food
Judging by the collocations in the general language DiaSketch for server (figure 3), senses 2
through 5 seem to be virtually absent (except for the collocation altar server which is
indicative of sense 4). So in this case semantic tagging was not a critical issue. Defining what
counts as a significant collocation can be difficult, but in this case study we only include those
co-occurrences which defeat the Null Hypothesis at a significance level of p < 0.00000118 (or
log(p) > 6) and where the number of co-occurrences (O11) exceeds 2% of the marginal
frequency of the term in question (O21). While the absolute number of occurrences of the
lemma server in this relation are approximately the same in the two corpora (some 800 per
time slice), the relative frequencies of course differ tremendously.
While figure 3 charts the most significant modifiers of the term server through time in
a 6M word slice of the British computer magazine PcPlus, figure 4 lists the most significant
modifiers of the same term in a 915M word slice of the New York Times (NYT) corpus.
Strength of association, as measured in negative logarithmic p-values, is indicated along the
y-axis and the collocation candidates are listed along the x-axis. A striking difference between
the two figures is that the three noun modifiers network, Internet and computer are prominent
collocations in the newspaper corpus (throughout the timeframe) but do not occur in the
DiaSketch for the corpus of computer magazines. While network server and Internet server
are infrequent19 variants of the term web server (which is highly salient in both corpora),
computer server is a case of determinologized usage. In this compound the noun modifier is
18
Mail correspondence with Stefan Evert has led me to believe that this is the standard significance level for
collocation strength as computed by means of contingency tables.
19
http://scholar.google.com (May 4, 2005): computer server (680 hits), Internet server (3,320 hits), network
server (3,800 hits), web server (62,800 hits)
58
used as a kind of domain label to distinguish the IT sense of the head noun from other senses
possible outside this domain (for example the altar server). The expression is no more fuzzy
than server on its own, but the modifier would be redundant in specialist communication and
might cause noise in an ATR system.
The only example of a collocation referring to a fuzzy category is powerful server
which climbs above the strict threshold values in two time slices of the NYT corpus (Fall of
1994 and 1999). While the adjectival modifier in PcPlus (virtual) combines with the mother
term to form a clear-cut subordinate concept in a generic relation to server, the collocation of
powerful with server modulates the special reference of server so that the combined phrase no
longer refers to a clear-cut domain specific concept. What exactly is denoted by powerful
server? Does powerful refer to the storage capacity of the server as measured in gigabytes or
to its clock rate as measured in MHz or rather to the data transmission rate of its network
interface as measured in gigabit per second?
Figure 3 Modifiers of the term server in PcPlus (2000-2003)
logp)
association strength (-
2000
2001
2002
2003
l
n
ai
io
m cat
i
pl
ap
y
ox
pr
eb
w
ur
yo
P
FT
S
N
D
nt
fo
is
N
l
tra
n
ce
al
ri tu
v
collocates
59
60
Conclusion
Strictly speaking, a lexical unit which functions as a term in specialized communicative contexts
the list in table 1 could be a starting point
61
References
[1] Barry, John (1993) A. Technobabble MIT Press
[2] Baroni, Marco (2004) "BootCat: BootStrapping Corpora and terms from the web" In:
Proceedings of LREC 2004
[3] Cabr Castellv, Mara Teresa (1999) "Do We Need an Autonomous Theory of
Terms?" In: Terminology 5:1, pp. 5-19, John Benjamins
[4] Cabr Castellv, Mara Teresa (2000) "Elements for a theory of terminology: Towards
an alternative paradigm" In: Terminology 6:1, pp. 35-57, John Benjamins
[5] Cabr Castellv, Mara Teresa (2003) "Theories of terminology - their description,
prescription and explanation" In: Terminology 9:2, pp. 163-199, John Benjamins
[6] Cruse, D.A. (1986) Lexical Semantics Cambridge University Press
[7] Evert, Stefan; Brigitte Krenn (2003) "Computational approaches to collocations"
Introductory course at the European Summer School on Logic, Language and Information
(ESSLLI)
[8] Evert, Stefan (2004) The Statistics of Word Co-occurrences: Word Pairs and
Collocations PhD thesis, University of Stuttgart
[9] Gambier, Yves (1991) Travail et vocabulaire spcialiss: Prolgomnes une socioterminologie In: Meta, 36(1), pp. 8-15, Les Presses de l'Universit de Montral
[10] Gambier, Yves (1987) Problmes terminologiques des pluies acides: Pour une socioterminologie In: Meta, 32(3), pp. 314-320, Les Presses de l'Universit de Montral
[11] Gaudin, Franois (1993) Socioterminologie: du signe au sens, construction dun
champ, In: Meta, 38(2), pp. 293-301, Les Presses de l'Universit de Montral
[12] Jrvi, Outi (2001) "From Precise Terms to Fuzzy Words - from Bad to Worse in
Terminology Science? In: IITF Journal vol. 12, no. 1-2, pp. 85-88
[13] Kageura, Kyo (2002) The Dynamics of Terminology - a descriptive theory of term
formation and terminological growth John Benjamins
[14] Kilgarriff, Adam; David Tugwell (2001) "WORD SKETCH: Extraction and Display
of Significant Collocations for Lexicography" In: Proceedings of ACL 2001, Toulouse,
France, pp. 32-38
[15] Kilgarriff, Adam (2004) "The sketch engine" In: Proceedings of the 11th EuraLex
International Congress
[16] Melby, Alan K. (1995) The Possibility of Language John Benjamins
62
[17] Meyer, Ingrid; Kristen Mackintosh (2000) "When terms move into our everyday lives:
An overview of de-terminologization" In: Terminology vol. 6:1, pp. 111-138, John
Benjamins
[18] Meyer, Ingrid; Kristen Mackintosh (2000) "L'tirement du sens terminologique:
apercu du phnomne de la dterminologisation" In: Le Sens en Terminologie Ed. by Henri
Bjoint and Philippe Thoiron, Presses universitaires de Lyon, pp. 198-217
[19] Myking, Johan (2000) "Sosioterminologi - Ein modell for Norden?" In: I
Terminologins tjnst - Festskrift fr Heribert Picht p 60-rsdagen, University of Vaasa
[20] Pearson, Jennifer (1998) Terms in Context John Benjamins
[21] Sager, Juan C.; David Dungworth (1980) English Special Languages Brandsetter
Verlag, Wiesbaden
[22] Sager, Juan C. (1990) A practical course in terminology processing John Benjamins
[23] Sager, Juan C. (1999) "In search of a foundation: Towards a theory of the term" In:
Terminology 5:1, pp. 41-57, John Benjamins
[24] Schulze, Bruno M. (1994) Entwurf und implementierung eines anfragesystems fr
maschinelle textcorpora. Master's thesis, Institut fr maschinelle Sprachverarbeitung
(IMS), Stuttgart University
[25] Schtze, Hinrich; Christopher D. Manning (1999) Foundations of Statistical Natural
Language Processing MIT Press
[26] Temmerman, Rita (2000) Towards New Ways of Terminology Description. The
sociocognitive approach John Benjamins
[27] Wittgenstein, Ludwig (1997) Philosophical Investigations translated by G. E. M.
Anscombe Basil Blackwell, orig. 1953
[28] Wster, Eugen (1991) Einfhrung in die allgemeine Terminologielehre und
terminologische Lexikographie. Bonn: Romanistischer Verlag, 3. Auflage, orig. 1979
[29] Yarowsky, David (1995) "Unsupervised Word Sense Disambiguation Rivaling
Supervised Methods" In: Proceedings of ACL 33, pp. 189-196
63