Skip to main content

2013 | OriginalPaper | Buchkapitel

2. Methods for Automatic WSD

verfasst von : Oi Yee Kwong

Erschienen in: New Perspectives on Computational and Cognitive Strategies for Word Sense Disambiguation

Verlag: Springer New York

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Research in automatic word sense disambiguation has a long history on a par with computational linguistics itself. In this chapter, we take a two-dimensional approach to review the development and state of the art of the field, by the knowledge sources used for disambiguation on the one hand, and the algorithmic mechanisms with which the knowledge sources are actually deployed on the other. The trend for the latter is relatively clear, correlating closely with the historical development of many other natural language processing subtasks, where conventional knowledge-based methods gradually give way to scalable, corpus-based statistical and supervised methods. While the importance of multiple knowledge sources has been realised at the outset, their effective use in disambiguation systems has nevertheless been constrained by the notorious problem of “knowledge acquisition bottleneck” and is therefore very much dependent on the availability of suitable lexical resources.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Fußnoten
1
The subject is also known by other names, e.g. word sense discrimination (McRoy 1992; Schütze 1998), lexical ambiguity resolution (Hirst 1987), automatic sense disambiguation (Lesk 1986), sense tagging (Wilks and Stevenson 1997), sense clustering (Chen and Chang 1998), word sense classification or supersense tagging (Ciaramita and Johnson 2003), word sense induction (Navigli and Crisafulli 2010), etc. Stevenson (2003) distinguishes sense disambiguation and sense tagging as different levels of WSD. However, as he also pointed out, the delimitation is not always clear-cut.
 
2
Thus in the past, some only considered a reduced set of senses (e.g. Leacock et al. 1993; Bruce and Wiebe 1994; Leacock et al. 1998; Towell and Voorhees 1998), or even only two distinct senses (e.g. Brown et al. 1991; Gale et al. 1992a; Yarowsky 1995), usually because of data availability and the respective methods used. The recent general understanding is that homonymy is more reasonably and usefully handled by WSD systems, and fine-grained senses in existing resources should better be merged to give more distinct groups of senses (e.g. Ide and Wilks 2006; McCarthy 2006).
 
3
In fact, some of the knowledge sources like selectional restrictions and subcategorisation frames may have a mutually beneficial relationship with WSD. On the one hand, these knowledge sources are useful for word sense disambiguation; and on the other hand, the ability to disambiguate word senses is also found to help the acquisition of selectional restriction and verb subcategorisation patterns (e.g. McCarthy 1997; Korhonen and Preiss 2003).
 
4
See Landes et al. (1998) for the associated project on building semantic concordances.
 
6
Some of the methods discussed in this section make use of knowledge acquired from large corpora and therefore involve statistical techniques. Since the probabilistic models are applied to the acquisition of particular knowledge sources like selectional preferences and subcategorisation patterns, we group them under knowledge-based methods and distinguish them from other corpus-based WSD methods which are really trained on corpus examples with sense information. The latter will be considered supervised methods as discussed in the next section.
 
7
Wilks and Stevenson (1996) remarked that a 12-word sentence could give rise to more than 109 sense combinations to evaluate.
 
8
The human judgement scores were from Miller and Charles (1991).
 
9
Precision and recall are common performance measures in NLP. See Sect.​ 3.​2.​3 for their definitions.
 
10
See Gale et al. (1992a) for the “one sense per discourse” property in WSD.
 
11
Slightly different is Pedersen (2000), who used one type of information, namely co-occurring words, but combined evidence from various window sizes.
 
13
Actually Stevenson’s (2003) system is more appropriately considered a hybrid system since most of the individual modules have a very prominent knowledge-based element while the supervised learning part serves to conveniently combine them. For example, the three partial taggers work on word overlaps with dictionary definitions using simulated annealing, broad context based on subject areas of words indicated by the pragmatic codes in LDOCE, and selectional restriction information found in LDOCE as expressed by semantic codes in the dictionary, respectively.
 
14
SVM attempts to find a hyperplane with the largest margin that separates training examples into two classes. AdaBoost attempts to boost the performance of an ensemble of weak learners by giving more weights to misclassified training examples so that the classifier will concentrate on these hard examples.
 
15
The estimation was based on 3,200 most frequent words in the Brown Corpus which cover 90% of all word occurrences, and each should have 1,000 instances tagged.
 
16
Methods like this which use a few examples for bootstrapping are sometimes counted as weakly/lightly/minimally supervised approaches, in contrast to genuine unsupervised methods which take no tagged training examples at all.
 
Metadaten
Titel
Methods for Automatic WSD
verfasst von
Oi Yee Kwong
Copyright-Jahr
2013
Verlag
Springer New York
DOI
https://doi.org/10.1007/978-1-4614-1320-2_2

Neuer Inhalt