Automatic Strategies for Sense Disambiguation
A broad category of disambiguation schemes is to start with a similarity metric that gives a score between every two synsets and seeks to maximize that score across all of the possible synsets.
When inquisitive readers encounter a word they don't know or a word that is used in a way that is unfamiliar to them, they look up the term in a dictionary. Confronted with multiple options for a term, they find that definition (or gloss) that best fits the context the word appeared in. Using this intuition, Mike Lesk developed an elegant scheme whereby the sense of a word that has the largest number of overlaps the other senses is chosen.
This work has been extended by Siddharth Patwardhan and Ted Pedersen to use context vectors for each of the words in a WordNet gloss to form a basis for a new similarity measure.
As the number of synsets increases and the
Hiding a Semantic Hierarchy in a Markov Model
Although Abney and Light focused on the problem of disambiguation in the context of selectional restrictions, their approach is very similar to our approach. They seek