Obviously, we can also use different strategies to combine a set of knowledge-based heuristics. For instance, all the heuristics described in the previous section can be applied in order passing to the next heuristic only the remaining ambiguity that previous heuristics were not able to solve.
In order to evaluate the performance of the knowledge-based heuristics previously defined, we used the SemCor collection Miller93semcor, in which all content words are annotated with the most appropriate WordNet sense.In this case, we used a window of fifteen nouns (seven context nouns before and after the target noun).
The results obtained for the specification marks method using the heuristics when applied one by one are shown in Table 2. This table shows the results for polysemous nouns only, and for polysemous and monosemous nouns combined.
The results obtained for the heuristics applied independently are shown in Table 3. As shown, all the heuristics perform differently, providing different precision/recall figures.
|
Another possibility is to combine all the heuristics using a majority voting schema Rigau1997. In this simple schema, each heuristic provides a vote, and the method selects the synset that obtains more votes. The results shown in Table 4 illustrate that when the heuristics are working independently, the method achieves a 39.1% recall for polysemous nouns (with full coverage), which represents an improvement of 8 percentual points over the method in which heuristics are applied in order (one by one).
|
We also show in Table 5 the results of our domain heuristic when applied on the English all-words task from SENSEVAL-2. In the table, the polysemy reduction caused by domain clustering can profitably help WSD. Since domains are coarser than synsets, word domain disambiguation (WDD) Magnini2000 can obtain better results than WSD. Our goal is to perform a preliminary domain disambiguation in order to provide an informed search-space reduction.
|