<?xml version="1.0"?><!DOCTYPE article SYSTEM "/project/take/software/searchbench_offline_processing/paperxml_generator/aclextractor/src/python/../resource/dtd/paperxml.dtd"><article><header><firstpageheader><page local="1"/><title>Role of Word Sense Disalnbiguation in Lexical Acquisition: Predicting Semantics from Syntactic Cues</title><author surname="Dorr" givenname="Bonnie J."><org  name="University of Maryland" country="USA" city="College Park"/></author><author surname="Jones" givenname="Doug"><org  name="University of Maryland" country="USA" city="College Park"/></author></firstpageheader><frontmatter><p>Role of Word Sense Disambiguation in Lexical Acquisition: Predicting Semantics from Syntactic Cues</p><p>Bonnie J. Dorr and Doug Jones Department of Computer Science and Institute for Advanced Computer Studies University of Maryland A.V. Williams Building College Park, MD 20742 {bonnie, jones}(3umiacs .umd.edu</p></frontmatter><abstract>This paper addresses the issue of word-sense ambiguity in extraction from machine-readable resources for the con­struction of large-scale knowledge sources. We describe two experiments: one which ignored word-sense distinctions, re­sulting in 6.3% accuracy for semantic classification of verbs based on (Levin, 1993); and one which exploited word-sense distinctions, resulting in 97.9% accuracy. These experiments were dual purpose: (1) to validate the central thesis of the work of (Levin, 1993), i.e., that verb semantics and syntactic behavior are predictably related; (2) to demonstrate that a 15-fold improvement can be achieved in deriving semantic information from syntactic cues if we first divide the syntac­tic cues into distinct groupings that correlate with different word senses. Finally, we show that we can provide effective acquisition techniques for novel word senses using a combi­nation of online sources. </abstract></header><body><section number="1" title="Introduction"><p>This paper addresses the issue of word-sense ambigu­ity in extraction from machine-readable resources for the construction of large-scale knowledge sources. We describe two experiments: one which ignored word-sense distinctions, resulting in 6.3% accuracy for seman­tic classification of verbs based on (Levin, 1993); and one which exploited word-sense distinctions, resulting in 97.9% accuracy. These experiments were dual pur­pose: (1) to validate the central thesis of the work of (Levin, 1993), i.e., that verb semantics and syntactic be­havior are predictably related; (2) to demonstrate that a 15-fold improvement can be achieved in deriving se­mantic information from syntactic cues if we first divide the syntactic cues into distinct groupings that correlate with different word senses. Finally, we show that we can provide effective acquisition techniques for novel word senses using a combination of online sources, in particular, Longman's Dictionary of Contemporary En­glish (LDOCE) (Procter, 1978), Levin's verb classifica­tion scheme (Levin, 1993), and WordNet (Miller, 1985). We have used these techniques to build a database of 10,000 English verb entries containing semantic infor­mation that we are currently porting into languages such as Arabic, Spanish, and Korean for multilingual NLP tasks such as foreign language tutoring and ma­chine translation.</p></section><section number="2" title="Automatic Lexical Acquisition for NLP Tasks"><p>As machine-readable resources (i.e., online dictionaries, thesauri, and other knowledge sources) become read­ily available to NLP researchers, automated acquisition has become increasingly more attractive. Several re­searchers have noted that the average time needed to construct a lexical entry can be as much as 30 min­utes (see, e.g., (Neff and McCord, 1990; Copestake et al., 1995; Walker and Amsler, 1986)). Given that we are aiming for large-scale lexicons of 20-60,000 words, automation of the acquisition process has become a ne­cessity.</p><p>Previous research in automatic acquisition focuses primarily on the use of statistical techniques, such as bilingual alignment (Church and Hanks, 1990; Kla-vans and Tzoukermann, 1996; Wu and Xia, 1995), or extraction of syntactic constructions from online dic­tionaries and corpora (Brent, 1993; Dorr, Garman, and Weinberg, 1995). Others who have taken a more knowledge-based (interlingual) approach (Lonsdale, Mi-tamura, and Nyberg, 1996) do not provide a means for systematically deriving the relation between sur­face syntactic structures and their underlying semantic representations. Those who have taken more argument structures into account, e.g., (Copestake et al., 1995), do not take full advantage of the systematic relation be­tween syntax and semantics during lexical acquisition.</p><p>We adopt the central thesis of Levin (1993), i.e., that the semantic class of a verb and its syntactic behav­ior are predictably related. We base our work on a correlation between semantic classes and patterns of grammar codes in the Longman's Dictionary of Con­temporary English (LDOCE) (Procter, 1978). While the LDOCE has been used previously in automatic ex­traction tasks (Alshawi, 1989; Farwell, Guthrie, and Wilks, 1993; Boguraev and Briscoe, 1989; Wilks et al., 1989; Wilks et al., 1990) these tasks are primarily con­cerned with the extraction of other types of informa­tion including syntactic phrase structure and broad ar­gument restrictions or with the derivation of semantic structures from definition analyses. The work of San-fdippo and Poznanski (1992) is more closely related to our approach in that it attempts to recover a syntactic-semantic relation from machine-readable dictionaries. However, they claim that the semantic classification of verbs based on standard machine-readable dictionaries (e.g., the LDOCE) is "a hopeless pursuit [since] stan­dard dictionaries arc simply not equipped to offer this kind of information with consistency and exhaustive-ness.<page local="2"/>"</p><doubt alpha="0.0" length="3" tooSmall="False" monospace="0.0">322</doubt><p>Others have also argued that the task of simplify­ing lexical entries on the basis of broad semantic class membership is complex and, perhaps, infeasiblc (see, e.g., Boguraev and Briscoe (1989)). However, a number of researchers (Fillmore, 1968; Grimshaw, 1990; Gru­ber, 1965; Guthrie et al., .1991; Hearst, 1991; Jackend-olF, 1983; Jackendoff, 1990; Levin, 1993; Pinker, 1989; Yarowsky, 1992) have demonstrated conclusively that there is a clear relationship between syntactic context and word senses; it is our aim to exploit this relationship for the acquisition of semantic lexicons.</p><p><b>3 Syntax-Semantics Relation: Verb Classification Based on Syntactic Behavior</b></p><p>The central thesis of (Levin, 1993) is that the seman­tics of a verb and its syntactic behavior are predictably related. As a demonstration that such predictable rela­tionships are not confined to an insignificant portion of the vocabulary, Levin surveys 4183 verbs, grouped into 191 semantic classes in Part Two of her book. The syn­tactic behavior of these classes is illustrated with 1668 example sentences, an average of 8 sentences per class.</p><p>Given the scope of Levin's work, it is not easy to verify the central thesis. To this end, we created a database of Levin's verb classes and example sentences from each class, and wrote a parser to extract basic syn­tactic patterns from the sentences.<footnote anchor="1"/> We then character­ized each semantic class by a set of syntactic patterns, which we call a <i>syntactic signature, </i>and used the re­sulting database as the basis of two experiments, both designed to to discover whether the syntactic signatures tell us anything about the meaning of the verbs.<footnote anchor="2"/> The first experiment, which we label <i>Verb-Based, </i>ignores word-sense distinctions by assigning one syntactic sig­nature to each verb, regardless of whether it occurred in multiple classes. The second experiment, which we label <i>Class-Based, </i>implicitly takes word-sense distinc­tions into account by considering each occurrence of a verb individually and assigning it a single syntactic sig­nature according to class membership.</p><p>The remainder of this section describes the assign­ment of signatures to semantic classes and the two ex­periments for determining the relation of syntactic in­formation to semantic classes. We will see that our clas­sification technique shows a 15-fold improvement in the experiment where we implicitly account for word-sense distinctions.</p><footnote label="1">Both the database and the parser are encoded in Quin-tus Prolog.</footnote><footnote label="2">The design of this experiment is inspired by the work of (Dubois and Saint-Dizier, 1995). In particular, we depart from the alternation-based data in (Levin, 1993), which is primarily binary in that sentences arc presented in pairs which constitute an alternation. Following Saint-Dizier's work, we construct N-ary syntactic characterizations. The choice is of no empirical consequence, but it simplifies the experiment by eliminating the problem of naming the syn­tactic patterns.</footnote><p>Verbs: break, chip, crack, crash, crush, fracture, rip, shatter, smash, snap, splinter, split, tear</p><p>Example Sentences:</p><p>Crystal vases break easily.</p><p>The hammer broke the window.</p><p>The window broke.</p><p>Tony broke her arm.</p><p>Tony broke his finger.</p><p>Tony broke the crystal vase.</p><p>Tony broke the cup against the wall.</p><p>Tony broke the glass to pieces.</p><p>Tony broke the piggy bank open.</p><p>Tony broke the window with a hammer.</p><p>Tony broke the window.</p><p>*Tony broke at the window.</p><p>*Tony broke herself on the arm.</p><p>*Tony broke himself.</p><p>*Tony broke the wall with the cup.</p><p>A break.</p><doubt alpha="58.6" length="251" tooSmall="False" monospace="0.0">Derived Syntactic Signature: l-[np,v] l-[np,v,np] 1-[np,v,np,adjective] l-[np,v,np,pp(against)] 1-[np,v,np,pp(to)] l-[np,v,np,pp(with)] l-[np,v,poss,np] l-[np,v,adv(easily)J l-[n] 0-[np,v,np,pp(with)] 0-[np,v,self] 0-[np,v,self,pp(on)] 0-[np,v,pp(at)]</doubt><table caption="Table 1: Syntactic Signature for Change of State -breaksubclass"></table><subsection number="3.1" title="Assignment of Signatures"><p>For the first experiment below, we construct a verb-based syntactic signature, while for the second experi­ment, we constructed a class-based signature.</p><p>The first step for constructing a signature is to decide what syntactic information to extract for the basic syntactic patterns that make up the signature. It turns out that a very simple strategy works well, namely, flat parses that contain lists of the major cat­egories in the sentence, the verb, and a handful of other elements. The "parse", then, for the sentence <b>Tony broke the crystal vase </b>is simply the syntac­tic pattern <b>[np.v.np]. </b>For <b>Tony broke the vase to pieces </b>we get <b>[np,v,np,pp(to)]</b>. Note that the <b>pp </b>node is marked with its head preposition. Table 1 shows an example class, the <i>break </i>subclass of the Change of State verbs (45.1), along with example sentences and the derived syntactic signature based on sentence pat­terns. Positive example sentences are denoted by the number 1 in the sentence patterns and negative example sentences are denoted by the number 0 (corresponding to sentences marked with a *).</p></subsection><subsection number="3.2" title="Experiment 1: Verb-based Approach"><p>In the first experiment, we ignored word sense distinc­tions and considered each verb only once, regardless of whether it occurred in multiple classes. In fact, 46% of the verbs appear more than once. In some cases, the verb appears to have a related sense even though it appears in different classes. For example, the verb <i>roll </i>appears in two subclasses of Manner of Motion Verbs that are distinguished on the basis of whether the gram­matical subject is animate or inanimate. In other cases, the verb may have (largely) unrelated senses. For ex­ample, the verb <i>move </i>is both a Manner of Motion verb and verb of Psychological State.<page local="3"/></p><doubt alpha="0.0" length="3" tooSmall="False" monospace="0.0">323</doubt><p>To compose the syntactic signatures for each verb, we collect all of the syntactic patterns associated with every class a particular verb appears in, regardless of the different classes are semantically related. A syntactic signature for a verb, by definition, is the union of the frames extracted from every example sentence for each verb. The outline of the verb-based experiment is as follows:</p><p>1. Automatically extract syntactic information from the example sentences.</p><p>2. Group the verbs according to their syntactic signature.</p></subsection></section><section number="3." title="Determine where the two ways of grouping verbs over­lap:"><p>(a) the semantic classification given by Levin.</p><p>(b) the syntactic classification based on the derived syntactic signatures.</p><p>To return to the Change of State verbs, we now con­sider the syntactic signature of the verb <i>break, </i>rather than the signature of the semantic class as a unit. The verb <i>break </i>belongs not only to the Change of State class, but also four other classes: 10.6 <i>Cheat, </i>23.2 <i>Split, </i>40.8.3 <i>Burt, </i>and 48.1.1 <i>Appear. </i>Each of these classes is characterized syntactically with a set of sentences. The union of the syntactic patterns corresponding to these sentences forms the syntactic signature for the verb. So although the signature for the Change of State class has 13 frames, the verb <i>break </i>has 39 frames from the other classes it appears in.</p><p>Conceptually, it is helpful to consider the difference between the <i>intension </i>of a function versus its <i>exten­sion. </i>In this case, we are interested in the functions that group the verbs syntactically and semantically. In-tensionally speaking, the definition of the function that groups verbs semantically would have something to do with the actual meaning of the verbs.<footnote anchor="3"/> Likewise, the in­tension of the function that groups verbs syntactically would be defined in terms of something strictly syntac­tic, such as subcategorization frames. But the inten­sions of these functions are matters of significant the­oretical investigation, and although much has been ac­complished in this area, the question of mapping syntax to semantics and vice versa is an open research topic. Therefore, we can turn to the <i>extensions </i>of the func­tions: the actual groupings of verbs, based on these two separate criteria. The semantic extensions are sets of verb tokens, and likewise, the syntactic extensions are sets of verb tokens. To the extent that these functions map between syntax and semantics intensionally, they will pick out the same verbs extensionally.</p><p>So for the verb-based experiment, our technique for establishing the relatedness between the syntactic signa­tures and the semantic classes, is mediated by the verbs themselves. We compare the two orthogonal groupings of the inventory of verbs: the semantic classes defined by Levin and the sets of verbs that correspond to each of the derived syntactic signatures. When these two groupings overlap, we have discovered a mapping from the syntax of the verbs to their semantics, via the verb tokens. More specifically, we define the overlap index as the number of overlapping verbs divided by the av­erage of the number of verbs in the semantic class and the number of verbs in the syntactic signature. Thus an overlap index of 1.00 is a complete overlap and an over­lap of 0 is completely disjoint. In this experiment, the sets of verbs with a high overlap index are of interest.</p><footnote label="3">An example of the intensional characterization of the Levin classes are the definitions of Lexical Conceptual Struc­tures which correspond to each of Levin's semantic classes. See (Dorr and Voss, to appear).</footnote><p>When we parsed the 1668 example sentences in Part Two of Levin's book (including the negative examples), these sentences reduce to 282 unique patterns. The 191 sets of sentences listed with each of the 191 semantic classes in turn reduces to 748 distinct syntactic signa­tures. Since there are far more syntactic signatures than the 191 semantic classes, it is clear that the mapping between signatures and semantic classes is not direct. Only 12 mappings have complete overlaps. That means 6.3% of the 191 semantic classes have a complete over­lap with a syntactic signature.</p><p>The results of this experiment are shown in Table 2. Three values are shown for each of the six variations in the experiment: the mean overlap, the median overlap, and the percentage of perfect overlaps (overlaps of value 1.00). In every case, the median is higher than the mean. Put another way, there is always a cluster of good overlaps, but the general tendency is to have fairly poor overlaps.</p><p>The six variations of the experiment are as follows. The first distinction is whether or not to count the neg­ative evidence. We note that the use of negative exam­ples, i.e., plausible uses of the verb in contexts which are disallowed, was a key component of this experi­ment. There are 1082 positive examples and 586 nega­tive examples. Although this evidence is useful, it is not available in dictionaries, corpora, or other convenient resources that could be used to extend Levin's classi­fication. Thus, to extend our approach to novel word senses (i.e., words not occurring in Levin), we would not be able to use negative evidence. For this reason, we felt it necessary to determine the importance of nega­tive evidence for building uniquely identifying syntactic signatures. As one might expect, throwing out the neg­ative evidence degrades the usefulness of the signatures across the board. The results which had the negative evidence are shown in the left-hand column of numbers in Table 2, and the results which had only positive evi­dence are shown in the right-hand side.</p><p>The second, three-way, distinction involves preposi­tions, and breaks the two previous distinctions involv­ing negative evidence into three sub-cases. Because we were interested in the role of prepositions in the sig­natures, we also ran the experiment with two dilferent parse types: ones that ignored the actual prepositions in the pp's, and ones that ignored all information except for the values of the prepositions. Interestingly, we still got useful results with these impoverished parses, al­though fewer semantic classes had uniquely-identifying syntactic signatures under these conditions. These re­sults are shown in the three major rows of Table 2.</p><p>The best result, using both positive and negative ev­idence to identify semantic classes, gives 6.3% of the verbs having perfect overlaps relating semantic classes to syntactic signatures. See Table 2 for the full results.</p><subsection number="3.3" title="Experiment 2: Class-based Approach"><p>In this experiment, we attempt to discover whether each class-based syntactic signature uniquely identifies a sin<b><u>Verb-based Experiment (No Disambiguation)</u> Class-based Experiment (Disambiguated Verbs)</b> gle semantic class.<page local="4"/> By focusing on the classes, the verbs are implicitly disambiguated: the word sense is by def­inition the sense of the verb as a member of a given class. To compare these signatures with the previous verb-based signatures, it may be helpful to note that a verb-based signature is the union of all of the class-based signatures of the semantic classes that the verb appears in.</p><doubt alpha="0.0" length="3" tooSmall="False" monospace="0.0">324</doubt><p>The outline for this class-based experiment is as fol­lows:</p><p>1. Automatically extract syntactic information from the example sentences to yield the syntactic signature for the class.</p><p>2. Determine which semantic classes have uniquely-identifying syntactic signatures.</p><p>If we use the class-based syntactic signatures contain­ing preposition-marked pp's and both positive and neg­ative evidence, the 1668 example sentences reduce to 282 syntactic patterns, .just as before. But now there are 189 class-based syntactic signatures, as compared with 748 verb-based signatures from before. 187 of them uniquely identify a semantic class, meaning that 97.9% of the classes have uniquely identifying syntactic signa­tures. Four of the semantic classes do not have enough syntactic information to distinguish them uniquely.<footnote anchor="4"/></p><p>Although the effects of the various distinctions were present in the verb-based experiment, these effects are much clearer in the class-based experiments. The effects of negative and positive evidence, as well as the three ways of handling prepositions show up much clearer here, as is clear in Table 4.</p><p>In the class-based experiment, we counted the per­centage of semantic classes that had uniquely identi­fying signatures. In the verb-based experiment, we counted the number of perfect overlaps (i.e., index of 1.00) between the verbs as grouped in the semantic classes and grouped by syntactic signature, 'the over­all results of the suite of experiments, illustrating the role of disambiguation, negative evidence, and preposi­tions, is shown in Table 4. There were three ways of treating prepositions: (i) mark the pp with the prepo­sition, (ii) ignore the preposition, and (iii) keep only the prepositions. For these different strategies, we see the percentage of perfect overlaps, as well as both the median and mean overlap ratios for each experiment. These data show that the most important factor in the experiments is word-sense disambiguation.</p><footnote label="4">Two of these classes correspond to one of the two non-unique signatures, and two correspond to the other non-unique signature.</footnote></subsection></section><section number="4" title="Semantic Classification of Novel Words"><p>As we saw above, word sense disambiguation is critical to the success of any lexical acquisition algorithm. The Levin-based verbs arc already disambiguated by virtue of their membership in différent classes. The difficulty, then, is to disambiguate and classify verbs that do not occur in Levin. Our current direction is to make use of the results of the first two experiments, i.e., the re­lation between syntactic patterns and semantic classes, but to use two additional techniques for disambiguation and classification of non-Levin verbs: (1) extraction of synonym sets provided in WordNet (Miller, 1985), an online lexical database containing thesaurus-like rela­tions such as synonymy; and (2) selection of appropri­ate synonyms based on correlations between syntactic information in Longman's Dictionary of Contemporary English (LDOCE) (Procter, 1978) and semantic classes in Levin. The basic idea is to first determine the most likely candidates for semantic classification of a verb by examining the verb's synonym sets, many of which in­tersect directly with the verbs classified by Levin. The "closest" synonyms are then selected from these sets by comparing the LDOCE grammar codes of the unknown word with those associated with each synonym candi­date. The use of LDOCE as a syntactic filter on the semantics derived from WordNet is the key to resolv­ing word-sense ambiguity during the acquisition pro­cess. The full acquisition algorithm is as follows:</p><doubt alpha="0.0" length="3" tooSmall="False" monospace="0.0">325</doubt><table caption="Table 2: Verb-based Results Table 3: Class-Based Results" class="main" frame="box" rules="all" border="1" regular="False"><tr class="row"><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p></p></td><td class="cell"><p></p></td><td class="cell"><p>With</p></td><td class="cell"><p>No</p></td><td class="cell"><p></p></td><td class="cell"><p></p></td><td class="cell"><p></p></td><td class="cell"><p>With</p></td><td class="cell"><p>No</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p></p></td><td class="cell"><p></p></td><td class="cell"><p>Negative</p></td><td class="cell"><p>Negative</p></td><td class="cell"><p></p></td><td class="cell"><p></p></td><td class="cell"><p></p></td><td class="cell"><p>Negative</p></td><td class="cell"><p>Negative</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p></p></td><td class="cell"><p>Overlap</p></td><td class="cell"><p>Evidence</p></td><td class="cell"><p>Evidence</p></td><td class="cell"><p></p></td><td class="cell"><p></p></td><td class="cell"><p>Overlap</p></td><td class="cell"><p>Evidence</p></td><td class="cell"><p>Evidence</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>Marked</p></td><td class="cell"><p>Median</p></td><td class="cell"><p>0.10</p></td><td class="cell"><p>Ü.09</p></td><td class="cell"><p></p></td><td class="cell"><p>Marked</p></td><td class="cell"><p>Median</p></td><td class="cell"><p>1.00</p></td><td class="cell"><p>1.00</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>Prepositions</p></td><td class="cell"><p>Mean</p></td><td class="cell"><p>0.17</p></td><td class="cell"><p>0.17</p></td><td class="cell"><p></p></td><td class="cell"><p>Prepositions</p></td><td class="cell"><p>Mean</p></td><td class="cell"><p>0.99</p></td><td class="cell"><p>0.93</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p></p></td><td class="cell"><p>Perfect</p></td><td class="cell"><p>6.3%</p></td><td class="cell"><p>5.2%</p></td><td class="cell"><p></p></td><td class="cell"><p></p></td><td class="cell"><p>Perfect</p></td><td class="cell"><p>97.9%</p></td><td class="cell"><p>88.0%</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>Ignored</p></td><td class="cell"><p>Median</p></td><td class="cell"><p>0.10</p></td><td class="cell"><p>0.09</p></td><td class="cell"><p></p></td><td class="cell"><p>Ignored</p></td><td class="cell"><p>Median</p></td><td class="cell"><p>1.00</p></td><td class="cell"><p>1.00</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>Prepositions</p></td><td class="cell"><p>Mean</p></td><td class="cell"><p>0.17</p></td><td class="cell"><p>0.16</p></td><td class="cell"><p></p></td><td class="cell"><p>Prepositions</p></td><td class="cell"><p>Mean</p></td><td class="cell"><p>0.96</p></td><td class="cell"><p>0.69</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p></p></td><td class="cell"><p>Perfect</p></td><td class="cell"><p>6.3%</p></td><td class="cell"><p>4.2%</p></td><td class="cell"><p></p></td><td class="cell"><p></p></td><td class="cell"><p>Perfect</p></td><td class="cell"><p>87.4%</p></td><td class="cell"><p>52.4%</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>Only</p></td><td class="cell"><p>Median</p></td><td class="cell"><p>0.10</p></td><td class="cell"><p>0.09</p></td><td class="cell"><p></p></td><td class="cell"><p>Only</p></td><td class="cell"><p>Median</p></td><td class="cell"><p>1.00</p></td><td class="cell"><p>0.54</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>Prepositions</p></td><td class="cell"><p>Mean</p></td><td class="cell"><p>0.16</p></td><td class="cell"><p>0.15</p></td><td class="cell"><p></p></td><td class="cell"><p>Prepositions</p></td><td class="cell"><p>Mean</p></td><td class="cell"><p>0.82</p></td><td class="cell"><p>0.57</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p></p></td><td class="cell"><p>Perfect</p></td><td class="cell"><p>3.1%</p></td><td class="cell"><p>3.1%</p></td><td class="cell"><p></p></td><td class="cell"><p></p></td><td class="cell"><p>Perfect</p></td><td class="cell"><p>66.5%</p></td><td class="cell"><p>42.9%</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td></tr></table><table caption="Table 4: Overall Results" class="main" frame="box" rules="all" border="1" regular="False"><tr class="row"><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p></p></td><td class="cell"><p>With</p></td><td class="cell"><p>No</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p></p></td><td class="cell"><p>Negative</p></td><td class="cell"><p>Negative</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p></p></td><td class="cell"><p>Evidence</p></td><td class="cell"><p>Evidence</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p><b>No Disambiguation</b></p></td><td class="cell"><p></p></td><td class="cell"><p></p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>Marked Prepositions</p></td><td class="cell"><p>6.3%</p></td><td class="cell"><p>5.2%</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>Ignored Prepositions</p></td><td class="cell"><p>6.3%</p></td><td class="cell"><p>4.2%</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>Only Prepositions</p></td><td class="cell"><p>3.1%</p></td><td class="cell"><p>3.1%</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p><b>With Disambiguation</b></p></td><td class="cell"><p></p></td><td class="cell"><p></p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>Marked Prepositions</p></td><td class="cell"><p>97.9%</p></td><td class="cell"><p>88.0%</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>Ignored Prepositions</p></td><td class="cell"><p>87.4%</p></td><td class="cell"><p>52.4%</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>Only Prepositions</p></td><td class="cell"><p>66.5%</p></td><td class="cell"><p>42.9%</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td></tr></table><page local="5"/><p>Given a verb, check Levin class.</p><p>1. If in Levin, classify directly.</p><p>2. If not in Levin, find synonym set from WordNet.</p><p>(a) If synonym in Levin, select the class that has the closest match with canonical LDOCE codes.</p><p>(b) If no synonyms in Levin or canonical LDOCE codes are completely mismatched, hypothesize new class.</p><p>Note that this algorithm assumes that there is a "canonical" set of LDOCE codes for each of Levin's semantic classes. Table 5 describes the significance of a subset of the syntactic codes in LDOCE. (The total number of codes is 174.) We have developed a relation between LDOCE codes and Levin classes, in much the same way that we associated syntactic signatures with the semantic classes in the earlier experiments. These canonical codes are for syntactic filtering (checking for the closest match) in the classification algorithm.</p><p>As an example of how the word-sense disambigua­tion process and classification, consider the non-Levin verb <i>attempt. </i>The LDOCE specification for this verb is: Tl T3 T4 WV5 N. Using the synonymy feature of WordNet, the algorithm automatically extracts five candidate classes associated with the synonyms of this word: (1) Class 29.6 "Masquerade Verbs" <i>(ad), </i>(2) Class 29.8 "Captain Verbs" <i>(pioneer), </i>(3) Class 31.1 "Amuse Verbs" <i>(try), </i>(4) Class 35.6 "Ferret Verbs" <i>(seek), </i>and (5) Class 55.2 "Complete Verbs" <i>(initiate). </i>The synonyms for each of these classes have the follow­ing LDOCE encodings, respectively: (1) I I-FOR I-ON I-UPON LI L9 Tl N; (2) L9 Tl N; (3) I Tl T3 T4 WV4 N; (4) I I-AFTER I-FOR TT T3; and (5) Tl T'1-INTO N. The largest intersection with the syntactic codes for <i>attempt </i>occurs with the verb <i>try </i>(Tl T3 T4 N). How­ever, Levin's class 31.1 is not the correct class for <i>at­tempt </i>since this sense of <i>try </i>has a "negative amuse" meaning (e.g., <i>John's behavior tried my patience. </i>In fact, the codes TT T3 T4 are not part of the canonical class-code mapping associated with class 31.1. Thus, <i>at­tempt </i>falls under case 2(b) of the algorithm, and a new class is hypothesized. This is a case where word-sense disambiguation has allowed us to classify a new word <i>and </i>to enhance Levin's verb classification by adding a new class to the word <i>try </i>as well. In our experiments, our algorithm found several additional non-Levin verbs that fell into this newly hypothesized class, including <i>aspire, attempt, dare, decide, desire, elect, need, </i>and <i>swear.</i></p><p>We have automatically classified 10,000 "unknown" verbs, i.e., those not occurring in the Levin classifica­tion, using this technique. These verbs arc taken from English "glosses" (i.e., translations) provided in bilin­gual dictionaries for Spanish and Arabic.<footnote anchor="5"/> As a pre­liminary measure of success, we picked out 84 LDOCE <i>control vocabulary </i>verbs, (i.e., primitive words used for defining dictionary entries) and hand-checked our re­sults. We found that 69 verbs were classified correctly,</p><p>''The Spanish-English dictionary was built at the Univer­sity of Maryland; The Arabic-English dictionary was pro­duced by Alpnet, a company in Utah that develops transla­tion aids. Wc are also in the process of developing bilingual dictionaries for Korean and French, and we will be porting our LCS acquisition technology to these languages in the near future.</p><doubt alpha="63.0" length="27" tooSmall="False" monospace="0.0">i.e., 82% accuracy.5Summary</doubt><p>Wc have conducted two experiments with the intent of addressing the issue of word-sense ambiguity in extrac­tion from machine-readable resources for the construc­tion of large-scale knowledge sources. In the first exper­iment, verbs that appeared in different classes collected the syntactic information from each class it appeared in. Therefore, the syntactic signature was composed from all of the example sentences from every class the verb appeared in. In some cases, the verbs were seman­tically unrelated and consequently the mapping from syntax to semantics was muddied. The second exper­iment attempted to determine a relationship between a semantic class and the syntactic information associ­ated with each class. Not surprisingly, but not insignif­icantly, this relationship was very clear, since this ex­periment avoided the problem of word sense ambiguity. These experiments served to validate Levin's claim that verb semantics and syntactic behavior are predictably related and also demonstrated that a significant com­ponent of any lexical acquisition program is the ability to perform word-sense disambiguation.</p><p>We have used the results of our first two experiments to help in constructing and augmenting online dictio­naries for novel verb senses. We have used the same syntactic signatures to categorize new verbs into Levin:s classes on the basis of WordNet and LDOCE. We arc; currently porting these results to new languages using online bilingual lexicons.</p></section><section title="Acknowledgements"><p>The research reported herein was supported, in part, by Army Research Office contract DAAL03-91-C-0034 through Battelle Corporation, NSF NYI IRI-9357731, Alfred P. Sloan Research Fellow Award BR3336, and a General Research Board Semester Award.</p><table caption="Table 5: Sample Syntactic Codes used in LDOCE" class="main" frame="box" rules="all" border="1" regular="False"><tr class="row"><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p><b>LDOCE Code</b></p></td><td class="cell"><p><b>Arguments</b></p></td><td class="cell"><p><b>Adjuncts</b></p></td><td class="cell"><p><b>Example</b></p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>I</p></td><td class="cell"><p></p></td><td class="cell"><p></p></td><td class="cell"><p>Olivier is <b>acting </b>tonight</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>I-AFTER</p></td><td class="cell"><p>—---------------</p></td><td class="cell"><p>PP[afl.er] PPjfor]</p></td><td class="cell"><p>She <b>sought </b>after the truth</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>FI'OR</p></td><td class="cell"><p>They <b>sought </b>for the right one</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>I-ON</p></td><td class="cell"><p>--</p></td><td class="cell"><p>PP[on]</p></td><td class="cell"><p>lie <b>acted </b>on our suggestion</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>I-UPON</p></td><td class="cell"><p></p></td><td class="cell"><p>PPfuponJ</p></td><td class="cell"><p>The drug <b>acted </b>upon the pain</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>l:i</p></td><td class="cell"><p>NP</p></td><td class="cell"><p></p></td><td class="cell"><p>Tie <b>acts </b>the experienced man</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>L9</p></td><td class="cell"><p>aüv/PP</p></td><td class="cell"><p></p></td><td class="cell"><p>The play <b>acts </b>well</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>Tl</p></td><td class="cell"><p>NP</p></td><td class="cell"><p></p></td><td class="cell"><p>I <b>pioneered </b>the new land</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>TMNTO</p></td><td class="cell"><p>NP</p></td><td class="cell"><p>PPhntoJ</p></td><td class="cell"><p>We <b>initiated </b>him into the group</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>T3</p></td><td class="cell"><p>VP[to+infJ</p></td><td class="cell"><p>-----------</p></td><td class="cell"><p>He <b>tried </b>to do it</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>T4</p></td><td class="cell"><p>VP[+]m&gt;gT</p></td><td class="cell"><p>She <b>tried </b>eating the new food</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>WV4</p></td><td class="cell"><p>-ing adjectival</p></td><td class="cell"><p></p></td><td class="cell"><p>I've had a <b>trying </b>day</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>WV5</p></td><td class="cell"><p>-ed adjectival</p></td><td class="cell"><p></p></td><td class="cell"><p>He was convicted for <b>attempted murder</b></p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"><p>N (dénommai verb)</p></td><td class="cell"><p></p></td><td class="cell"><p></p></td><td class="cell"><p><b>pioneer </b>(noun)</p></td><td class="cell"></td></tr><tr class="row"><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td><td class="cell"></td></tr></table></section><references><p>Alshawi, II. 1.989. Analysing the Dictionary Defini­tions. In B. Boguraev and T. Briscoe, editor, <i>Compu­tational Lexicography for Natural Language Process­ing. </i>Longman, London, pages 153- 169.</p><p>Boguraev, B. and T. Briscoe. 1989. Utilising the LDOCE Grammar Codes. In B. Boguraev and T. Briscoe, editor, <i>Computational Lexicography for Nat­ural Language Processing. </i>Longman, London, pages 85-116.</p><p>Brent, M. 1993. Unsupervised Learning of Lexical Syn­tax. <i>Computational Linguistics, </i>19:243-262.</p><p>Church, K. and P. Hanks. 1990. Word Association Norms, Mutual Information and Lexicography. <i>Com­putational Linguistics, </i>16:22 29.</p><p>Copestake, A., T. Briscoe, P. Vossen, A. Ageno, I. Castellon, F. Ribas, G. Rigau, II. Rodriguez, and A. Samiotou. 1995. Acquisition of Lexical Transla­tion Relations from MRDS. <i>Machine Translation, </i>9.</p><p>Dorr, B., J. Garman, and A. Weinberg. 1995. From Syntactic Encodings to Thematic Roles: Building Lexical Entries for Interlingual MT. <i>Machine Trans­lation, </i>9.</p><doubt alpha="0.0" length="3" tooSmall="False" monospace="0.0">326</doubt><page local="6"/><p>Dorr, B. and C. Voss, to appear. A Multi-Level Ap­proach to Interlingual MT: Defining the Interface between Representational Languages. <i>International Journal of Expert Systems.</i></p><p>Dubois, D. and P. Saint-Dinier. 1995. Construction cf representation de classes sémantiques de verbes: une coopération entre syntaxe et cognition, manuscript, I Ii 11' CNRS, Toulouse, France.</p><p>Farwell, D., L. Guthrie, and Y. Wilks. 1993. Automat­ically Creating Lexical Entries for ULTRA, a Multi­lingual MT System. <i>Machine 'Translation, </i>8(3).</p><p>Fillmore, C.J. 1968. The (Jase for Case. In 10. Bach and R.'T. Harms, editor, <i>Univers als in Linguistic Theory. </i>Holt, Rinchart, and Winston, pages 1 88.</p><p>Grimshaw, J. 1990. <i>Argument Structure. </i>MIT Press, Cambridge, MA.</p><p>Gruber, J.S. 1965. <i>Studies in Lexical Relations. </i>Ph.D. thesis, MIT, Cambridge, MA.</p><p>Guthrie, J., L. Guthrie, Y. Wilks, and II. Aidinejad. 1991. Subject-Dependent Co-occurrence and Word Sense Disambiguation. In <i>Proceedings of the 29th An­nual Meeting of the Association for Computational Linguistics, </i>pages 146 152, University of California, Berkeley, CA.</p><p>Hearst, M. 1991. Noun Homograph Disambiguation Using Local Context in Large Text Corpora. In <i>Using Corpora, </i>University of Waterloo, Waterloo, Ontario.</p><p>Jackcndoff, R. 1983. <i>Semantics and Cognition. </i>MIT Press, Cambridge, MA.</p><p>Jackcndoff, R. 1990. <i>Semantic Structures. </i>MIT Press, Cambridge, MA.</p><p>Klavans, J.L. and E. Tzoukcrmann.  1996. Dictionar­ies and Corpora: Combining Corpus and Machine-readable Dictionary Data for Building Bilingual Lex icons. <i>Machine Translation, </i>10.</p><p>Levin, B. 1993. <i>English Verb Classes and Alternations: A Preliminary Investigation. </i>Chicago, IL.</p><p>Lonsdale, D., T. Mitamura, and 10. Nyberg. 1996. Ac­quisition of Large Lexicons for Practical Knowledge Based MT. <i>Machine Translation; </i>9.</p><p>Miller, G. 1985. WORDNET: A Dictionary Browser. In <i>Proceedings of the First International Conference</i> <i>on Information in Data, </i>University of Waterloo Cen­tre for the New OED, Waterloo, Ontario.</p><p>Neff, M. and M. McCord. 1990. Acquiring Lexical Data from Machine-Readable Dictionary Resources for Machine 'Translation. In <i>Third International Con­ference on Theoretical and Methodological Issues in Machine 'translation of Natural Languages (TMl-90), </i>Austin, Texas.</p><p>Pinker, S. 1989. <i>Learnability and Cognition: The Ac­quisition of Argument Structure. </i>MIT Press, Cam­bridge, MA.</p><p>Procter, P. 1978. <i>Longman Dictionary of Contempo­rary English. </i>Longman, London.</p><p>Sanfilippo, A. and V. Po/,nanski.   1992. The Acquisi­tion of Lexical Knowledge from Combined Machine-Readable Dictionary Resources.   In <i>Proceedings of the Applied Natural Language Processing Conference, </i>pages 80-87, Trcnto, Italy.</p><p>Walker, D. and R. Amsler. 1986. The Use of Machine-readable Dictionaries in Sublanguage Analysis. In R. (irishman and Ii. Kittrcdge, editors, <i>Analyzing Language in Restricted Domains. </i>Lawrence lOrlbaum Associates, Hillsdale, New Jersey, pages 69 83.</p><p>Wilks, Y., I). Pass, CM. Guo, ,1.10. McDonald, and T. Plate. 1990. Providing Machine Tractable Dic­tionary Tools. <i>Machine 'translation, </i>5(2):9Ü 154.</p><p>Wilks, Y., I). Fass, CM. Guo, .1.10. McDonald, T. Plate, and B.M. Slator. 1989. A Tractable Machine Dictio­nary as a. Resource for Computational Semantics. In li. Boguraev and T. Briscoe, editor, <i>Computational Lexicography for Natural Language Processing. </i>Long­man, London, pages 85 116.</p><p>Wu, D. and X. Xia. J995. Largo-Scale Automatic Ex­traction of an English-Chinese Translation Lexicon. <i>Machine Translation, </i>9.</p><p>Yarowsky, 1). 1992. Word-Sense Disambiguation:, Us­ing Statistical Models of Roget's (Jategories Trained on Large Corpora. In <i>Proceedings of the Fourteenth International Conference on Computational Linguis­tics, </i>pages 454-460, Nantes, France.</p><doubt alpha="0.0" length="3" tooSmall="False" monospace="0.0">327</doubt></references></body></article>