acl acl2013 acl2013-6 knowledge-graph by maker-knowledge-mining

6 acl-2013-A Java Framework for Multilingual Definition and Hypernym Extraction


Source: pdf

Author: Stefano Faralli ; Roberto Navigli

Abstract: In this paper we present a demonstration of a multilingual generalization of Word-Class Lattices (WCLs), a supervised lattice-based model used to identify textual definitions and extract hypernyms from them. Lattices are learned from a dataset of automatically-annotated definitions from Wikipedia. We release a Java API for the programmatic use of multilingual WCLs in three languages (English, French and Italian), as well as a Web application for definition and hypernym extraction from user-provided sentences.

Reference: text


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 it i Abstract In this paper we present a demonstration of a multilingual generalization of Word-Class Lattices (WCLs), a supervised lattice-based model used to identify textual definitions and extract hypernyms from them. [sent-3, score-0.255]

2 Lattices are learned from a dataset of automatically-annotated definitions from Wikipedia. [sent-4, score-0.113]

3 We release a Java API for the programmatic use of multilingual WCLs in three languages (English, French and Italian), as well as a Web application for definition and hypernym extraction from user-provided sentences. [sent-5, score-0.688]

4 1 Introduction Electronic dictionaries and domain glossaries are definition repositories which prove very useful not only for lookup purposes, but also for automatic tasks such as Question Answering (Cui et al. [sent-6, score-0.177]

5 , 2013), domain Word Sense Disambiguation (Duan and Yates, 2010; Faralli and Navigli, 2012), automatic acquisition of semantic predicates (Flati and Navigli, 2013), relation extraction (Yap and Baldwin, 2009) and, more in general, knowledge acquisition (Hovy et al. [sent-9, score-0.15]

6 Nonetheless, raw text often contains several definitional sentences, that is, it provides within itselfformal explanations for terms ofinterest. [sent-13, score-0.282]

7 Whilst it is not feasible to search texts manually for definitions in several languages, the task of extracting definitional information can be automatized by means of Machine Learning (ML) and Natural Language Processing (NLP) techniques. [sent-14, score-0.369]

8 However, these methods suffer from two signifiicant drawbacks: on the one hand, low recall (as definitional sentences occur in highly variable syntactic structures), and, on the other hand, noise (because the most frequent definitional pattern X is a Y is inherently very – – noisy). [sent-17, score-0.603]

9 A recent approach to definition and hypernym extraction, called Word-Class Lattices (Navigli and Velardi, 2010, WCLs), overcomes these issues by addressing the variability of definitional sentences and providing a flexible way of automatically extracting hypernyms from them. [sent-18, score-0.936]

10 To do so, lattice-based classifiers are learned from a training set of textual definitions. [sent-19, score-0.059]

11 Training sentences are automatically clustered by similarity and, for each such cluster, a lattice classifier is learned which models the variants of the definition template detected. [sent-20, score-0.326]

12 A lattice is a directed acyclic graph, a subclass ofnon-deterministic finite state automata. [sent-21, score-0.06]

13 The purpose of the lattice structure is to preserve (in a compact form) the salient differences among distinct sequences. [sent-22, score-0.06]

14 Since multilinguality is a key need in today’s information society, and because WCLs have been tested overwhelmingly only with the English language, we provide experiments for three different languages, namely English, French and Italian. [sent-24, score-0.041]

15 To do so, in contrast to Navigli and Velardi (2010), who created a manually annotated training set of definitions, we provide a heuristic method for the au- tomatic acquisition of reliable training sets from Wikipedia, and use them to determine the robustness and generalization power of WCLs. [sent-25, score-0.136]

16 We show high performance in definition and hypernym extraction for our three languages. [sent-26, score-0.581]

17 c A2s0s1o3ci Aatsiosonc fioartio Cno fmorpu Ctoamtiopnuatalt Lioinngauli Lsitnicgsu,i psatgices 103–108, [In geography, a country]DF [is]V F [a political division]GF . [sent-34, score-0.032]

18 [In finance, a bond]DF [is]V F [a negotiable certificate]GF [that that acknowledges. [sent-35, score-0.046]

19 Table 1: Example definitions (defined terms are marked in bold face, their hypernyms in italics). [sent-43, score-0.19]

20 geography finance division certificate InpoNetNr1y,ahTARGETi aJJNNm2easure country bond foot Figure 1: The DF and GF Word-Class Lattices for the political negotiable sentences in Table 1. [sent-44, score-0.399]

21 following fields (Storrer and Wellinghoff, 2006): definiendum (DF), definitor (VF), definiens (GF) and rest (REST), where DF is the part of the definition including the word being defined (e. [sent-45, score-0.378]

22 , “In computer science, a closure”), VF is the verb phrase used to introduce the definition (e. [sent-47, score-0.167]

23 , “a firstclass function”, hypernym marked in italics) and RF includes additional clauses (e. [sent-51, score-0.408]

24 Consider a set of training sentences T, each of which is automatically part-of-speech tagged and manually bracketed with the DF, VF, GF and REST fields (examples are shown in Table 1). [sent-54, score-0.158]

25 , w|s| , where wi is the i-th word of s, we generalize |s it|s words wi to word classes ωi as follows: ωi=? [sent-66, score-0.08]

26 wPiOS(wi) oift hweirw∈i Fse that is, a word wi is left unchanged if it occurs frequently in the training corpus (i. [sent-67, score-0.073]

27 , wi ∈ F) or is transformed to its part of speech tag (P∈O FS)( owri) is) otherwise. [sent-69, score-0.04]

28 As a result, we obtain a generalized sentence s0 = ω1 , ω2, . [sent-70, score-0.037]

29 For instance, given the first sentence in Tabl|es| 1, we obtain the corresponding generalized sentence: “In NN, a hTARGETi is a JJ NN”, where NN and JJ indicate thhTeA nRoGunE Tanid is sa adj JeJc tNivNe” c,l washseerse, NreNsp aencdtiv JeJly in. [sent-74, score-0.037]

30 2 Learning The WCL learning algorithm consists of 3 steps: • Star patterns: each sentence in the training sSetta irs ppartetperroncse:ss eeadc han sden gteenncerea ilnize thde t tor ain sintagr pattern by replacing with * all the words wi ∈ F, i. [sent-76, score-0.11]

31 In Figure 1 we show the resulting lattices for the DF and GF fields built for the cluster of sentences of Table 1. [sent-80, score-0.325]

32 Note that during the construction of the lattice the nodes associated with the hypernym words in the learning sentences (i. [sent-81, score-0.474]

33 , division, certificate and measure) are marked as hypernyms in order to determine the hypernym of a test sentence at classification time (see (Navigli and Velardi, 2010) for details). [sent-83, score-0.571]

34 Given a test sentence s, we consider all possible combinations of definiendum, definitor and definiens lattices and select the combination ofthe three WCLs that best fits the sentence, if such a combination exists. [sent-86, score-0.41]

35 In fact, choosing the most appropriate combination of lattices impacts the performance of hypernym extraction. [sent-87, score-0.633]

36 selecting the words in the input sentence that are marked as hypernyms in the WCL selected for GF. [sent-90, score-0.14]

37 As a result, we are able to fully automatize the definition and hypernym extraction by utilizing collaboratively-curated encyclopedia content. [sent-92, score-0.581]

38 candidate definition extraction: we iterate through the collection of Wikipedia pages for the language of interest. [sent-95, score-0.198]

39 For each article we extract the first paragraph, which usually, but not always, contains a definitional sentence for the concept expressed by the page title. [sent-96, score-0.356]

40 We discard all those pages for which the title corresponds to a special page (i. [sent-97, score-0.095]

41 part-of-speech tagging and phrase chunking: for each candidate definition we perform part-of-speech tagging and chunking, thus automatically identifying noun, verb, and prepositional phrases (we use TreeTagger (Schmid, 1997)). [sent-111, score-0.225]

42 automatic annotation: we replace all the occurrences in the candidate definition of the target term (i. [sent-113, score-0.205]

43 h, definiens field) the phrase which contains the hypernym and as DF (i. [sent-123, score-0.469]

44 , definiendum field) the phrase which starts at the beginning of the sentence and ends right before the start of the VP tag. [sent-125, score-0.131]

45 Finally we mark as REST the remaining phrases after the phrase already tagged as GF. [sent-126, score-0.055]

46 For example, given the sentence “Albert Einstein was a German-born theoretical physicist. [sent-127, score-0.037]

47 ”, we produce the following sentence annotation: “[Albert Einstein]DF [was]V F [a German-born theoretical physicist]GF . [sent-128, score-0.037]

48 ” (target term marked in bold and hypernym in italics). [sent-129, score-0.441]

49 filtering: we finally discard all the candidate definitions for which not all fields could be found during the previous step (i. [sent-131, score-0.145]

50 , either the hTARGETi, hypernym or any DF, VF, GF, hRTEASRTG GtEagT iis, m hiyspseinrngy). [sent-133, score-0.375]

51 The numbers are shown in Table 2: starting with 3,904,360 Wikipedia pages for English, 1,617,359 for French and 1,008,044 for Italian (first column), we obtained 1,552,493, 447,772, and 291,259 automatically tagged sentences, respectively, for the three languages (second column in the Table). [sent-135, score-0.137]

52 Since we next had to use these sentences for training our WCLs, we took out a random sample of 1000 sentences for each language which we used for testing purposes. [sent-136, score-0.111]

53 We manually annotated each of these sentences as definitional or nondefinitional2 and, in the case of the former, also with the correct hypernym. [sent-137, score-0.321]

54 2 Evaluation We tested the newly acquired training dataset against two test datasets. [sent-139, score-0.08]

55 The first dataset was our random sampling of 1000 Wikipedia test sentences which we had set aside for each language (no intersection with the training set, see previous section). [sent-140, score-0.072]

56 The second dataset was the same one used in Navigli and Velardi (2010), made up of sentences from the ukWaC Web corpus (Ferraresi et al. [sent-141, score-0.039]

57 , 2008) and used to estimate the definition and hypernym extraction performance on an open text corpus. [sent-142, score-0.581]

58 3 Results Table 3 shows the results obtained on definition (column 2-4) and hypernym extraction (column 57) in terms of precision (P), recall (R) and accuracy (A) on our first dataset. [sent-144, score-0.581]

59 Note that accuracy also takes into account candidate definitions in the test set which were tagged as non-definitional (see Section 3. [sent-145, score-0.147]

60 In the Table we compare the performance of our English WCL trained from Wikipedia sentences using our automatic procedure against the original performance of WCL 1We used the 21-09-2012 (EN), 17-09-2012 (FR), 21-092012 (IT) dumps. [sent-147, score-0.039]

61 2Note that the first sentence of a Wikipedia page might seldom be non-definitional, such as “Basmo fortress is located in the north-western part . [sent-148, score-0.074]

62 a ncy (A) of definition and hypernym extraction when testing on our dataset of 1000 randomly sampled Wikipedia first-paragraph sentences. [sent-158, score-0.581]

63 EN (2010) refers to the WCL learned from the original manually-curated training set from Navigli and Velardi (2010), while EN, FR and IT refer to WCL trained, respectively, with one of the three training sets automatically acquired from Wikipedia. [sent-159, score-0.167]

64 56 Table 4: Estimated WCL definition extraction precision (P) and recall (R), testing a sample of ukWaC sentences. [sent-162, score-0.206]

65 on It c 1a,9n0 b8e m saeneuna tlhlya-ts tehle c atuetdom tra ti ncianlgly s aecn- quired training set considerably improves the performance, as it covers higher variability. [sent-164, score-0.033]

66 We note that the recall in both definition and hypernym extraction is higher for French and Italian. [sent-165, score-0.581]

67 We remark that the presented results were obtained without any human effort, except for the independent collaborative editing and hyperlinking of Wikipedia pages, and that the overall performances can be improved by manually checking the automatically annotated training datasets. [sent-167, score-0.061]

68 As can be seen in Table 4, the estimated precision and recall for WCL definition extraction with the 2010 training set were 94. [sent-169, score-0.239]

69 5%, respectively, while with our automatically acquired English training set we obtained a higher precision of 98. [sent-171, score-0.108]

70 This second experiment shows that learning WCLs from hundreds of thousands of defini- tion candidates does not overfit to Wikipedia-style definitional sentences. [sent-174, score-0.282]

71 After looking at the automatically acquired training datasets, we noted some erroneous annotations mainly due to the following factors: i) some Wikipedia pages do not start with defini3Available from http : / / l . [sent-175, score-0.134]

72 tional sentences; ii) they may contain more than one verbal phrase between the defined term and the hypernym; iii) the first link after the verbal phrase does not cover, or partially covers, the correct hypernym. [sent-178, score-0.083]

73 The elimination of the above wrongly acquired definitional patterns can be implemented with some language-dependent heuristics or can be done by human annotators. [sent-179, score-0.39]

74 In any case, given the presence of a high number of correct annotated sentences, these wrong definitional patterns have a very low impact on the definition and hypernym extraction precision as shown in the above experiments (see Table 3 and Table 4). [sent-180, score-0.924]

75 4 Multilingual WCL API Together with the training and test sets of the above experiments, we also release here our implementation of Word-Class Lattices, available as a Java API. [sent-181, score-0.068]

76 As a result the WCL classifier can easily be used programmatically in any Java project. [sent-182, score-0.069]

77 After the selection of the target language (line 2), we load the training dataset for the target language (line 4). [sent-184, score-0.033]

78 Then an instance of WCLClas s i fier is created (line 8) and the training phase is launched on the input training corpora (line 9). [sent-185, score-0.066]

79 Now the classifier is ready to be tested on any given sentence in the target language (lines 11-16). [sent-186, score-0.068]

80 If the classifier output is positive (line 18) we can print the extracted hypernym (line 19). [sent-187, score-0.406]

81 The output of the presented code is the string “classifier” which corresponds to the hypernym extracted by WCL for the input sentence “WCL is a kind of classifier”. [sent-188, score-0.412]

82 1 Web user interface We also release a Web interface to enable online usage of our WCLs for the English, French and Italian languages. [sent-190, score-0.035]

83 term of interest, the candidate definition, select the language of interest and, after submission, in the case of positive response from WCL, obtain the corresponding hypernym and a graphical representation of the lattices matching the given sentence, as shown in the bottom part of the Figure. [sent-193, score-0.696]

84 The graphical representation shows the concatenation of the learned lattices which match the DF, VF, GF parts of the given sentence (see Section 2). [sent-194, score-0.321]

85 We also allow the user not to provide the term of interest: in this case all the nouns in the sentence are considered as candidate defined terms. [sent-195, score-0.1]

86 The server side produces an HTML page (like the one shown in Figure 3), using the WCL API (see Section 4) to process and test the submitted definition candidate. [sent-197, score-0.179]

87 5 Related Work A great deal of work is concerned with the language independent extraction of definitions. [sent-198, score-0.064]

88 Much recent work uses symbolic methods that depend on lexico-syntactic patterns or features, which are manually created or semi-automatically learned as recently done in (Zhang and Jiang, 2009; Westerhout, 2009). [sent-199, score-0.087]

89 (2008), in order to improve precision while keeping pattern generality, prune candidates using more refined stylistic patterns and lexical filters. [sent-203, score-0.061]

90 (2007) propose the use of probabilistic lexico-semantic patterns, for definitional question answering in the TREC contest4. [sent-205, score-0.282]

91 gov Hypernym extraction methods vary from simple lexical patterns (Hearst, 1992; Oakes, 2005) to statistical and machine learning techniques (Agirre et al. [sent-209, score-0.125]

92 , 2013), where given a definitional sentence the hypernym is identified as the first occuring noun after the defined term. [sent-214, score-0.694]

93 They first search sentences that contain two terms which are known to be in a taxonomic relation (term pairs are taken from WordNet (Miller et al. [sent-217, score-0.039]

94 , 1990)); then they parse the sentences, and automatically extract patterns from the parse trees. [sent-218, score-0.089]

95 Finally, they train a hypernym classifier based on these features. [sent-219, score-0.406]

96 Lexico-syntactic patterns are generated for each sentence relating a term to its hypernym, and a dependency parser is used to represent them. [sent-220, score-0.131]

97 Using Hearst’s rules for the automatic acquisition of hyponyms for mining a pharmaceutical corpus. [sent-303, score-0.073]

98 The acquisition of lexical knowledge from combined machinereadable dictionary sources. [sent-315, score-0.043]

99 Automated detection and annotation of term definitions in German text corpora. [sent-332, score-0.12]

100 In Proceedings of the RANLP 2009 Workshop on Definition Extraction, page 61–67, Borovets, Bulgaria. [sent-345, score-0.037]


similar papers computed by tfidf model

tfidf for this paper:

wordName wordTfidf (topN-words)

[('hypernym', 0.375), ('wcl', 0.346), ('wcls', 0.322), ('definitional', 0.282), ('lattices', 0.258), ('gf', 0.235), ('velardi', 0.202), ('df', 0.193), ('navigli', 0.185), ('htargeti', 0.161), ('definition', 0.142), ('vf', 0.109), ('faralli', 0.088), ('definitions', 0.087), ('roberto', 0.087), ('stefano', 0.084), ('api', 0.078), ('wikipedia', 0.077), ('hypernyms', 0.07), ('definiendum', 0.069), ('definiens', 0.069), ('ukwac', 0.067), ('java', 0.066), ('roma', 0.064), ('extraction', 0.064), ('french', 0.062), ('patterns', 0.061), ('lattice', 0.06), ('paola', 0.058), ('certificate', 0.056), ('italian', 0.054), ('division', 0.049), ('acquired', 0.047), ('geography', 0.047), ('italics', 0.047), ('multilingual', 0.046), ('definitor', 0.046), ('negotiable', 0.046), ('storrer', 0.046), ('yap', 0.046), ('acquisition', 0.043), ('borovets', 0.041), ('hearst', 0.041), ('flati', 0.041), ('multilinguality', 0.041), ('sanfilippo', 0.041), ('wi', 0.04), ('sentences', 0.039), ('web', 0.039), ('cui', 0.039), ('benedictis', 0.038), ('borg', 0.038), ('einstein', 0.038), ('programmatically', 0.038), ('page', 0.037), ('sentence', 0.037), ('hovy', 0.036), ('glossaries', 0.035), ('release', 0.035), ('country', 0.034), ('foot', 0.034), ('bond', 0.034), ('ferraresi', 0.034), ('term', 0.033), ('training', 0.033), ('snow', 0.033), ('marked', 0.033), ('political', 0.032), ('title', 0.032), ('kozareva', 0.031), ('classifier', 0.031), ('candidate', 0.03), ('albert', 0.03), ('hyponyms', 0.03), ('screenshot', 0.03), ('tagged', 0.03), ('automatically', 0.028), ('closure', 0.028), ('fields', 0.028), ('eduard', 0.028), ('dolan', 0.028), ('finance', 0.028), ('star', 0.028), ('ritter', 0.027), ('generalization', 0.027), ('column', 0.027), ('pages', 0.026), ('line', 0.026), ('languages', 0.026), ('learned', 0.026), ('demonstration', 0.025), ('duan', 0.025), ('trec', 0.025), ('phrase', 0.025), ('uppsala', 0.024), ('chunking', 0.024), ('ranlp', 0.024), ('agirre', 0.024), ('rest', 0.024), ('field', 0.024)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 1.0000007 6 acl-2013-A Java Framework for Multilingual Definition and Hypernym Extraction

Author: Stefano Faralli ; Roberto Navigli

Abstract: In this paper we present a demonstration of a multilingual generalization of Word-Class Lattices (WCLs), a supervised lattice-based model used to identify textual definitions and extract hypernyms from them. Lattices are learned from a dataset of automatically-annotated definitions from Wikipedia. We release a Java API for the programmatic use of multilingual WCLs in three languages (English, French and Italian), as well as a Web application for definition and hypernym extraction from user-provided sentences.

2 0.52061415 152 acl-2013-Extracting Definitions and Hypernym Relations relying on Syntactic Dependencies and Support Vector Machines

Author: Guido Boella ; Luigi Di Caro

Abstract: In this paper we present a technique to reveal definitions and hypernym relations from text. Instead of using pattern matching methods that rely on lexico-syntactic patterns, we propose a technique which only uses syntactic dependencies between terms extracted with a syntactic parser. The assumption is that syntactic information are more robust than patterns when coping with length and complexity of the sentences. Afterwards, we transform such syntactic contexts in abstract representations, that are then fed into a Support Vector Machine classifier. The results on an annotated dataset of definitional sentences demonstrate the validity of our approach overtaking current state-of-the-art techniques.

3 0.2084866 306 acl-2013-SPred: Large-scale Harvesting of Semantic Predicates

Author: Tiziano Flati ; Roberto Navigli

Abstract: We present SPred, a novel method for the creation of large repositories of semantic predicates. We start from existing collocations to form lexical predicates (e.g., break ∗) and learn the semantic classes that best f∗it) tahned ∗ argument. Taon idco this, we extract failtl thhee ∗ occurrences ion Wikipedia ewxthraiccht match the predicate and abstract its arguments to general semantic classes (e.g., break BODY PART, break AGREEMENT, etc.). Our experiments show that we are able to create a large collection of semantic predicates from the Oxford Advanced Learner’s Dictionary with high precision and recall, and perform well against the most similar approach.

4 0.14786683 170 acl-2013-GlossBoot: Bootstrapping Multilingual Domain Glossaries from the Web

Author: Flavio De Benedictis ; Stefano Faralli ; Roberto Navigli

Abstract: We present GlossBoot, an effective minimally-supervised approach to acquiring wide-coverage domain glossaries for many languages. For each language of interest, given a small number of hypernymy relation seeds concerning a target domain, we bootstrap a glossary from the Web for that domain by means of iteratively acquired term/gloss extraction patterns. Our experiments show high performance in the acquisition of domain terminologies and glossaries for three different languages.

5 0.083744586 43 acl-2013-Align, Disambiguate and Walk: A Unified Approach for Measuring Semantic Similarity

Author: Mohammad Taher Pilehvar ; David Jurgens ; Roberto Navigli

Abstract: Semantic similarity is an essential component of many Natural Language Processing applications. However, prior methods for computing semantic similarity often operate at different levels, e.g., single words or entire documents, which requires adapting the method for each data type. We present a unified approach to semantic similarity that operates at multiple levels, all the way from comparing word senses to comparing text documents. Our method leverages a common probabilistic representation over word senses in order to compare different types of linguistic data. This unified representation shows state-ofthe-art performance on three tasks: seman- tic textual similarity, word similarity, and word sense coarsening.

6 0.05546264 154 acl-2013-Extracting bilingual terminologies from comparable corpora

7 0.055405781 291 acl-2013-Question Answering Using Enhanced Lexical Semantic Models

8 0.054491244 238 acl-2013-Measuring semantic content in distributional vectors

9 0.052095495 7 acl-2013-A Lattice-based Framework for Joint Chinese Word Segmentation, POS Tagging and Parsing

10 0.048600797 62 acl-2013-Automatic Term Ambiguity Detection

11 0.048256356 160 acl-2013-Fine-grained Semantic Typing of Emerging Entities

12 0.045629926 234 acl-2013-Linking and Extending an Open Multilingual Wordnet

13 0.044330966 96 acl-2013-Creating Similarity: Lateral Thinking for Vertical Similarity Judgments

14 0.044006828 311 acl-2013-Semantic Neighborhoods as Hypergraphs

15 0.043144047 292 acl-2013-Question Classification Transfer

16 0.042229321 98 acl-2013-Cross-lingual Transfer of Semantic Role Labeling Models

17 0.042223681 8 acl-2013-A Learner Corpus-based Approach to Verb Suggestion for ESL

18 0.041441407 373 acl-2013-Using Conceptual Class Attributes to Characterize Social Media Users

19 0.040538222 162 acl-2013-FrameNet on the Way to Babel: Creating a Bilingual FrameNet Using Wiktionary as Interlingual Connection

20 0.040186454 105 acl-2013-DKPro WSD: A Generalized UIMA-based Framework for Word Sense Disambiguation


similar papers computed by lsi model

lsi for this paper:

topicId topicWeight

[(0, 0.145), (1, 0.039), (2, 0.001), (3, -0.091), (4, -0.018), (5, -0.018), (6, -0.083), (7, 0.002), (8, 0.054), (9, -0.028), (10, -0.027), (11, 0.018), (12, -0.056), (13, 0.004), (14, 0.032), (15, -0.051), (16, 0.054), (17, 0.01), (18, -0.016), (19, 0.014), (20, -0.036), (21, 0.047), (22, -0.069), (23, 0.09), (24, 0.236), (25, 0.072), (26, 0.004), (27, -0.153), (28, 0.054), (29, 0.241), (30, -0.127), (31, 0.325), (32, 0.087), (33, 0.083), (34, 0.321), (35, 0.017), (36, 0.009), (37, 0.241), (38, 0.006), (39, -0.021), (40, 0.027), (41, 0.011), (42, -0.012), (43, -0.049), (44, -0.018), (45, -0.049), (46, -0.092), (47, 0.136), (48, 0.023), (49, 0.001)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.92897916 6 acl-2013-A Java Framework for Multilingual Definition and Hypernym Extraction

Author: Stefano Faralli ; Roberto Navigli

Abstract: In this paper we present a demonstration of a multilingual generalization of Word-Class Lattices (WCLs), a supervised lattice-based model used to identify textual definitions and extract hypernyms from them. Lattices are learned from a dataset of automatically-annotated definitions from Wikipedia. We release a Java API for the programmatic use of multilingual WCLs in three languages (English, French and Italian), as well as a Web application for definition and hypernym extraction from user-provided sentences.

2 0.88613737 152 acl-2013-Extracting Definitions and Hypernym Relations relying on Syntactic Dependencies and Support Vector Machines

Author: Guido Boella ; Luigi Di Caro

Abstract: In this paper we present a technique to reveal definitions and hypernym relations from text. Instead of using pattern matching methods that rely on lexico-syntactic patterns, we propose a technique which only uses syntactic dependencies between terms extracted with a syntactic parser. The assumption is that syntactic information are more robust than patterns when coping with length and complexity of the sentences. Afterwards, we transform such syntactic contexts in abstract representations, that are then fed into a Support Vector Machine classifier. The results on an annotated dataset of definitional sentences demonstrate the validity of our approach overtaking current state-of-the-art techniques.

3 0.80402666 170 acl-2013-GlossBoot: Bootstrapping Multilingual Domain Glossaries from the Web

Author: Flavio De Benedictis ; Stefano Faralli ; Roberto Navigli

Abstract: We present GlossBoot, an effective minimally-supervised approach to acquiring wide-coverage domain glossaries for many languages. For each language of interest, given a small number of hypernymy relation seeds concerning a target domain, we bootstrap a glossary from the Web for that domain by means of iteratively acquired term/gloss extraction patterns. Our experiments show high performance in the acquisition of domain terminologies and glossaries for three different languages.

4 0.51415259 306 acl-2013-SPred: Large-scale Harvesting of Semantic Predicates

Author: Tiziano Flati ; Roberto Navigli

Abstract: We present SPred, a novel method for the creation of large repositories of semantic predicates. We start from existing collocations to form lexical predicates (e.g., break ∗) and learn the semantic classes that best f∗it) tahned ∗ argument. Taon idco this, we extract failtl thhee ∗ occurrences ion Wikipedia ewxthraiccht match the predicate and abstract its arguments to general semantic classes (e.g., break BODY PART, break AGREEMENT, etc.). Our experiments show that we are able to create a large collection of semantic predicates from the Oxford Advanced Learner’s Dictionary with high precision and recall, and perform well against the most similar approach.

5 0.4114413 61 acl-2013-Automatic Interpretation of the English Possessive

Author: Stephen Tratz ; Eduard Hovy

Abstract: The English ’s possessive construction occurs frequently in text and can encode several different semantic relations; however, it has received limited attention from the computational linguistics community. This paper describes the creation of a semantic relation inventory covering the use of ’s, an inter-annotator agreement study to calculate how well humans can agree on the relations, a large collection of possessives annotated according to the relations, and an accurate automatic annotation system for labeling new examples. Our 21,938 example dataset is by far the largest annotated possessives dataset we are aware of, and both our automatic classification system, which achieves 87.4% accuracy in our classification experiment, and our annotation data are publicly available.

6 0.37769002 198 acl-2013-IndoNet: A Multilingual Lexical Knowledge Network for Indian Languages

7 0.36041835 62 acl-2013-Automatic Term Ambiguity Detection

8 0.35719335 356 acl-2013-Transfer Learning Based Cross-lingual Knowledge Extraction for Wikipedia

9 0.34658411 65 acl-2013-BRAINSUP: Brainstorming Support for Creative Sentence Generation

10 0.34062105 373 acl-2013-Using Conceptual Class Attributes to Characterize Social Media Users

11 0.33166137 154 acl-2013-Extracting bilingual terminologies from comparable corpora

12 0.32822102 242 acl-2013-Mining Equivalent Relations from Linked Data

13 0.31631923 178 acl-2013-HEADY: News headline abstraction through event pattern clustering

14 0.31388751 234 acl-2013-Linking and Extending an Open Multilingual Wordnet

15 0.2999022 365 acl-2013-Understanding Tables in Context Using Standard NLP Toolkits

16 0.29903787 268 acl-2013-PATHS: A System for Accessing Cultural Heritage Collections

17 0.29566535 340 acl-2013-Text-Driven Toponym Resolution using Indirect Supervision

18 0.2917428 215 acl-2013-Large-scale Semantic Parsing via Schema Matching and Lexicon Extension

19 0.28033641 360 acl-2013-Translating Italian connectives into Italian Sign Language

20 0.27972135 367 acl-2013-Universal Conceptual Cognitive Annotation (UCCA)


similar papers computed by lda model

lda for this paper:

topicId topicWeight

[(0, 0.043), (6, 0.024), (11, 0.101), (15, 0.013), (24, 0.041), (26, 0.057), (35, 0.073), (42, 0.036), (48, 0.071), (64, 0.101), (70, 0.044), (85, 0.218), (88, 0.026), (90, 0.01), (95, 0.065)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.78940749 6 acl-2013-A Java Framework for Multilingual Definition and Hypernym Extraction

Author: Stefano Faralli ; Roberto Navigli

Abstract: In this paper we present a demonstration of a multilingual generalization of Word-Class Lattices (WCLs), a supervised lattice-based model used to identify textual definitions and extract hypernyms from them. Lattices are learned from a dataset of automatically-annotated definitions from Wikipedia. We release a Java API for the programmatic use of multilingual WCLs in three languages (English, French and Italian), as well as a Web application for definition and hypernym extraction from user-provided sentences.

2 0.78672051 324 acl-2013-Smatch: an Evaluation Metric for Semantic Feature Structures

Author: Shu Cai ; Kevin Knight

Abstract: The evaluation of whole-sentence semantic structures plays an important role in semantic parsing and large-scale semantic structure annotation. However, there is no widely-used metric to evaluate wholesentence semantic structures. In this paper, we present smatch, a metric that calculates the degree of overlap between two semantic feature structures. We give an efficient algorithm to compute the metric and show the results of an inter-annotator agreement study.

3 0.7620703 133 acl-2013-Efficient Implementation of Beam-Search Incremental Parsers

Author: Yoav Goldberg ; Kai Zhao ; Liang Huang

Abstract: Beam search incremental parsers are accurate, but not as fast as they could be. We demonstrate that, contrary to popular belief, most current implementations of beam parsers in fact run in O(n2), rather than linear time, because each statetransition is actually implemented as an O(n) operation. We present an improved implementation, based on Tree Structured Stack (TSS), in which a transition is performed in O(1), resulting in a real lineartime algorithm, which is verified empiri- cally. We further improve parsing speed by sharing feature-extraction and dotproduct across beam items. Practically, our methods combined offer a speedup of ∼2x over strong baselines on Penn Treeb∼a2nxk sentences, a bnads are eosrd oenrs P eofn magnitude faster on much longer sentences.

4 0.70305932 130 acl-2013-Domain-Specific Coreference Resolution with Lexicalized Features

Author: Nathan Gilbert ; Ellen Riloff

Abstract: Most coreference resolvers rely heavily on string matching, syntactic properties, and semantic attributes of words, but they lack the ability to make decisions based on individual words. In this paper, we explore the benefits of lexicalized features in the setting of domain-specific coreference resolution. We show that adding lexicalized features to off-the-shelf coreference resolvers yields significant performance gains on four domain-specific data sets and with two types of coreference resolution architectures.

5 0.66122663 152 acl-2013-Extracting Definitions and Hypernym Relations relying on Syntactic Dependencies and Support Vector Machines

Author: Guido Boella ; Luigi Di Caro

Abstract: In this paper we present a technique to reveal definitions and hypernym relations from text. Instead of using pattern matching methods that rely on lexico-syntactic patterns, we propose a technique which only uses syntactic dependencies between terms extracted with a syntactic parser. The assumption is that syntactic information are more robust than patterns when coping with length and complexity of the sentences. Afterwards, we transform such syntactic contexts in abstract representations, that are then fed into a Support Vector Machine classifier. The results on an annotated dataset of definitional sentences demonstrate the validity of our approach overtaking current state-of-the-art techniques.

6 0.657929 228 acl-2013-Leveraging Domain-Independent Information in Semantic Parsing

7 0.64125031 22 acl-2013-A Structured Distributional Semantic Model for Event Co-reference

8 0.63764781 268 acl-2013-PATHS: A System for Accessing Cultural Heritage Collections

9 0.63709778 15 acl-2013-A Novel Graph-based Compact Representation of Word Alignment

10 0.61780971 355 acl-2013-TransDoop: A Map-Reduce based Crowdsourced Translation for Complex Domain

11 0.61357778 361 acl-2013-Travatar: A Forest-to-String Machine Translation Engine based on Tree Transducers

12 0.61181492 17 acl-2013-A Random Walk Approach to Selectional Preferences Based on Preference Ranking and Propagation

13 0.60250056 265 acl-2013-Outsourcing FrameNet to the Crowd

14 0.59345061 7 acl-2013-A Lattice-based Framework for Joint Chinese Word Segmentation, POS Tagging and Parsing

15 0.59155995 159 acl-2013-Filling Knowledge Base Gaps for Distant Supervision of Relation Extraction

16 0.59102577 188 acl-2013-Identifying Sentiment Words Using an Optimization-based Model without Seed Words

17 0.59022373 318 acl-2013-Sentiment Relevance

18 0.59019542 215 acl-2013-Large-scale Semantic Parsing via Schema Matching and Lexicon Extension

19 0.58784914 99 acl-2013-Crowd Prefers the Middle Path: A New IAA Metric for Crowdsourcing Reveals Turker Biases in Query Segmentation

20 0.58679992 176 acl-2013-Grounded Unsupervised Semantic Parsing