acl acl2010 acl2010-124 knowledge-graph by maker-knowledge-mining

124 acl-2010-Generating Image Descriptions Using Dependency Relational Patterns


Source: pdf

Author: Ahmet Aker ; Robert Gaizauskas

Abstract: This paper presents a novel approach to automatic captioning of geo-tagged images by summarizing multiple webdocuments that contain information related to an image’s location. The summarizer is biased by dependency pattern models towards sentences which contain features typically provided for different scene types such as those of churches, bridges, etc. Our results show that summaries biased by dependency pattern models lead to significantly higher ROUGE scores than both n-gram language models reported in previous work and also Wikipedia baseline summaries. Summaries generated using dependency patterns also lead to more readable summaries than those generated without dependency patterns.

Reference: text


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 Generating image descriptions using dependency relational patterns Ahmet Aker University of Sheffield a . [sent-1, score-0.847]

2 The summarizer is biased by dependency pattern models towards sentences which contain features typically provided for different scene types such as those of churches, bridges, etc. [sent-6, score-0.471]

3 Our results show that summaries biased by dependency pattern models lead to significantly higher ROUGE scores than both n-gram language models reported in previous work and also Wikipedia baseline summaries. [sent-7, score-0.578]

4 Summaries generated using dependency patterns also lead to more readable summaries than those generated without dependency patterns. [sent-8, score-0.691]

5 This typically small amount of textual information associated with the image is of limited usefulness for image indexing, organization and search. [sent-11, score-0.74]

6 Therefore methods which could automatically supplement the information available for image indexing and lead to improved image retrieval would be extremely useful. [sent-12, score-0.74]

7 Following the general approach proposed by Aker and Gaizauskas (2009), in this paper we describe a method for automatic image captioning or caption enhancement starting with only a scene or subject type and a set of place names pertaining to an image for example hchurch, {St. [sent-13, score-0.906]

8 However, our technique is suitable not only for image captioning but in any application context that requires summary descriptions of instances of object classes, where the instance is to be characterized in terms of the features typically mentioned in describing members of the class. [sent-26, score-0.91]

9 They also experimented with representing such conceptual models using n-gram language models derived from corpora consisting of collections of descriptions of instances of specific object types (e. [sent-31, score-0.336]

10 a corpus of descriptions of churches, a corpus of bridge descriptions, and so on) and reported results showing that incorporating such n-gram language models as a feature in a feature-based extractive summarizer improves the quality of automatically generated summaries. [sent-33, score-0.425]

11 For example one common and important feature of object descriptions is the simple specification of the object type, e. [sent-35, score-0.523]

12 Intuitively, what is important in both these cases is that there is a predication whose subject is the object instance of interest and the head of whose complement is the object type: London Bridge . [sent-49, score-0.346]

13 This intuition suggests that rather than representing object type conceptual models via corpus-derived language models as do Aker and Gaizauskas (2009), we do so instead using corpus-derived dependency patterns. [sent-63, score-0.409]

14 We explore this hypothesis by developing a method for deriving common dependency patterns from object type corpora (Section 2) and then incorporating these patterns into an extractive summarization system (Section 3). [sent-65, score-0.72]

15 In Section 4 we evaluate the approach both by scoring against model summaries and via a readability assessment. [sent-66, score-0.303]

16 Instead, since specific dependency patterns express specific types of inTable 2: Object types and the number of articles in each object type corpus. [sent-71, score-0.52]

17 Object types which are bold are covered by the evaluation image set. [sent-72, score-0.37]

18 1 Object type corpora We derive n-gram language and dependency pattern models using object type corpora made available to us by Aker and Gaizauskas. [sent-76, score-0.547]

19 Aker and Gaizauskas (2009) define an object type corpus as a collection of texts about a specific static object type such as church, bridge, etc. [sent-77, score-0.472]

20 To build such object type corpora the authors categorized Wikipedia articles places by object type. [sent-80, score-0.449]

21 The object type of each article was identified automatically by running Is-A patterns over the first five sentences of the article. [sent-81, score-0.415]

22 2 N-gram language models Aker and Gaizauskas (2009) experimented with uni-gram and bi-gram language models to capture the features commonly used when describing an object type and used these to bias the sentence selection of the summarizer towards the sentences that contain these features. [sent-85, score-0.478]

23 3 Dependency patterns We use the same object type corpora to derive dependency patterns. [sent-93, score-0.52]

24 The final two rows of the table show the output of the Stanford dependency parser and the relational patterns identified for this example. [sent-103, score-0.339]

25 We continue this until we cover all direct relations with built resulting in two more patterns (OBJECTTYPE built DATE and OBJECTTYPE built W). [sent-131, score-0.334]

26 Following these steps we extracted relational patterns for each object type corpus along with the frequency of occurrence of the pattern in the entire corpus. [sent-133, score-0.527]

27 1 Pattern categorization In addition to using dependency patterns as models for biasing sentence selection, we can also use them to control the kind of information to be included in the final summary (see Section 3. [sent-138, score-0.507]

28 We may want to ensure that the summary contains a sentence describing the object type of the object, its location and some background information. [sent-140, score-0.466]

29 To be able to do so, we categorize dependency patterns according to the type of information they express. [sent-142, score-0.382]

30 We manually analyzed human written descriptions about instances of different object types and recorded for each sentence in the descriptions the kind of information it contained about the object. [sent-143, score-0.495]

31 We analyzed descriptions of 3 10 different objects where each object had up to four different human written descriptions (Section 4. [sent-144, score-0.474]

32 We also manually assigned each dependency pattern in each corpus-derived model to one of the above categories, provided it occurred five or more times in the object type corpora. [sent-149, score-0.484]

33 The patterns extracted for our example sentence shown in Table 3, for instance, are all categorized by year category because all of them contain information about the foundation date of an object. [sent-150, score-0.286]

34 3 Summarizer We adopted the same overall approach to summarization used by Aker and Gaizauskas to generate the image descriptions. [sent-151, score-0.41]

35 It is given two inputs: a toponym associated with an image and a set of documents to be summarized which have been retrieved from the web using the toponym as a query. [sent-153, score-0.486]

36 The summarizer creates image descriptions in a three step process. [sent-154, score-0.63]

37 In our experiments we extend this feature set by two dependency pattern related features: DpMSim and DepCat. [sent-174, score-0.287]

38 To compute this score, we first parse the sentence on the fly with the Stanford parser and obtain the dependency patterns for the sentence. [sent-177, score-0.33]

39 We then associate each dependency pattern of the sentence with the occurrence frequency of that pattern in the dependency pattern model (DpM). [sent-178, score-0.642]

40 It is a sum of all occurrence frequencies of the dependency patterns detected in a sentence S that are also contained in the DpM. [sent-180, score-0.33]

41 DpMSim(S,DpM) =XfDpM(p) (1) pX∈ XS The second feature, DepCat, uses dependency patterns to categorize the sentences rather than ranking them. [sent-181, score-0.362]

42 To do this, we obtain the relational patterns for the current sentence, check whether for each such pattern whether it is included in the DpM, and, if so, we add to the sentence the category the pattern was manually associated with. [sent-185, score-0.437]

43 We used 32 of the 3 10 images from our image set (see Section 4. [sent-202, score-0.507]

44 The image descriptions from this data set are used as model summaries. [sent-204, score-0.508]

45 Our training data contains for each image a set of image descriptions taken from the VirtualTourist travel community web-site 4. [sent-205, score-0.878]

46 From this web-site we took all existing image descriptions about a particular image or object. [sent-206, score-0.878]

47 Note that some of these descriptions about a particular object were used to derive the model summaries for that object (see Section 4. [sent-207, score-0.743]

48 Assuming that model summaries contain the most relevant sentences about an object we perform ROUGE comparisons between the sentences in all the image descriptions and the model summaries, i. [sent-209, score-1.026]

49 we pair each sentence from all image descriptions about a particular place with every sentence from all the model 4www. [sent-211, score-0.6]

50 In this way, we have for each sentence from all existing image de- scriptions about an object a ROUGE score5 indicating its relevance. [sent-215, score-0.589]

51 1 Data sets For evaluation we use the image collection de- scribed in Aker and Gaizauskas (2010). [sent-225, score-0.37]

52 The image collection contains 3 10 different images with manually assigned toponyms. [sent-226, score-0.507]

53 The images cover 60 of the 107 object types identified from Wikipedia (see Table 2). [sent-227, score-0.31]

54 For each image there are up to four short descriptions or model summaries. [sent-228, score-0.508]

55 The model summaries were created manually based on image descriptions taken from VirtualTourist and contain a minimum of 190 and a maximum of 210 words. [sent-229, score-0.767]

56 of this image collection was used to train the weights and the remaining (105 images) for evaluation. [sent-231, score-0.37]

57 To generate automatic captions for the images we automatically retrieved the top 30 related web-documents for each image using the Yahoo! [sent-232, score-0.545]

58 search engine and the toponym associated with the image as a query. [sent-233, score-0.428]

59 Named after its designer, engineer Gustave Eiffel, the tower was built as the entrance arch for the 1889 World’s Fair. [sent-250, score-0.352]

60 Although it was the world’s tallest structure when completed in 1889, the Eiffel Tower has since lost its standing both as the tallest lattice tower and as the tallest structure in France. [sent-260, score-0.46]

61 The tower has two restaurants: Altitude 95, on the first floor 311ft (95m) above sea level; and the Jules Verne, an expensive gastronomical restaurant on the second floor, with a private lift. [sent-261, score-0.302]

62 2 ROUGE assessment In the first assessment we compared the automatically generated summaries against model summaries written by humans using ROUGE (Lin, 2004). [sent-273, score-0.604]

63 ROUGE 2 gives recall scores for bi-gram overlap between the automatically generated summaries and the reference ones. [sent-275, score-0.303]

64 Firstly, we generated summaries for each image using the top-ranked non Wikipedia document retrieved in the Yahoo! [sent-278, score-0.655]

65 From this document we create a baseline summary by selecting sentences from the beginning until the summary reaches a length of 200 words. [sent-280, score-0.321]

66 Secondly, we separately ran the summarizer over the top ten documents for each single feature and compared the automated summaries against the model ones. [sent-290, score-0.449]

67 Table 5 shows that the dependency model feature (DpMSim) contributes most to the summary quality according to the ROUGE metrics. [sent-292, score-0.313]

68 To see how the ROUGE scores change when features are combined with each other we performed different combinations of the features, ran the summarizer for each combination and compared the automated summaries against the model ones. [sent-297, score-0.454]

69 1255 Table 6: ROUGE scores of feature combinations which score moderately or significantly higher than dependency pattern model (DpMSim) feature and Wikipedia baseline. [sent-303, score-0.4]

70 W1049i7k also included the dependency pattern categorization (DepCat) feature explained in Section 3. [sent-308, score-0.338]

71 Table 6 shows the results of feature combinations which score moderately or significantly higher than the dependency pattern model (DpMSim) feature score shown in Table 5. [sent-310, score-0.386]

72 The summaries categorized by dependency patterns (starterSimilarity+LMSim+DepCat) achieve significantly higher ROUGE scores than the Wikipedia baseline. [sent-312, score-0.627]

73 It can be seen that this combination without the dependency patterns lead to lower ROUGE scores in ROUGE 2 and only moderate improvement in ROUGE SU4 if compared with Wikipedia baseline ROUGE scores. [sent-317, score-0.328]

74 3 Readability assessment We also evaluated our summaries using a readability assessment as in DUC and TAC. [sent-319, score-0.389]

75 DUC and TAC manually assess the quality of automatically generated summaries by asking human subjects to score each summary using five criteria grammaticality, redundancy, clarity, focus and structure criteria. [sent-320, score-0.415]

76 For comparison we also evaluated summaries which were not structured by dependency patterns (starterSimilarity + LMSim) and also the Wikipedia baseline summaries. [sent-324, score-0.543]

77 Each person was shown all 3 15 summaries (105 from each summary type) in a random way and was asked to assess them according to the DUC and TAC manual assessment scheme. [sent-326, score-0.428]

78 We see from Table 7 that using dependency patterns to categorize the sentences and produce a structured summary helps to obtain better readable summaries. [sent-328, score-0.488]

79 The scores of our automated summaries were better than the Wikipedia baseline summaries in the grammar feature. [sent-330, score-0.591]

80 However, in other features the Wikipedia baseline summaries obtained better scores than our automated summaries. [sent-331, score-0.332]

81 5 Related Work Our approach has an advantage over related work in automatic image captioning in that it requires only GPS information associated with the image in order to generate captions. [sent-333, score-0.812]

82 Other attempts towards automatic generation of image captions generate captions based on the immediate textual context of the image with or without consideration of image related features such as colour, shape or texture (Deschacht and Moens, 2007; Mori et al. [sent-334, score-1.186]

83 However, Marsch & White (2003) argue that the content of an image and its immediate text have little semantic agreement and this can, according to Purves et al. [sent-341, score-0.37]

84 Furthermore, these approaches assume that the image has been obtained from a document. [sent-343, score-0.37]

85 However, dependency patterns have not been used extensively in summarization tasks. [sent-370, score-0.324]

86 (2002) who used dependency patterns in combination with other features to generate extracts in a single document summarization task. [sent-372, score-0.35]

87 The authors found that when learning weights in a simple feature weigthing scheme, the weight assigned to dependency patterns was lower than that assigned to other features. [sent-373, score-0.323]

88 The small contribution of the dependency patterns may have been due to the small number of documents they used to derive their dependency patterns they gathered dependency patterns from only ten domain specific documents which are unlikely to be sufficient to capture repeated features in a domain. [sent-374, score-0.852]

89 – 6 Discussion and Conclusion We have proposed a method by which dependency patterns extracted from corpora of descriptions of instances ofparticular object types can be used in a multi-document summarizer to automatically generate image descriptions. [sent-375, score-1.087]

90 Our evaluations show that such an approach yields summaries which score more highly than an approach which uses a simpler representation of an object type model in the form of a n-gram language model. [sent-376, score-0.525]

91 When used as the sole feature for sentence ranking, dependency pattern models (DpMSim) produced summaries with higher ROUGE scores than those obtained using the features reported in Aker and Gaizauskas (2009). [sent-377, score-0.636]

92 Furthermore, we showed that using dependency patterns in combination with features reported in Aker and Gaizauskas to produce a structured summary led to significantly better results than Wikipedia baseline summaries as assessed by ROUGE. [sent-379, score-0.669]

93 These results indicate that dependency patterns are worth investigating for object focused auto- mated summarization tasks. [sent-381, score-0.497]

94 Such investigations should in particular concentrate on how dependency patterns can be used to structure information within the summary, as our best results were achieved when dependency patterns were used for this purpose. [sent-382, score-0.568]

95 One is to explore how dependency patterns could be used to produce generative summaries and/or perform sentence trimming. [sent-384, score-0.589]

96 Another is to investigate how dependency patterns might be automatically clustered into groups expressing similar or related facts, rather than relying on manual categorization of dependency patterns into categories such as “type”, “year”, etc. [sent-385, score-0.652]

97 Evaluation should be extended to investigate the utility of the automatically generated image descriptions for image retrieval. [sent-387, score-0.878]

98 what is the flow of facts to describe a location) from existing image descriptions to produce better summaries. [sent-390, score-0.508]

99 Automatic word assignment to images based on image division and vector quantization. [sent-519, score-0.507]

100 Describing the where–improving image annotation and search through geography. [sent-547, score-0.37]


similar papers computed by tfidf model

tfidf for this paper:

wordName wordTfidf (topN-words)

[('image', 0.37), ('aker', 0.303), ('rouge', 0.296), ('summaries', 0.259), ('tower', 0.232), ('eiffel', 0.202), ('gaizauskas', 0.173), ('lmsim', 0.173), ('objecttype', 0.173), ('object', 0.173), ('depcat', 0.159), ('dpmsim', 0.159), ('dependency', 0.148), ('descriptions', 0.138), ('images', 0.137), ('patterns', 0.136), ('startersimilarity', 0.13), ('summary', 0.126), ('summarizer', 0.122), ('bridge', 0.102), ('pattern', 0.1), ('wikipedia', 0.099), ('tallest', 0.076), ('captioning', 0.072), ('built', 0.066), ('barnard', 0.063), ('type', 0.063), ('duc', 0.061), ('rhine', 0.058), ('toponym', 0.058), ('virtualtourist', 0.058), ('relational', 0.055), ('categorization', 0.051), ('duygulu', 0.051), ('multimedia', 0.046), ('abbey', 0.046), ('gps', 0.046), ('sentence', 0.046), ('scores', 0.044), ('readability', 0.044), ('assessment', 0.043), ('champ', 0.043), ('gustave', 0.043), ('monument', 0.043), ('sentences', 0.043), ('categorized', 0.04), ('summarization', 0.04), ('feature', 0.039), ('ascend', 0.038), ('captions', 0.038), ('dpm', 0.038), ('floor', 0.038), ('sudo', 0.038), ('stevenson', 0.038), ('categorize', 0.035), ('berg', 0.035), ('paris', 0.033), ('date', 0.033), ('categories', 0.033), ('restaurant', 0.032), ('scene', 0.031), ('describing', 0.031), ('year', 0.031), ('score', 0.03), ('reached', 0.03), ('indirect', 0.029), ('altitude', 0.029), ('arch', 0.029), ('ecole', 0.029), ('jules', 0.029), ('kurtic', 0.029), ('militaire', 0.029), ('mori', 0.029), ('nobata', 0.029), ('purves', 0.029), ('railway', 0.029), ('satoh', 0.029), ('shop', 0.029), ('slmd', 0.029), ('stairs', 0.029), ('verne', 0.029), ('world', 0.029), ('automated', 0.029), ('biased', 0.027), ('location', 0.027), ('visited', 0.026), ('visiting', 0.026), ('document', 0.026), ('located', 0.026), ('conceptual', 0.025), ('rivers', 0.025), ('westminster', 0.025), ('tour', 0.025), ('churches', 0.025), ('entrance', 0.025), ('freitas', 0.025), ('greenwood', 0.025), ('yangarber', 0.025), ('objects', 0.025), ('extractive', 0.024)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.99999982 124 acl-2010-Generating Image Descriptions Using Dependency Relational Patterns

Author: Ahmet Aker ; Robert Gaizauskas

Abstract: This paper presents a novel approach to automatic captioning of geo-tagged images by summarizing multiple webdocuments that contain information related to an image’s location. The summarizer is biased by dependency pattern models towards sentences which contain features typically provided for different scene types such as those of churches, bridges, etc. Our results show that summaries biased by dependency pattern models lead to significantly higher ROUGE scores than both n-gram language models reported in previous work and also Wikipedia baseline summaries. Summaries generated using dependency patterns also lead to more readable summaries than those generated without dependency patterns.

2 0.27514613 136 acl-2010-How Many Words Is a Picture Worth? Automatic Caption Generation for News Images

Author: Yansong Feng ; Mirella Lapata

Abstract: In this paper we tackle the problem of automatic caption generation for news images. Our approach leverages the vast resource of pictures available on the web and the fact that many of them are captioned. Inspired by recent work in summarization, we propose extractive and abstractive caption generation models. They both operate over the output of a probabilistic image annotation model that preprocesses the pictures and suggests keywords to describe their content. Experimental results show that an abstractive model defined over phrases is superior to extractive methods.

3 0.2050204 38 acl-2010-Automatic Evaluation of Linguistic Quality in Multi-Document Summarization

Author: Emily Pitler ; Annie Louis ; Ani Nenkova

Abstract: To date, few attempts have been made to develop and validate methods for automatic evaluation of linguistic quality in text summarization. We present the first systematic assessment of several diverse classes of metrics designed to capture various aspects of well-written text. We train and test linguistic quality models on consecutive years of NIST evaluation data in order to show the generality of results. For grammaticality, the best results come from a set of syntactic features. Focus, coherence and referential clarity are best evaluated by a class of features measuring local coherence on the basis of cosine similarity between sentences, coreference informa- tion, and summarization specific features. Our best results are 90% accuracy for pairwise comparisons of competing systems over a test set of several inputs and 70% for ranking summaries of a specific input.

4 0.18788332 264 acl-2010-Wrapping up a Summary: From Representation to Generation

Author: Josef Steinberger ; Marco Turchi ; Mijail Kabadjov ; Ralf Steinberger ; Nello Cristianini

Abstract: The main focus of this work is to investigate robust ways for generating summaries from summary representations without recurring to simple sentence extraction and aiming at more human-like summaries. This is motivated by empirical evidence from TAC 2009 data showing that human summaries contain on average more and shorter sentences than the system summaries. We report encouraging preliminary results comparable to those attained by participating systems at TAC 2009.

5 0.18276834 188 acl-2010-Optimizing Informativeness and Readability for Sentiment Summarization

Author: Hitoshi Nishikawa ; Takaaki Hasegawa ; Yoshihiro Matsuo ; Genichiro Kikui

Abstract: We propose a novel algorithm for sentiment summarization that takes account of informativeness and readability, simultaneously. Our algorithm generates a summary by selecting and ordering sentences taken from multiple review texts according to two scores that represent the informativeness and readability of the sentence order. The informativeness score is defined by the number of sentiment expressions and the readability score is learned from the target corpus. We evaluate our method by summarizing reviews on restaurants. Our method outperforms an existing algorithm as indicated by its ROUGE score and human readability experiments.

6 0.16964795 125 acl-2010-Generating Templates of Entity Summaries with an Entity-Aspect Model and Pattern Mining

7 0.16084477 77 acl-2010-Cross-Language Document Summarization Based on Machine Translation Quality Prediction

8 0.14517522 8 acl-2010-A Hybrid Hierarchical Model for Multi-Document Summarization

9 0.1375386 11 acl-2010-A New Approach to Improving Multilingual Summarization Using a Genetic Algorithm

10 0.12681915 14 acl-2010-A Risk Minimization Framework for Extractive Speech Summarization

11 0.095351569 39 acl-2010-Automatic Generation of Story Highlights

12 0.090273321 113 acl-2010-Extraction and Approximation of Numerical Attributes from the Web

13 0.089543365 261 acl-2010-Wikipedia as Sense Inventory to Improve Diversity in Web Search Results

14 0.081546031 122 acl-2010-Generating Fine-Grained Reviews of Songs from Album Reviews

15 0.076071657 171 acl-2010-Metadata-Aware Measures for Answer Summarization in Community Question Answering

16 0.074917912 166 acl-2010-Learning Word-Class Lattices for Definition and Hypernym Extraction

17 0.069865644 83 acl-2010-Dependency Parsing and Projection Based on Word-Pair Classification

18 0.065497436 143 acl-2010-Importance of Linguistic Constraints in Statistical Dependency Parsing

19 0.06483268 185 acl-2010-Open Information Extraction Using Wikipedia

20 0.063755862 48 acl-2010-Better Filtration and Augmentation for Hierarchical Phrase-Based Translation Rules


similar papers computed by lsi model

lsi for this paper:

topicId topicWeight

[(0, -0.182), (1, 0.074), (2, -0.121), (3, 0.007), (4, 0.016), (5, -0.024), (6, 0.079), (7, -0.338), (8, -0.078), (9, 0.028), (10, -0.032), (11, -0.074), (12, -0.174), (13, 0.088), (14, 0.068), (15, -0.032), (16, 0.022), (17, 0.13), (18, 0.021), (19, 0.061), (20, 0.028), (21, 0.054), (22, -0.009), (23, 0.005), (24, 0.035), (25, 0.028), (26, -0.052), (27, -0.039), (28, -0.026), (29, -0.066), (30, -0.02), (31, 0.042), (32, -0.004), (33, -0.04), (34, 0.007), (35, -0.071), (36, 0.058), (37, 0.01), (38, 0.17), (39, 0.036), (40, -0.058), (41, 0.198), (42, -0.015), (43, 0.105), (44, -0.127), (45, 0.036), (46, -0.008), (47, -0.178), (48, 0.01), (49, 0.007)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.93727386 124 acl-2010-Generating Image Descriptions Using Dependency Relational Patterns

Author: Ahmet Aker ; Robert Gaizauskas

Abstract: This paper presents a novel approach to automatic captioning of geo-tagged images by summarizing multiple webdocuments that contain information related to an image’s location. The summarizer is biased by dependency pattern models towards sentences which contain features typically provided for different scene types such as those of churches, bridges, etc. Our results show that summaries biased by dependency pattern models lead to significantly higher ROUGE scores than both n-gram language models reported in previous work and also Wikipedia baseline summaries. Summaries generated using dependency patterns also lead to more readable summaries than those generated without dependency patterns.

2 0.70071739 136 acl-2010-How Many Words Is a Picture Worth? Automatic Caption Generation for News Images

Author: Yansong Feng ; Mirella Lapata

Abstract: In this paper we tackle the problem of automatic caption generation for news images. Our approach leverages the vast resource of pictures available on the web and the fact that many of them are captioned. Inspired by recent work in summarization, we propose extractive and abstractive caption generation models. They both operate over the output of a probabilistic image annotation model that preprocesses the pictures and suggests keywords to describe their content. Experimental results show that an abstractive model defined over phrases is superior to extractive methods.

3 0.68127435 125 acl-2010-Generating Templates of Entity Summaries with an Entity-Aspect Model and Pattern Mining

Author: Peng Li ; Jing Jiang ; Yinglin Wang

Abstract: In this paper, we propose a novel approach to automatic generation of summary templates from given collections of summary articles. This kind of summary templates can be useful in various applications. We first develop an entity-aspect LDA model to simultaneously cluster both sentences and words into aspects. We then apply frequent subtree pattern mining on the dependency parse trees of the clustered and labeled sentences to discover sentence patterns that well represent the aspects. Key features of our method include automatic grouping of semantically related sentence patterns and automatic identification of template slots that need to be filled in. We apply our method on five Wikipedia entity categories and compare our method with two baseline methods. Both quantitative evaluation based on human judgment and qualitative comparison demonstrate the effectiveness and advantages of our method.

4 0.60354775 38 acl-2010-Automatic Evaluation of Linguistic Quality in Multi-Document Summarization

Author: Emily Pitler ; Annie Louis ; Ani Nenkova

Abstract: To date, few attempts have been made to develop and validate methods for automatic evaluation of linguistic quality in text summarization. We present the first systematic assessment of several diverse classes of metrics designed to capture various aspects of well-written text. We train and test linguistic quality models on consecutive years of NIST evaluation data in order to show the generality of results. For grammaticality, the best results come from a set of syntactic features. Focus, coherence and referential clarity are best evaluated by a class of features measuring local coherence on the basis of cosine similarity between sentences, coreference informa- tion, and summarization specific features. Our best results are 90% accuracy for pairwise comparisons of competing systems over a test set of several inputs and 70% for ranking summaries of a specific input.

5 0.60142595 264 acl-2010-Wrapping up a Summary: From Representation to Generation

Author: Josef Steinberger ; Marco Turchi ; Mijail Kabadjov ; Ralf Steinberger ; Nello Cristianini

Abstract: The main focus of this work is to investigate robust ways for generating summaries from summary representations without recurring to simple sentence extraction and aiming at more human-like summaries. This is motivated by empirical evidence from TAC 2009 data showing that human summaries contain on average more and shorter sentences than the system summaries. We report encouraging preliminary results comparable to those attained by participating systems at TAC 2009.

6 0.58320481 8 acl-2010-A Hybrid Hierarchical Model for Multi-Document Summarization

7 0.5653879 188 acl-2010-Optimizing Informativeness and Readability for Sentiment Summarization

8 0.55765444 122 acl-2010-Generating Fine-Grained Reviews of Songs from Album Reviews

9 0.50391304 14 acl-2010-A Risk Minimization Framework for Extractive Speech Summarization

10 0.46447024 77 acl-2010-Cross-Language Document Summarization Based on Machine Translation Quality Prediction

11 0.45549601 166 acl-2010-Learning Word-Class Lattices for Definition and Hypernym Extraction

12 0.45166329 11 acl-2010-A New Approach to Improving Multilingual Summarization Using a Genetic Algorithm

13 0.42989573 39 acl-2010-Automatic Generation of Story Highlights

14 0.39425302 157 acl-2010-Last but Definitely Not Least: On the Role of the Last Sentence in Automatic Polarity-Classification

15 0.36117706 185 acl-2010-Open Information Extraction Using Wikipedia

16 0.35594609 140 acl-2010-Identifying Non-Explicit Citing Sentences for Citation-Based Summarization.

17 0.34540406 64 acl-2010-Complexity Assumptions in Ontology Verbalisation

18 0.34232751 113 acl-2010-Extraction and Approximation of Numerical Attributes from the Web

19 0.33583945 138 acl-2010-Hunting for the Black Swan: Risk Mining from Text

20 0.32766789 63 acl-2010-Comparable Entity Mining from Comparative Questions


similar papers computed by lda model

lda for this paper:

topicId topicWeight

[(25, 0.053), (33, 0.012), (39, 0.01), (42, 0.024), (44, 0.021), (59, 0.093), (61, 0.278), (72, 0.014), (73, 0.031), (76, 0.02), (78, 0.016), (80, 0.011), (83, 0.114), (84, 0.044), (98, 0.153)]

similar papers list:

simIndex simValue paperId paperTitle

1 0.77643692 121 acl-2010-Generating Entailment Rules from FrameNet

Author: Roni Ben Aharon ; Idan Szpektor ; Ido Dagan

Abstract: Idan Szpektor Ido Dagan Yahoo! Research Department of Computer Science Haifa, Israel Bar-Ilan University idan @ yahoo- inc .com Ramat Gan, Israel dagan @ c s .biu . ac . i l FrameNet is a manually constructed database based on Frame Semantics. It models the semantic Many NLP tasks need accurate knowledge for semantic inference. To this end, mostly WordNet is utilized. Yet WordNet is limited, especially for inference be- tween predicates. To help filling this gap, we present an algorithm that generates inference rules between predicates from FrameNet. Our experiment shows that the novel resource is effective and complements WordNet in terms of rule coverage.

same-paper 2 0.77595186 124 acl-2010-Generating Image Descriptions Using Dependency Relational Patterns

Author: Ahmet Aker ; Robert Gaizauskas

Abstract: This paper presents a novel approach to automatic captioning of geo-tagged images by summarizing multiple webdocuments that contain information related to an image’s location. The summarizer is biased by dependency pattern models towards sentences which contain features typically provided for different scene types such as those of churches, bridges, etc. Our results show that summaries biased by dependency pattern models lead to significantly higher ROUGE scores than both n-gram language models reported in previous work and also Wikipedia baseline summaries. Summaries generated using dependency patterns also lead to more readable summaries than those generated without dependency patterns.

3 0.69506872 146 acl-2010-Improving Chinese Semantic Role Labeling with Rich Syntactic Features

Author: Weiwei Sun

Abstract: Developing features has been shown crucial to advancing the state-of-the-art in Semantic Role Labeling (SRL). To improve Chinese SRL, we propose a set of additional features, some of which are designed to better capture structural information. Our system achieves 93.49 Fmeasure, a significant improvement over the best reported performance 92.0. We are further concerned with the effect of parsing in Chinese SRL. We empirically analyze the two-fold effect, grouping words into constituents and providing syntactic information. We also give some preliminary linguistic explanations.

4 0.61290443 133 acl-2010-Hierarchical Search for Word Alignment

Author: Jason Riesa ; Daniel Marcu

Abstract: We present a simple yet powerful hierarchical search algorithm for automatic word alignment. Our algorithm induces a forest of alignments from which we can efficiently extract a ranked k-best list. We score a given alignment within the forest with a flexible, linear discriminative model incorporating hundreds of features, and trained on a relatively small amount of annotated data. We report results on Arabic-English word alignment and translation tasks. Our model outperforms a GIZA++ Model-4 baseline by 6.3 points in F-measure, yielding a 1.1 BLEU score increase over a state-of-the-art syntax-based machine translation system.

5 0.61280537 211 acl-2010-Simple, Accurate Parsing with an All-Fragments Grammar

Author: Mohit Bansal ; Dan Klein

Abstract: We present a simple but accurate parser which exploits both large tree fragments and symbol refinement. We parse with all fragments of the training set, in contrast to much recent work on tree selection in data-oriented parsing and treesubstitution grammar learning. We require only simple, deterministic grammar symbol refinement, in contrast to recent work on latent symbol refinement. Moreover, our parser requires no explicit lexicon machinery, instead parsing input sentences as character streams. Despite its simplicity, our parser achieves accuracies of over 88% F1 on the standard English WSJ task, which is competitive with substantially more complicated state-of-theart lexicalized and latent-variable parsers. Additional specific contributions center on making implicit all-fragments parsing efficient, including a coarse-to-fine inference scheme and a new graph encoding.

6 0.61234546 51 acl-2010-Bilingual Sense Similarity for Statistical Machine Translation

7 0.61153114 39 acl-2010-Automatic Generation of Story Highlights

8 0.61092019 109 acl-2010-Experiments in Graph-Based Semi-Supervised Learning Methods for Class-Instance Acquisition

9 0.61040366 136 acl-2010-How Many Words Is a Picture Worth? Automatic Caption Generation for News Images

10 0.60994017 184 acl-2010-Open-Domain Semantic Role Labeling by Modeling Word Spans

11 0.60983366 195 acl-2010-Phylogenetic Grammar Induction

12 0.60936165 218 acl-2010-Structural Semantic Relatedness: A Knowledge-Based Method to Named Entity Disambiguation

13 0.60935163 55 acl-2010-Bootstrapping Semantic Analyzers from Non-Contradictory Texts

14 0.60847592 245 acl-2010-Understanding the Semantic Structure of Noun Phrase Queries

15 0.60799354 153 acl-2010-Joint Syntactic and Semantic Parsing of Chinese

16 0.60744655 102 acl-2010-Error Detection for Statistical Machine Translation Using Linguistic Features

17 0.6074332 261 acl-2010-Wikipedia as Sense Inventory to Improve Diversity in Web Search Results

18 0.60723889 93 acl-2010-Dynamic Programming for Linear-Time Incremental Parsing

19 0.6065619 145 acl-2010-Improving Arabic-to-English Statistical Machine Translation by Reordering Post-Verbal Subjects for Alignment

20 0.60654438 101 acl-2010-Entity-Based Local Coherence Modelling Using Topological Fields