acl acl2010 acl2010-264 knowledge-graph by maker-knowledge-mining

264 acl-2010-Wrapping up a Summary: From Representation to Generation


Source: pdf

Author: Josef Steinberger ; Marco Turchi ; Mijail Kabadjov ; Ralf Steinberger ; Nello Cristianini

Abstract: The main focus of this work is to investigate robust ways for generating summaries from summary representations without recurring to simple sentence extraction and aiming at more human-like summaries. This is motivated by empirical evidence from TAC 2009 data showing that human summaries contain on average more and shorter sentences than the system summaries. We report encouraging preliminary results comparable to those attained by participating systems at TAC 2009.

Reference: text


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 eu Abstract The main focus of this work is to investigate robust ways for generating summaries from summary representations without recurring to simple sentence extraction and aiming at more human-like summaries. [sent-8, score-0.728]

2 This is motivated by empirical evidence from TAC 2009 data showing that human summaries contain on average more and shorter sentences than the system summaries. [sent-9, score-0.441]

3 We report encouraging preliminary results comparable to those attained by participating systems at TAC 2009. [sent-10, score-0.052]

4 The advantages of looking at the summarization problem in terms of distinct processing phases are numerous. [sent-12, score-0.237]

5 It not only serves as a common ground for comparing different systems and understanding better the underlying logic and assumptions, but it also provides a neat framework for developing systems based on clean and extendable designs. [sent-13, score-0.042]

6 (2007) showed that solely by enhancing the first source interpretation phase, one is already able to pro– – duce better summaries. [sent-16, score-0.118]

7 There has been limited work on the last summary generation phase due to the fact that it is unarguably a very challenging problem. [sent-17, score-0.347]

8 net l amount of approaches assume simple sentence selection, a type of extractive summarization, where often the summary representation and the end summary are, indeed, conflated. [sent-19, score-0.513]

9 The main focus of this work is, thus, to investigate robust ways for generating summaries from summary representations without recurring to simple sentence extraction and aiming at more human-like summaries. [sent-20, score-0.728]

10 This decision is also motivated by empirical evidence from TAC 2009 data (see table 1) showing that human summaries contain on average more and shorter sentences than the system summaries. [sent-21, score-0.441]

11 The intuition behind this is that, by containing more sentences, a summary is able to capture more of the important content from the source. [sent-22, score-0.192]

12 Our initial experimental results show that our approach is feasible, since it produces summaries, which when evaluated against the TAC 2009 data1 yield ROUGE scores (Lin and Hovy, 2003) comparable to the participating systems in the Summarization task at TAC 2009. [sent-23, score-0.103]

13 2 Related Work There is a large body of literature on summarization (Hovy, 2005; Erkan and Radev, 2004; Kupiec et al. [sent-26, score-0.201]

14 The most closely related work to the approach presented hereby is work on summarization attempting to go beyond simple sentence ex1http : / /www . [sent-28, score-0.297]

15 c C2o0n1f0er Aenscseoc Sihatoirotn P faopre Crso,m papguetsat 3io8n2a–l3 L8i6n,guistics traction and to a lesser degree work on sentence compression. [sent-32, score-0.096]

16 Although our approach is related to sentence compression (Knight and Marcu, 2002; Clarke and Lapata, 2008), it is subtly different. [sent-34, score-0.159]

17 Firstly, we reduce the number of terms to be used in the summary at a global level, not at a local per-sentence level. [sent-35, score-0.228]

18 Secondly, we directly exploit the resulting structures from the SVD making the last generation step fully aware ofprevious processing stages, as opposed to tackling the problem of sentence compression in isolation. [sent-36, score-0.215]

19 A similar approach to our sentence reconstruction method has been developed by Quirk et al. [sent-37, score-0.448]

20 In their work, training and test sets contain sentence pairs that are composed of two different proper English sentences and a paraphrase of a source sentence is generated by finding the optimal path through a paraphrases lattice. [sent-39, score-0.342]

21 3 Three-fold Summarization: Interpretation, Transformation and Generation We chose the LSA paradigm for summarization, since it provides a clear and direct instantiation of Sp¨ arck-Jones’ three-stage framework. [sent-41, score-0.043]

22 In LSA-based summarization the interpretation phase takes the form of building a term-bysentence matrix A = [A1, A2, . [sent-42, score-0.45]

23 , anj]T represents the weighted term-frequency vector of sentence j in a given set of documents. [sent-48, score-0.096]

24 We adopt the same weighting scheme as the one described in (Steinberger et al. [sent-49, score-0.033]

25 , 2007), as well as their more general definition of term entailing not only unigrams and bigrams, but also named entities. [sent-50, score-0.116]

26 The transformation phase is done by applying singular value decomposition (SVD) to the initial term-by-sentence matrix defined as A = UΣVT. [sent-51, score-0.362]

27 The generation phase is where our main contribution comes in. [sent-52, score-0.155]

28 At this point we depart from standard LSA-based approaches and aim at producing a succinct summary representation comprised only of salient terms Term Summary Representation (TSR). [sent-53, score-0.341]

29 Then this TSR is passed on to another module which attempts to produce complete sentences. [sent-54, score-0.041]

30 The module for sentence reconstruction is described in detail in section 4, in what follows we explain the method for producing a TSR. [sent-55, score-0.566]

31 1 Term Summary Representation To explain how a term summary representation (TSR) is produced, we first need to define two concepts: salience score of a given term and salience threshold. [sent-57, score-0.753]

32 Salience score for each term in matrix A is given by the magnitude of the corresponding vector in the matrix resulting from the dot product of the matrix of left singular vectors with the diagonal matrix of singular values. [sent-58, score-0.649]

33 More formally, let T = U · Σ and then for each term i, the salience score Uis given by |. [sent-59, score-0.247]

34 f S Sthalei top k ththr term, w ish eeqn uaall terms are sorted in descending order on the basis of their salience scores and a cutoff is defined as a percentage (e. [sent-61, score-0.396]

35 In other words, if the total number of terms is n, then 100 ∗ k/n must be equal utom mtbhee percentage c nu,to thfef specified. [sent-64, score-0.108]

36 The generation of a TSR is performed in two steps. [sent-65, score-0.056]

37 First, an initial pool of sentences is selected by using the same technique as in (Steinberger and Je˘ zek, 2009) which exploits the dot product of the diagonal matrix of singular values with the right singular vectors: Σ · VT. [sent-66, score-0.496]

38 2 This initial pool of sentseinncgeusla irs vthecet output ·oVf standard LSA approaches. [sent-67, score-0.126]

39 Second, the terms from the source matrix A are identified in the initial pool of sentences and those terms whose salience score is above the salience h|Te~ni threshold are copied across to the TSR. [sent-68, score-0.747]

40 Thus, the TSR is formed by the most (globally) salient terms from each one of the sentences. [sent-69, score-0.036]

41 For example: • Extracted Sentence: “Irish Prime Minister Bertie Ahern admitted on Tuesday that he had held a series of private one-on-one meetings on the Northern Ireland peace process with Sinn Fein leader Gerry Adams, but denied they had been secret in any way. [sent-70, score-0.153]

42 ” • TSR Sentence at 10%: “Irish Prime Minister Bertie Ahern Tuesday had held one-on-one meetings Northern Ireland peace process Sinn Fein leader Gerry Adams”3 2Due to space constraints, full details on that step are omitted here, see (Steinberger and Je˘ zek, 2009). [sent-71, score-0.153]

43 3The TSR sentence is stemmed just before feeding it to the reconstruction module discussed in the next section. [sent-72, score-0.602]

44 4 Noisy-channel model for sentence reconstruction This section describes a probabilistic approach to the reconstruction problem. [sent-96, score-0.8]

45 We adopt the noisychannel framework that has been widely used in a number of other NLP applications. [sent-97, score-0.094]

46 Our interpretation of the noisy channel consists of looking at a stemmed string without stopwords and imagining that it was originally a long string and that someone removed or stemmed some text from it. [sent-98, score-0.405]

47 In our framework, reconstruction consists of identifying the original long string. [sent-99, score-0.352]

48 To model our interpretation of the noisy channel, we make use of one of the most popular classes of SMT systems: the Phrase Based Model (PBM) (Zens et al. [sent-100, score-0.064]

49 It is an extension of the noisychannel model and was introduced by Brown et al. [sent-103, score-0.061]

50 In PBM, a source sentence f is segmented into a sequence of I phrases fI = [f1, f2, . [sent-105, score-0.15]

51 fI] and the same is done for the target sentence e, where the notion of phrase is not related to any grammatical assumption; a phrase is an n-gram. [sent-108, score-0.178]

52 The best translation ebest of f is obtained by: YI ebest= arg meaxp(e|f) = arg maexiY=1φ(fi|ei)λφ d(ai− bi−1)λdiY|=e1|pLM(ei|e1. [sent-109, score-0.188]

53 ei−1)λLM where φ(fi |ei) is the probability of translating a phrase ei einto a phrase fi. [sent-111, score-0.17]

54 In the reordering model, ai denotes the start position of the source phrase that was translated into the ith target phrase, and bi−1 denotes the end position of the source phrase translated into the (i− 1th) target phrase. [sent-113, score-0.224]

55 In our reconstruction problem, the difference between the source and target sentences is not in terms of languages, but in terms of forms. [sent-121, score-0.521]

56 In fact, our source sentence f is a stemmed sentence without stopwords, while the target sentence e is a complete English sentence. [sent-122, score-0.455]

57 “Translate” means to reconstruct the most probable sentence e given f inserting new words and reproducing the inflected surface forms of the source words. [sent-123, score-0.15]

58 1 Training of the model In Statistical Machine Translation, a PBM system is trained using parallel sentences, where each sentence in a language is paired with another sentence in a different language and one is the translation of the other. [sent-125, score-0.239]

59 In the reconstruction problem, we use a set, S1 of 2,487,414 English sentences extracted from the news. [sent-126, score-0.395]

60 This set is duplicated, S2, and for each sentence in S2, stopwords are removed and the re- maining words are stemmed using Porter’s stemmer (Porter, 1980). [sent-127, score-0.277]

61 Verbs are not included in this list, because they are relevant for the reconstruction task. [sent-129, score-0.352]

62 , 2007), a complete phrase-based translation toolkit for academic purposes. [sent-132, score-0.047]

63 It provides all the state-of-theart components needed to create a phrase-based machine translation system. [sent-133, score-0.047]

64 • 5 Experimental Results For our experiments we made use of the TAC 2009 data which conveniently contains humanproduced summaries against which we could evaluate the output of our system (NIST, 2009). [sent-135, score-0.343]

65 To begin our inquiry we carried out a phase of exploratory data analysis, in which we measured the average number of sentences per summary, words per sentence and words per summary in human vs. [sent-136, score-0.43]

66 Additionally, we also measured these statistics of summaries produced by our system at five different percentage cutoffs: 100%, 15%, 10%, 5% and 1%. [sent-138, score-0.415]

67 The most notable thing is that human summaries contain on average more and shorter sentences than the system summaries (see 2nd and 3rd column from left to right). [sent-140, score-0.854]

68 Secondly, we note that as the percentage cutoff decreases (from 4th column rightwards) the characteristics of the summaries produced by our system are increasingly more similar to those of the human summaries. [sent-141, score-0.59]

69 In other words, within the 100- word window imposed by the TAC guidelines, our system is able to fit more (and hence shorter) sentences as we decrease the percentage cutoff. [sent-142, score-0.153]

70 The last five columns of table 2 (from left to right) correspond to summaries produced by our system at various percentage cutoffs. [sent-147, score-0.415]

71 The 2nd column, LSAextract, corresponds to the performance of our system at producing summaries by sentence extraction only. [sent-148, score-0.482]

72 5 In the light of the above, the decrease in performance from column LSAextract to column ‘At 100%’ can be regarded as reconstruction error. [sent-149, score-0.53]

73 6 Then, as we decrease the percentage cutoff (from 4th column rightwards) we are increasingly covering more of the content comprised by the human summaries (as far as the ROUGE metrics are able to gauge this, of course). [sent-150, score-0.665]

74 In other words, the improvement of content coverage makes up for the reconstruction error, and at 5% cutoff we already obtain ROUGE scores comparable to LSAextract. [sent-151, score-0.457]

75 This suggests that if we improve the quality of our sentence reconstruction we would potentially end up with a better performing system than a typical LSA system based on sentence selection. [sent-152, score-0.544]

76 Finally, we admittedly note that by applying a percentage cutoff on the initial term set and further performing the sentence reconstruction we gain in content coverage, to a certain extent, on the expense of sentence readability. [sent-154, score-0.836]

77 6 Conclusion In this paper we proposed a novel approach to summary generation from summary representation based on the LSA summarization framework and on a machine-translation-inspired technique for sentence reconstruction. [sent-155, score-0.77]

78 Our preliminary results show that our approach is feasible, since it produces summaries which resemble better human summaries in terms ofthe average number of sentences per summary and yield ROUGE scores comparable to the participating systems in the Summarization task at TAC 2009. [sent-156, score-1.009]

79 In future work we plan on working towards improving the quality of our sentence reconstruction step in order to produce better and more readable sentences. [sent-158, score-0.483]

80 5These are, effectively, what we called initial pool of sentences in section 3, before the TSR generation. [sent-159, score-0.169]

81 6The only difference between the two types of summaries is the reconstruction step, since we are including 100% of the terms. [sent-160, score-0.695]

82 Global inference for sentence compression: An integer linear programming approach. [sent-180, score-0.096]

83 Generic text summarization using relevance measure and latent semantic analysis. [sent-192, score-0.201]

84 Summarization beyond sentence extraction: A probabilistic approach to sentence compression. [sent-204, score-0.192]


similar papers computed by tfidf model

tfidf for this paper:

wordName wordTfidf (topN-words)

[('reconstruction', 0.352), ('summaries', 0.343), ('tsr', 0.273), ('tac', 0.217), ('steinberger', 0.208), ('summarization', 0.201), ('summary', 0.192), ('salience', 0.183), ('rouge', 0.122), ('stemmed', 0.113), ('cutoff', 0.105), ('pbm', 0.104), ('phase', 0.099), ('je', 0.098), ('sentence', 0.096), ('lsa', 0.093), ('ei', 0.088), ('matrix', 0.086), ('singular', 0.083), ('pool', 0.075), ('percentage', 0.072), ('column', 0.07), ('ahern', 0.069), ('bertie', 0.069), ('ebest', 0.069), ('fein', 0.069), ('lsaextract', 0.069), ('rightwards', 0.069), ('sinn', 0.069), ('zek', 0.069), ('stopwords', 0.068), ('term', 0.064), ('interpretation', 0.064), ('compression', 0.063), ('minister', 0.061), ('adams', 0.061), ('leader', 0.061), ('irish', 0.061), ('bristol', 0.061), ('noisychannel', 0.061), ('mail', 0.061), ('tuesday', 0.061), ('koehn', 0.058), ('generation', 0.056), ('gong', 0.056), ('plm', 0.056), ('northern', 0.056), ('boguraev', 0.056), ('shorter', 0.055), ('source', 0.054), ('bi', 0.053), ('paraphrase', 0.053), ('participating', 0.052), ('unigrams', 0.052), ('recurring', 0.052), ('initial', 0.051), ('sp', 0.05), ('gerry', 0.049), ('kupiec', 0.049), ('mani', 0.049), ('kabadjov', 0.049), ('ireland', 0.049), ('meetings', 0.047), ('channel', 0.047), ('translation', 0.047), ('fi', 0.046), ('erkan', 0.045), ('aiming', 0.045), ('clarke', 0.045), ('peace', 0.045), ('zens', 0.045), ('producing', 0.043), ('och', 0.043), ('ste', 0.043), ('svd', 0.043), ('porter', 0.043), ('instantiation', 0.043), ('transformation', 0.043), ('sentences', 0.043), ('oxford', 0.042), ('clean', 0.042), ('royal', 0.042), ('phrase', 0.041), ('module', 0.041), ('diagonal', 0.039), ('hovy', 0.039), ('decrease', 0.038), ('comprised', 0.037), ('emphasis', 0.037), ('dot', 0.036), ('quirk', 0.036), ('terms', 0.036), ('arg', 0.036), ('prime', 0.035), ('readable', 0.035), ('bigrams', 0.034), ('explain', 0.034), ('ai', 0.034), ('adopt', 0.033), ('representation', 0.033)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 1.0000002 264 acl-2010-Wrapping up a Summary: From Representation to Generation

Author: Josef Steinberger ; Marco Turchi ; Mijail Kabadjov ; Ralf Steinberger ; Nello Cristianini

Abstract: The main focus of this work is to investigate robust ways for generating summaries from summary representations without recurring to simple sentence extraction and aiming at more human-like summaries. This is motivated by empirical evidence from TAC 2009 data showing that human summaries contain on average more and shorter sentences than the system summaries. We report encouraging preliminary results comparable to those attained by participating systems at TAC 2009.

2 0.26963827 77 acl-2010-Cross-Language Document Summarization Based on Machine Translation Quality Prediction

Author: Xiaojun Wan ; Huiying Li ; Jianguo Xiao

Abstract: Cross-language document summarization is a task of producing a summary in one language for a document set in a different language. Existing methods simply use machine translation for document translation or summary translation. However, current machine translation services are far from satisfactory, which results in that the quality of the cross-language summary is usually very poor, both in readability and content. In this paper, we propose to consider the translation quality of each sentence in the English-to-Chinese cross-language summarization process. First, the translation quality of each English sentence in the document set is predicted with the SVM regression method, and then the quality score of each sentence is incorporated into the summarization process. Finally, the English sentences with high translation quality and high informativeness are selected and translated to form the Chinese summary. Experimental results demonstrate the effectiveness and usefulness of the proposed approach. 1

3 0.24952902 38 acl-2010-Automatic Evaluation of Linguistic Quality in Multi-Document Summarization

Author: Emily Pitler ; Annie Louis ; Ani Nenkova

Abstract: To date, few attempts have been made to develop and validate methods for automatic evaluation of linguistic quality in text summarization. We present the first systematic assessment of several diverse classes of metrics designed to capture various aspects of well-written text. We train and test linguistic quality models on consecutive years of NIST evaluation data in order to show the generality of results. For grammaticality, the best results come from a set of syntactic features. Focus, coherence and referential clarity are best evaluated by a class of features measuring local coherence on the basis of cosine similarity between sentences, coreference informa- tion, and summarization specific features. Our best results are 90% accuracy for pairwise comparisons of competing systems over a test set of several inputs and 70% for ranking summaries of a specific input.

4 0.22970699 14 acl-2010-A Risk Minimization Framework for Extractive Speech Summarization

Author: Shih-Hsiang Lin ; Berlin Chen

Abstract: In this paper, we formulate extractive summarization as a risk minimization problem and propose a unified probabilistic framework that naturally combines supervised and unsupervised summarization models to inherit their individual merits as well as to overcome their inherent limitations. In addition, the introduction of various loss functions also provides the summarization framework with a flexible but systematic way to render the redundancy and coherence relationships among sentences and between sentences and the whole document, respectively. Experiments on speech summarization show that the methods deduced from our framework are very competitive with existing summarization approaches. 1

5 0.19459786 11 acl-2010-A New Approach to Improving Multilingual Summarization Using a Genetic Algorithm

Author: Marina Litvak ; Mark Last ; Menahem Friedman

Abstract: Automated summarization methods can be defined as “language-independent,” if they are not based on any languagespecific knowledge. Such methods can be used for multilingual summarization defined by Mani (2001) as “processing several languages, with summary in the same language as input.” In this paper, we introduce MUSE, a languageindependent approach for extractive summarization based on the linear optimization of several sentence ranking measures using a genetic algorithm. We tested our methodology on two languages—English and Hebrew—and evaluated its performance with ROUGE-1 Recall vs. state- of-the-art extractive summarization approaches. Our results show that MUSE performs better than the best known multilingual approach (TextRank1) in both languages. Moreover, our experimental results on a bilingual (English and Hebrew) document collection suggest that MUSE does not need to be retrained on each language and the same model can be used across at least two different languages.

6 0.18788332 124 acl-2010-Generating Image Descriptions Using Dependency Relational Patterns

7 0.17218986 8 acl-2010-A Hybrid Hierarchical Model for Multi-Document Summarization

8 0.16806854 39 acl-2010-Automatic Generation of Story Highlights

9 0.15261948 188 acl-2010-Optimizing Informativeness and Readability for Sentiment Summarization

10 0.11207879 125 acl-2010-Generating Templates of Entity Summaries with an Entity-Aspect Model and Pattern Mining

11 0.10981645 171 acl-2010-Metadata-Aware Measures for Answer Summarization in Community Question Answering

12 0.094240569 122 acl-2010-Generating Fine-Grained Reviews of Songs from Album Reviews

13 0.081476577 240 acl-2010-Training Phrase Translation Models with Leaving-One-Out

14 0.070113733 192 acl-2010-Paraphrase Lattice for Statistical Machine Translation

15 0.068891414 119 acl-2010-Fixed Length Word Suffix for Factored Statistical Machine Translation

16 0.067839943 51 acl-2010-Bilingual Sense Similarity for Statistical Machine Translation

17 0.057643473 220 acl-2010-Syntactic and Semantic Factors in Processing Difficulty: An Integrated Measure

18 0.055028513 249 acl-2010-Unsupervised Search for the Optimal Segmentation for Statistical Machine Translation

19 0.054424915 66 acl-2010-Compositional Matrix-Space Models of Language

20 0.053813465 140 acl-2010-Identifying Non-Explicit Citing Sentences for Citation-Based Summarization.


similar papers computed by lsi model

lsi for this paper:

topicId topicWeight

[(0, -0.204), (1, -0.004), (2, -0.152), (3, -0.017), (4, -0.005), (5, 0.001), (6, 0.001), (7, -0.43), (8, -0.063), (9, -0.034), (10, 0.064), (11, -0.009), (12, -0.103), (13, 0.05), (14, -0.043), (15, -0.068), (16, -0.027), (17, 0.028), (18, 0.022), (19, 0.042), (20, 0.092), (21, 0.09), (22, -0.023), (23, 0.014), (24, 0.043), (25, 0.016), (26, -0.058), (27, -0.02), (28, 0.047), (29, 0.0), (30, -0.018), (31, -0.044), (32, -0.034), (33, -0.047), (34, 0.065), (35, -0.002), (36, -0.085), (37, -0.017), (38, 0.044), (39, 0.034), (40, 0.001), (41, -0.031), (42, 0.016), (43, -0.018), (44, 0.02), (45, 0.011), (46, -0.003), (47, 0.015), (48, -0.019), (49, -0.042)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.9585166 264 acl-2010-Wrapping up a Summary: From Representation to Generation

Author: Josef Steinberger ; Marco Turchi ; Mijail Kabadjov ; Ralf Steinberger ; Nello Cristianini

Abstract: The main focus of this work is to investigate robust ways for generating summaries from summary representations without recurring to simple sentence extraction and aiming at more human-like summaries. This is motivated by empirical evidence from TAC 2009 data showing that human summaries contain on average more and shorter sentences than the system summaries. We report encouraging preliminary results comparable to those attained by participating systems at TAC 2009.

2 0.8777892 14 acl-2010-A Risk Minimization Framework for Extractive Speech Summarization

Author: Shih-Hsiang Lin ; Berlin Chen

Abstract: In this paper, we formulate extractive summarization as a risk minimization problem and propose a unified probabilistic framework that naturally combines supervised and unsupervised summarization models to inherit their individual merits as well as to overcome their inherent limitations. In addition, the introduction of various loss functions also provides the summarization framework with a flexible but systematic way to render the redundancy and coherence relationships among sentences and between sentences and the whole document, respectively. Experiments on speech summarization show that the methods deduced from our framework are very competitive with existing summarization approaches. 1

3 0.8373127 11 acl-2010-A New Approach to Improving Multilingual Summarization Using a Genetic Algorithm

Author: Marina Litvak ; Mark Last ; Menahem Friedman

Abstract: Automated summarization methods can be defined as “language-independent,” if they are not based on any languagespecific knowledge. Such methods can be used for multilingual summarization defined by Mani (2001) as “processing several languages, with summary in the same language as input.” In this paper, we introduce MUSE, a languageindependent approach for extractive summarization based on the linear optimization of several sentence ranking measures using a genetic algorithm. We tested our methodology on two languages—English and Hebrew—and evaluated its performance with ROUGE-1 Recall vs. state- of-the-art extractive summarization approaches. Our results show that MUSE performs better than the best known multilingual approach (TextRank1) in both languages. Moreover, our experimental results on a bilingual (English and Hebrew) document collection suggest that MUSE does not need to be retrained on each language and the same model can be used across at least two different languages.

4 0.81601435 77 acl-2010-Cross-Language Document Summarization Based on Machine Translation Quality Prediction

Author: Xiaojun Wan ; Huiying Li ; Jianguo Xiao

Abstract: Cross-language document summarization is a task of producing a summary in one language for a document set in a different language. Existing methods simply use machine translation for document translation or summary translation. However, current machine translation services are far from satisfactory, which results in that the quality of the cross-language summary is usually very poor, both in readability and content. In this paper, we propose to consider the translation quality of each sentence in the English-to-Chinese cross-language summarization process. First, the translation quality of each English sentence in the document set is predicted with the SVM regression method, and then the quality score of each sentence is incorporated into the summarization process. Finally, the English sentences with high translation quality and high informativeness are selected and translated to form the Chinese summary. Experimental results demonstrate the effectiveness and usefulness of the proposed approach. 1

5 0.79207206 38 acl-2010-Automatic Evaluation of Linguistic Quality in Multi-Document Summarization

Author: Emily Pitler ; Annie Louis ; Ani Nenkova

Abstract: To date, few attempts have been made to develop and validate methods for automatic evaluation of linguistic quality in text summarization. We present the first systematic assessment of several diverse classes of metrics designed to capture various aspects of well-written text. We train and test linguistic quality models on consecutive years of NIST evaluation data in order to show the generality of results. For grammaticality, the best results come from a set of syntactic features. Focus, coherence and referential clarity are best evaluated by a class of features measuring local coherence on the basis of cosine similarity between sentences, coreference informa- tion, and summarization specific features. Our best results are 90% accuracy for pairwise comparisons of competing systems over a test set of several inputs and 70% for ranking summaries of a specific input.

6 0.73049772 39 acl-2010-Automatic Generation of Story Highlights

7 0.72588009 8 acl-2010-A Hybrid Hierarchical Model for Multi-Document Summarization

8 0.67161155 124 acl-2010-Generating Image Descriptions Using Dependency Relational Patterns

9 0.64384872 188 acl-2010-Optimizing Informativeness and Readability for Sentiment Summarization

10 0.59065354 140 acl-2010-Identifying Non-Explicit Citing Sentences for Citation-Based Summarization.

11 0.58705002 125 acl-2010-Generating Templates of Entity Summaries with an Entity-Aspect Model and Pattern Mining

12 0.52780205 122 acl-2010-Generating Fine-Grained Reviews of Songs from Album Reviews

13 0.47288477 171 acl-2010-Metadata-Aware Measures for Answer Summarization in Community Question Answering

14 0.46511188 136 acl-2010-How Many Words Is a Picture Worth? Automatic Caption Generation for News Images

15 0.43008193 196 acl-2010-Plot Induction and Evolutionary Search for Story Generation

16 0.38555577 166 acl-2010-Learning Word-Class Lattices for Definition and Hypernym Extraction

17 0.36105829 101 acl-2010-Entity-Based Local Coherence Modelling Using Topological Fields

18 0.35852128 119 acl-2010-Fixed Length Word Suffix for Factored Statistical Machine Translation

19 0.31627846 7 acl-2010-A Generalized-Zero-Preserving Method for Compact Encoding of Concept Lattices

20 0.31597403 157 acl-2010-Last but Definitely Not Least: On the Role of the Last Sentence in Automatic Polarity-Classification


similar papers computed by lda model

lda for this paper:

topicId topicWeight

[(3, 0.31), (4, 0.015), (16, 0.016), (25, 0.04), (33, 0.022), (39, 0.019), (42, 0.019), (44, 0.012), (59, 0.109), (73, 0.036), (78, 0.027), (80, 0.019), (83, 0.099), (84, 0.033), (98, 0.155)]

similar papers list:

simIndex simValue paperId paperTitle

1 0.79914176 81 acl-2010-Decision Detection Using Hierarchical Graphical Models

Author: Trung H. Bui ; Stanley Peters

Abstract: We investigate hierarchical graphical models (HGMs) for automatically detecting decisions in multi-party discussions. Several types of dialogue act (DA) are distinguished on the basis of their roles in formulating decisions. HGMs enable us to model dependencies between observed features of discussions, decision DAs, and subdialogues that result in a decision. For the task of detecting decision regions, an HGM classifier was found to outperform non-hierarchical graphical models and support vector machines, raising the F1-score to 0.80 from 0.55.

same-paper 2 0.79240316 264 acl-2010-Wrapping up a Summary: From Representation to Generation

Author: Josef Steinberger ; Marco Turchi ; Mijail Kabadjov ; Ralf Steinberger ; Nello Cristianini

Abstract: The main focus of this work is to investigate robust ways for generating summaries from summary representations without recurring to simple sentence extraction and aiming at more human-like summaries. This is motivated by empirical evidence from TAC 2009 data showing that human summaries contain on average more and shorter sentences than the system summaries. We report encouraging preliminary results comparable to those attained by participating systems at TAC 2009.

3 0.65225416 133 acl-2010-Hierarchical Search for Word Alignment

Author: Jason Riesa ; Daniel Marcu

Abstract: We present a simple yet powerful hierarchical search algorithm for automatic word alignment. Our algorithm induces a forest of alignments from which we can efficiently extract a ranked k-best list. We score a given alignment within the forest with a flexible, linear discriminative model incorporating hundreds of features, and trained on a relatively small amount of annotated data. We report results on Arabic-English word alignment and translation tasks. Our model outperforms a GIZA++ Model-4 baseline by 6.3 points in F-measure, yielding a 1.1 BLEU score increase over a state-of-the-art syntax-based machine translation system.

4 0.58019495 184 acl-2010-Open-Domain Semantic Role Labeling by Modeling Word Spans

Author: Fei Huang ; Alexander Yates

Abstract: Most supervised language processing systems show a significant drop-off in performance when they are tested on text that comes from a domain significantly different from the domain of the training data. Semantic role labeling techniques are typically trained on newswire text, and in tests their performance on fiction is as much as 19% worse than their performance on newswire text. We investigate techniques for building open-domain semantic role labeling systems that approach the ideal of a train-once, use-anywhere system. We leverage recently-developed techniques for learning representations of text using latent-variable language models, and extend these techniques to ones that provide the kinds of features that are useful for semantic role labeling. In experiments, our novel system reduces error by 16% relative to the previous state of the art on out-of-domain text.

5 0.57708168 51 acl-2010-Bilingual Sense Similarity for Statistical Machine Translation

Author: Boxing Chen ; George Foster ; Roland Kuhn

Abstract: This paper proposes new algorithms to compute the sense similarity between two units (words, phrases, rules, etc.) from parallel corpora. The sense similarity scores are computed by using the vector space model. We then apply the algorithms to statistical machine translation by computing the sense similarity between the source and target side of translation rule pairs. Similarity scores are used as additional features of the translation model to improve translation performance. Significant improvements are obtained over a state-of-the-art hierarchical phrase-based machine translation system. 1

6 0.57434165 55 acl-2010-Bootstrapping Semantic Analyzers from Non-Contradictory Texts

7 0.573645 144 acl-2010-Improved Unsupervised POS Induction through Prototype Discovery

8 0.57364166 145 acl-2010-Improving Arabic-to-English Statistical Machine Translation by Reordering Post-Verbal Subjects for Alignment

9 0.57363105 87 acl-2010-Discriminative Modeling of Extraction Sets for Machine Translation

10 0.5734067 54 acl-2010-Boosting-Based System Combination for Machine Translation

11 0.5724597 56 acl-2010-Bridging SMT and TM with Translation Recommendation

12 0.57176787 76 acl-2010-Creating Robust Supervised Classifiers via Web-Scale N-Gram Data

13 0.57147032 148 acl-2010-Improving the Use of Pseudo-Words for Evaluating Selectional Preferences

14 0.57123011 39 acl-2010-Automatic Generation of Story Highlights

15 0.57110304 79 acl-2010-Cross-Lingual Latent Topic Extraction

16 0.57095313 109 acl-2010-Experiments in Graph-Based Semi-Supervised Learning Methods for Class-Instance Acquisition

17 0.57091969 88 acl-2010-Discriminative Pruning for Discriminative ITG Alignment

18 0.57062936 218 acl-2010-Structural Semantic Relatedness: A Knowledge-Based Method to Named Entity Disambiguation

19 0.5699423 244 acl-2010-TrustRank: Inducing Trust in Automatic Translations via Ranking

20 0.56980956 102 acl-2010-Error Detection for Statistical Machine Translation Using Linguistic Features