acl acl2011 acl2011-21 knowledge-graph by maker-knowledge-mining

21 acl-2011-A Pilot Study of Opinion Summarization in Conversations


Source: pdf

Author: Dong Wang ; Yang Liu

Abstract: This paper presents a pilot study of opinion summarization on conversations. We create a corpus containing extractive and abstractive summaries of speaker’s opinion towards a given topic using 88 telephone conversations. We adopt two methods to perform extractive summarization. The first one is a sentence-ranking method that linearly combines scores measured from different aspects including topic relevance, subjectivity, and sentence importance. The second one is a graph-based method, which incorporates topic and sentiment information, as well as additional information about sentence-to-sentence relations extracted based on dialogue structure. Our evaluation results show that both methods significantly outperform the baseline approach that extracts the longest utterances. In particular, we find that incorporating dialogue structure in the graph-based method contributes to the improved system performance.

Reference: text


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 edu Abstract This paper presents a pilot study of opinion summarization on conversations. [sent-3, score-0.672]

2 We create a corpus containing extractive and abstractive summaries of speaker’s opinion towards a given topic using 88 telephone conversations. [sent-4, score-1.298]

3 The second one is a graph-based method, which incorporates topic and sentiment information, as well as additional information about sentence-to-sentence relations extracted based on dialogue structure. [sent-7, score-0.439]

4 1 Introduction Both sentiment analysis (opinion recognition) and summarization have been well studied in recent years in the natural language processing (NLP) community. [sent-10, score-0.449]

5 Summarization has been applied to different genres, such as news articles, scientific articles, and speech domains including broadcast news, meetings, conversations and lectures. [sent-12, score-0.37]

6 This kind of questions can be treated as a topic-oriented opinion summarization task. [sent-16, score-0.62]

7 Opinion summarization was run as a pilot task in Text Analysis Conference (TAC) in 2008. [sent-17, score-0.393]

8 The task was to produce summaries of opinions on specified targets from a set of blog documents. [sent-18, score-0.2]

9 The problem is defined as, given a conversation and a topic, a summarization system needs to generate a summary of the speaker’s opinion towards the topic. [sent-20, score-0.852]

10 This task is built upon opinion recognition and topic or query based summarization. [sent-21, score-0.471]

11 , 2008); (c) In conversational speech, information density is low and there are often off topic discussions, therefore presenting a need to identify utterances that are relevant to the topic. [sent-24, score-0.334]

12 In this paper we perform an exploratory study on opinion summarization in conversations. [sent-25, score-0.62]

13 Ac s2s0o1ci1a Atiosnso fcoirat Cioonm foprut Caotimonpaulta Lti nognuails Lti cnsg,u piasgteics 331–339, widely used in extractive summarization: sentenceranking and graph-based methods. [sent-28, score-0.371]

14 Our system attempts to incorporate more information about topic relevancy and sentiment scores. [sent-29, score-0.326]

15 Furthermore, in the graph-based method, we propose to better incorporate the dialogue structure information in the graph in order to select salient summary utterances. [sent-30, score-0.355]

16 We explain our opinion-oriented conversation summarization system in Section 4 and present experimental results and analysis in Section 5. [sent-36, score-0.46]

17 2 Related Work Research in document summarization has been well established over the past decades. [sent-38, score-0.341]

18 Recently there is an increasing research interest in speech summarization, such as conversational telephone speech (Zhu and Penn, 2006; Zechner, 2002), broadcast news (Maskey and Hirschberg, 2005; Lin et al. [sent-41, score-0.312]

19 Prior research has also explored using speech specific information, including prosodic features, dialog structure, and speech recognition confidence. [sent-48, score-0.303]

20 In order to provide a summary over opinions, we need to find out which utterances in the conversation contain opinion. [sent-49, score-0.351]

21 Most of prior work used classification methods such as naive Bayes or SVMs to perform the polarity classification or opinion detection. [sent-53, score-0.279]

22 Only a handful studies have used conversational speech for opinion recognition (Murray and Carenini, 2009; Raaijmakers et al. [sent-54, score-0.431]

23 Our work is also related to question answering (QA), especially opinion question answering. [sent-56, score-0.343]

24 , 2010) answers some specific opinion questions like “Why do people criticize Richard Branson? [sent-60, score-0.279]

25 Our work is different in that we are not going to answer specific opinion questions, instead, we provide a summary on the speaker’s opinion towards a given topic. [sent-62, score-0.713]

26 , 2010) have explored opinion summarization in review domain, and (Paul et al. [sent-65, score-0.651]

27 However, opinion summarization in spontaneous conversation is seldom studied. [sent-67, score-0.821]

28 3 Corpus Creation Though there are many annotated data sets for the research of speech summarization and sentiment analysis, there is no corpus available for opinion summarization on spontaneous speech. [sent-68, score-1.222]

29 1 These are conversational telephone speech between two strangers that were assigned a topic to talk about for around 5 minutes. [sent-70, score-0.34]

30 pus, we selected 88 conversations from 6 topics for this study. [sent-74, score-0.236]

31 Table 1lists the number of conversations in each topic, their average length (measured in the unit of dialogue acts (DA)) and standard deviation of length. [sent-75, score-0.368]

32 4erv0of conversations in each topic, average length (number of dialog acts), and standard deviation. [sent-78, score-0.265]

33 The rest of the conversations has only one annotation. [sent-81, score-0.2]

34 The annotators have access to both conversation transcripts and audio files. [sent-82, score-0.224]

35 For each conversation, the annotator writes an abstractive summary of up to 100 words for each speaker about his/her opinion or attitude on the given topic. [sent-83, score-0.871]

36 Then the annotator selects up to 15 DAs (no minimum limit) in the transcripts for each speaker, from which their abstractive summary is derived. [sent-85, score-0.495]

37 The selected DAs are used as the human generated extractive summary. [sent-86, score-0.365]

38 In addition, the annotator is asked to select an overall opinion towards the topic for each speaker among five categories: strongly support, somewhat support, neutral, somewhat against, strongly against. [sent-87, score-0.617]

39 Therefore for each conversation, we have an abstractive summary, an extractive summary, and an overall opinion for each speaker. [sent-88, score-0.912]

40 The following shows an example of such annotation for speaker B in a dialogue about “capital punishment”: [Extractive Summary] I think I’ve seen some statistics that say that, uh, it’s more expensive to kill somebody than to keep them in prison for life. [sent-89, score-0.398]

41 I ’t think he could ever redeem himself, don but if you look at who gets accused and who are the ones who actually get executed, it’s very racially related and ethnically related [Abstractive Summary] B is against capital punishment except under certain circumstances. [sent-91, score-0.24]

42 B finds that crimes deserving of capital punishment are “crimes of the moment” and as a result feels that capital punishment is not an effective deterrent. [sent-92, score-0.353]

43 [Overall Opinion] Somewhat against Table 2 shows the compression ratio of the extractive summaries and abstractive summaries as well as their standard deviation. [sent-94, score-1.175]

44 06 Table 2: Compression ratio and standard deviation of extractive and abstractive summaries. [sent-100, score-0.699]

45 We measured the inter-annotator agreement among the three annotators for the 18 conversations (each has two speakers, thus 36 “documents” in total). [sent-101, score-0.301]

46 For the extractive or abstractive summaries, we use ROUGE scores (Lin, 2004), a metric used to evaluate automatic summarization performance, to measure the pairwise agreement of summaries from different annotators. [sent-103, score-1.181]

47 We notice that the inter-annotator agreement for extractive summaries is comparable to other speech Table3:Int r-xsaovncetriavlos upamignreoamisntRαf o-2=1Lre0x . [sent-106, score-0.607]

48 The agreement on abstractive summaries is much lower than extractive summaries, which is as expected. [sent-109, score-0.84]

49 Even for the same opinion or sentence, annotators use different words in the abstractive summaries. [sent-110, score-0.646]

50 The agreement for the overall opinion annotation is similar to other opinion/emotion studies (Wilson, 2008b), but slightly lower than the level recommended by Krippendorff for reliable data (α = 0. [sent-111, score-0.317]

51 8) (Hayes and Krippendorff, 2007), which shows it is even difficult for humans to determine what opinion a person holds (support or against something). [sent-112, score-0.279]

52 Therefore this also demonstrates that it is more appropriate to provide a summary rather than a simple opinion category to answer questions about a person’s opinion towards something. [sent-114, score-0.713]

53 4 Opinion Summarization Methods Automatic summarization can be divided into extractive summarization and abstractive summarization. [sent-115, score-1.315]

54 Extractive summarization selects sentences from the original documents to form a summary; whereas abstractive summarization requires generation of new sentences that represent the most salient content in the original documents like humans do. [sent-116, score-1.03]

55 Often extractive summarization is used as the first step to generate abstractive summary. [sent-117, score-0.974]

56 As a pilot study for the problem of opinion summarization in conversations, we treat this problem as an extractive summarization task. [sent-118, score-1.342]

57 This section describes two approaches we have explored in generating extractive summaries. [sent-119, score-0.36]

58 In addition, they do not require a large labeled data set for modeling training, as needed in some classification or feature based summarization approaches. [sent-123, score-0.341]

59 score(s) = λsimsim(s, D) + λrelREL(s, topic) +λsentsentiment(s) + λlenlength(s) Xλi = 1 Xi (1) • • sim(s, D) is the cosine similarity between DA s amn(ds ,aDll )th ise uhtete croasnicnees s i mn tlharei dialogue nfr DomA the same speaker, D. [sent-126, score-0.246]

60 It measures the relevancy of s to the entire dialogue from the target speaker. [sent-127, score-0.223]

61 It has been shown to be an important indicator in summarization for various domains. [sent-129, score-0.341]

62 REL(s, topic) measures the topic relevance of RDAE s. [sent-132, score-0.244]

63 html where all the statistics are collected from the Switchboard corpus: p(w&topic;) denotes the probability that word w appears in a dialogue of topic t, and p(w) is the probability of w appearing in a dialogue of any topic. [sent-140, score-0.499]

64 In this approach, a document 335 is modeled as an adjacency matrix, where each node represents a sentence, and the weight of the edge between each pair of sentences is their similarity (cosine similarity is typically used). [sent-153, score-0.213]

65 The basic framework we use in this study is similar to the query-based graph summarization system in (Zhao et al. [sent-156, score-0.371]

66 We also consider sentiment and topic relevance information, and propose to incorporate information obtained from dialog structure in this framework. [sent-158, score-0.417]

67 • • If s and v are from the same speaker, and separated only by one DsaAm efr sopmea aern,ot ahnedr speaker with length less than 3 words (usually backchannel), there is an edge from s to v as well as an edge from v to s with weight 1 (ADJ(s, v) = ADJ(v, s) = 1). [sent-164, score-0.269]

68 Since we are using a directed graph for the sentence connections to model dialog structure, the resulting adjacency matrix is asymmetric. [sent-174, score-0.257]

69 Also note that in the first sentence ranking method or the basic graph methods, summarization is conducted for each speaker separately. [sent-176, score-0.54]

70 Utterances from one speaker have no influence on the summary decision for the other speaker. [sent-177, score-0.252]

71 1 Experimental Setup The 18 conversations annotated by all 3 annotators are used as test set, and the rest of 70 conversations are used as development set to tune the parameters (determining the best combination weights). [sent-180, score-0.463]

72 We perform extractive summarization using different word compression ratios (ranging from 10% to 25%). [sent-182, score-0.808]

73 The system-generated summaries are compared to human annotated extractive and abstractive summaries. [sent-184, score-0.802]

74 We use ROUGE as the evaluation metrics for summarization performance. [sent-185, score-0.341]

75 This has been shown to be a relatively strong baseline for speech summarization (Gillick et al. [sent-188, score-0.412]

76 We treat each annotator’s extractive summary as a system summary, and compare to the other two annotators’ extractive and abstractive summaries. [sent-191, score-1.075]

77 2 Results From the development set, we used the grid search method to obtain the best combination weights for the two summarization methods. [sent-194, score-0.341]

78 In addition, REL score is already able to catch the topic relevancy of the sentence. [sent-201, score-0.261]

79 This is different from graph-based summarization systems for text domains. [sent-208, score-0.341]

80 tween utterances does not perform well in conversation summarization. [sent-211, score-0.238]

81 Figure 1 shows the ROUGE-1 F-scores comparing to human extractive and abstractive summaries for different compression ratios. [sent-212, score-0.94]

82 When compared to extractive reference summaries, sentence-ranking is slightly better except for the compression ratio of 0. [sent-218, score-0.565]

83 When compared to abstractive reference summaries, the graphbased method is slightly better. [sent-220, score-0.336]

84 3 Analysis To analyze the effect of dialogue structure we introduce in the graph-based summarization method, we compare two configurations: λadj = 0 (only using REL score and sentiment score in ranking) and λadj = 0. [sent-223, score-0.703]

85 We generate summaries using these two setups and compare with human selected sentences. [sent-225, score-0.205]

86 This results in a large number of reference summary DAs (because of low human agreement), and thus the number of false negatives in the system output is very high. [sent-228, score-0.205]

87 As expected, a smaller compression ratio (fewer selected DAs in the system output) yields a higher false negative rate and a lower false positive rate. [sent-229, score-0.36]

88 We use the statistics from the Switchboard corpus to measure the relevance of each word to a given topic (PMI score), therefore only when people use the same word in different conversations of the topic, the PMI score of this word and the topic is high. [sent-242, score-0.65]

89 We used DA segments as units for extractive summarization, which can be problematic. [sent-247, score-0.329]

90 The two DAs from speaker B are not selected by our system but selected by human anno338 tators, causing false negative errors. [sent-251, score-0.271]

91 – – 6 Conclusion and Future Work This paper investigates two unsupervised methods in opinion summarization on spontaneous conversations by incorporating topic score and sentiment score in existing summarization techniques. [sent-253, score-1.6]

92 In the graph-based method, we use an adjacency matrix to model the dialogue structure and utilize it to find salient utterances in conversations. [sent-255, score-0.429]

93 Our experiments show that both methods are able to improve the baseline approach, and we find that the cosine similarity between utterances or between an utterance and the whole docu- ment is not as useful as in other document summarization tasks. [sent-256, score-0.579]

94 Going beyond traditional QA systems: challenges and keys in opinion question answering. [sent-265, score-0.311]

95 Auto- matic summarization of voicemail messages using lexical and prosodic features. [sent-299, score-0.377]

96 An analysis of human extractive summaries in meeting corpus. [sent-312, score-0.498]

97 Opinion summarization with integer linear programming formulation for sentence extraction and ordering. [sent-338, score-0.371]

98 A sentiment education: sentiment analysis using subjectivity summarization based on minimum cuts. [sent-342, score-0.624]

99 Improving supervised learning for meeting summarization using sampling and regression. [sent-380, score-0.341]

100 Automatic summarization of open-domain multiparty dialogues in dive rse genres. [sent-384, score-0.385]


similar papers computed by tfidf model

tfidf for this paper:

wordName wordTfidf (topN-words)

[('summarization', 0.341), ('extractive', 0.329), ('abstractive', 0.304), ('opinion', 0.279), ('conversations', 0.2), ('adj', 0.176), ('das', 0.174), ('summaries', 0.169), ('dialogue', 0.168), ('topic', 0.163), ('da', 0.161), ('speaker', 0.139), ('compression', 0.138), ('conversation', 0.119), ('utterances', 0.119), ('rel', 0.116), ('summary', 0.113), ('sentiment', 0.108), ('pmi', 0.088), ('switchboard', 0.087), ('punishment', 0.085), ('spontaneous', 0.082), ('relevance', 0.081), ('speech', 0.071), ('murray', 0.068), ('subjectivity', 0.067), ('ratio', 0.066), ('edge', 0.065), ('dialog', 0.065), ('rouge', 0.064), ('adjacency', 0.064), ('capital', 0.064), ('raaijmakers', 0.063), ('annotators', 0.063), ('false', 0.06), ('sim', 0.059), ('crimes', 0.055), ('relevancy', 0.055), ('telephone', 0.054), ('conversational', 0.052), ('pilot', 0.052), ('qa', 0.051), ('think', 0.049), ('theresa', 0.048), ('krippendorff', 0.046), ('uh', 0.046), ('salient', 0.044), ('multiparty', 0.044), ('score', 0.043), ('similarity', 0.042), ('furui', 0.042), ('koumpis', 0.042), ('maskey', 0.042), ('prison', 0.042), ('redeem', 0.042), ('relorig', 0.042), ('sentenceranking', 0.042), ('longest', 0.042), ('transcripts', 0.042), ('answer', 0.042), ('utterance', 0.041), ('opinionated', 0.039), ('agreement', 0.038), ('wilson', 0.038), ('hirschberg', 0.037), ('subjective', 0.037), ('gabriel', 0.037), ('korbinian', 0.037), ('renals', 0.037), ('annotator', 0.036), ('selected', 0.036), ('prosodic', 0.036), ('cosine', 0.036), ('domains', 0.035), ('connections', 0.034), ('balahur', 0.034), ('godfrey', 0.034), ('nishikawa', 0.034), ('stoyanov', 0.034), ('matrix', 0.034), ('broadcast', 0.034), ('salience', 0.033), ('julia', 0.032), ('reference', 0.032), ('question', 0.032), ('garg', 0.032), ('hayes', 0.032), ('moment', 0.032), ('xie', 0.032), ('connection', 0.031), ('explored', 0.031), ('opinions', 0.031), ('janyce', 0.031), ('news', 0.03), ('viewpoints', 0.03), ('sentence', 0.03), ('graph', 0.03), ('recognition', 0.029), ('benoit', 0.029), ('gillick', 0.029)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.99999982 21 acl-2011-A Pilot Study of Opinion Summarization in Conversations

Author: Dong Wang ; Yang Liu

Abstract: This paper presents a pilot study of opinion summarization on conversations. We create a corpus containing extractive and abstractive summaries of speaker’s opinion towards a given topic using 88 telephone conversations. We adopt two methods to perform extractive summarization. The first one is a sentence-ranking method that linearly combines scores measured from different aspects including topic relevance, subjectivity, and sentence importance. The second one is a graph-based method, which incorporates topic and sentiment information, as well as additional information about sentence-to-sentence relations extracted based on dialogue structure. Our evaluation results show that both methods significantly outperform the baseline approach that extracts the longest utterances. In particular, we find that incorporating dialogue structure in the graph-based method contributes to the improved system performance.

2 0.2928099 187 acl-2011-Jointly Learning to Extract and Compress

Author: Taylor Berg-Kirkpatrick ; Dan Gillick ; Dan Klein

Abstract: We learn a joint model of sentence extraction and compression for multi-document summarization. Our model scores candidate summaries according to a combined linear model whose features factor over (1) the n-gram types in the summary and (2) the compressions used. We train the model using a marginbased objective whose loss captures end summary quality. Because of the exponentially large set of candidate summaries, we use a cutting-plane algorithm to incrementally detect and add active constraints efficiently. Inference in our model can be cast as an ILP and thereby solved in reasonable time; we also present a fast approximation scheme which achieves similar performance. Our jointly extracted and compressed summaries outperform both unlearned baselines and our learned extraction-only system on both ROUGE and Pyramid, without a drop in judged linguistic quality. We achieve the highest published ROUGE results to date on the TAC 2008 data set.

3 0.2535181 159 acl-2011-Identifying Noun Product Features that Imply Opinions

Author: Lei Zhang ; Bing Liu

Abstract: Identifying domain-dependent opinion words is a key problem in opinion mining and has been studied by several researchers. However, existing work has been focused on adjectives and to some extent verbs. Limited work has been done on nouns and noun phrases. In our work, we used the feature-based opinion mining model, and we found that in some domains nouns and noun phrases that indicate product features may also imply opinions. In many such cases, these nouns are not subjective but objective. Their involved sentences are also objective sentences and imply positive or negative opinions. Identifying such nouns and noun phrases and their polarities is very challenging but critical for effective opinion mining in these domains. To the best of our knowledge, this problem has not been studied in the literature. This paper proposes a method to deal with the problem. Experimental results based on real-life datasets show promising results. 1

4 0.22058949 308 acl-2011-Towards a Framework for Abstractive Summarization of Multimodal Documents

Author: Charles Greenbacker

Abstract: We propose a framework for generating an abstractive summary from a semantic model of a multimodal document. We discuss the type of model required, the means by which it can be constructed, how the content of the model is rated and selected, and the method of realizing novel sentences for the summary. To this end, we introduce a metric called information density used for gauging the importance of content obtained from text and graphical sources.

5 0.19527498 185 acl-2011-Joint Identification and Segmentation of Domain-Specific Dialogue Acts for Conversational Dialogue Systems

Author: Fabrizio Morbini ; Kenji Sagae

Abstract: Individual utterances often serve multiple communicative purposes in dialogue. We present a data-driven approach for identification of multiple dialogue acts in single utterances in the context of dialogue systems with limited training data. Our approach results in significantly increased understanding of user intent, compared to two strong baselines.

6 0.18476075 270 acl-2011-SciSumm: A Multi-Document Summarization System for Scientific Articles

7 0.17869842 326 acl-2011-Using Bilingual Information for Cross-Language Document Summarization

8 0.17093836 76 acl-2011-Comparative News Summarization Using Linear Programming

9 0.1687566 98 acl-2011-Discovery of Topically Coherent Sentences for Extractive Summarization

10 0.16748151 18 acl-2011-A Latent Topic Extracting Method based on Events in a Document and its Application

11 0.16663174 251 acl-2011-Probabilistic Document Modeling for Syntax Removal in Text Summarization

12 0.1664646 255 acl-2011-Query Snowball: A Co-occurrence-based Approach to Multi-document Summarization for Question Answering

13 0.159141 272 acl-2011-Semantic Information and Derivation Rules for Robust Dialogue Act Detection in a Spoken Dialogue System

14 0.15174191 156 acl-2011-IMASS: An Intelligent Microblog Analysis and Summarization System

15 0.14990728 4 acl-2011-A Class of Submodular Functions for Document Summarization

16 0.14955951 95 acl-2011-Detection of Agreement and Disagreement in Broadcast Conversations

17 0.14390448 260 acl-2011-Recognizing Authority in Dialogue with an Integer Linear Programming Constrained Model

18 0.14130831 45 acl-2011-Aspect Ranking: Identifying Important Product Aspects from Online Consumer Reviews

19 0.14059241 47 acl-2011-Automatic Assessment of Coverage Quality in Intelligence Reports

20 0.13777617 33 acl-2011-An Affect-Enriched Dialogue Act Classification Model for Task-Oriented Dialogue


similar papers computed by lsi model

lsi for this paper:

topicId topicWeight

[(0, 0.233), (1, 0.254), (2, 0.03), (3, 0.121), (4, -0.246), (5, 0.144), (6, -0.18), (7, 0.242), (8, 0.019), (9, -0.035), (10, -0.035), (11, 0.016), (12, -0.157), (13, -0.053), (14, -0.11), (15, -0.025), (16, 0.071), (17, 0.021), (18, 0.025), (19, 0.029), (20, -0.094), (21, -0.018), (22, 0.058), (23, 0.092), (24, 0.014), (25, 0.016), (26, 0.125), (27, -0.127), (28, 0.033), (29, 0.125), (30, 0.17), (31, -0.057), (32, -0.009), (33, -0.037), (34, 0.05), (35, 0.025), (36, 0.038), (37, 0.017), (38, -0.002), (39, -0.088), (40, -0.059), (41, 0.05), (42, 0.023), (43, -0.028), (44, -0.028), (45, 0.021), (46, 0.005), (47, -0.081), (48, 0.008), (49, 0.021)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.96385121 21 acl-2011-A Pilot Study of Opinion Summarization in Conversations

Author: Dong Wang ; Yang Liu

Abstract: This paper presents a pilot study of opinion summarization on conversations. We create a corpus containing extractive and abstractive summaries of speaker’s opinion towards a given topic using 88 telephone conversations. We adopt two methods to perform extractive summarization. The first one is a sentence-ranking method that linearly combines scores measured from different aspects including topic relevance, subjectivity, and sentence importance. The second one is a graph-based method, which incorporates topic and sentiment information, as well as additional information about sentence-to-sentence relations extracted based on dialogue structure. Our evaluation results show that both methods significantly outperform the baseline approach that extracts the longest utterances. In particular, we find that incorporating dialogue structure in the graph-based method contributes to the improved system performance.

2 0.65855318 76 acl-2011-Comparative News Summarization Using Linear Programming

Author: Xiaojiang Huang ; Xiaojun Wan ; Jianguo Xiao

Abstract: Comparative News Summarization aims to highlight the commonalities and differences between two comparable news topics. In this study, we propose a novel approach to generating comparative news summaries. We formulate the task as an optimization problem of selecting proper sentences to maximize the comparativeness within the summary and the representativeness to both news topics. We consider semantic-related cross-topic concept pairs as comparative evidences, and consider topic-related concepts as representative evidences. The optimization problem is addressed by using a linear programming model. The experimental results demonstrate the effectiveness of our proposed model.

3 0.65177643 187 acl-2011-Jointly Learning to Extract and Compress

Author: Taylor Berg-Kirkpatrick ; Dan Gillick ; Dan Klein

Abstract: We learn a joint model of sentence extraction and compression for multi-document summarization. Our model scores candidate summaries according to a combined linear model whose features factor over (1) the n-gram types in the summary and (2) the compressions used. We train the model using a marginbased objective whose loss captures end summary quality. Because of the exponentially large set of candidate summaries, we use a cutting-plane algorithm to incrementally detect and add active constraints efficiently. Inference in our model can be cast as an ILP and thereby solved in reasonable time; we also present a fast approximation scheme which achieves similar performance. Our jointly extracted and compressed summaries outperform both unlearned baselines and our learned extraction-only system on both ROUGE and Pyramid, without a drop in judged linguistic quality. We achieve the highest published ROUGE results to date on the TAC 2008 data set.

4 0.63731319 308 acl-2011-Towards a Framework for Abstractive Summarization of Multimodal Documents

Author: Charles Greenbacker

Abstract: We propose a framework for generating an abstractive summary from a semantic model of a multimodal document. We discuss the type of model required, the means by which it can be constructed, how the content of the model is rated and selected, and the method of realizing novel sentences for the summary. To this end, we introduce a metric called information density used for gauging the importance of content obtained from text and graphical sources.

5 0.60912138 270 acl-2011-SciSumm: A Multi-Document Summarization System for Scientific Articles

Author: Nitin Agarwal ; Ravi Shankar Reddy ; Kiran GVR ; Carolyn Penstein Rose

Abstract: In this demo, we present SciSumm, an interactive multi-document summarization system for scientific articles. The document collection to be summarized is a list of papers cited together within the same source article, otherwise known as a co-citation. At the heart of the approach is a topic based clustering of fragments extracted from each article based on queries generated from the context surrounding the co-cited list of papers. This analysis enables the generation of an overview of common themes from the co-cited papers that relate to the context in which the co-citation was found. SciSumm is currently built over the 2008 ACL Anthology, however the gen- eralizable nature of the summarization techniques and the extensible architecture makes it possible to use the system with other corpora where a citation network is available. Evaluation results on the same corpus demonstrate that our system performs better than an existing widely used multi-document summarization system (MEAD).

6 0.60046798 156 acl-2011-IMASS: An Intelligent Microblog Analysis and Summarization System

7 0.59082609 47 acl-2011-Automatic Assessment of Coverage Quality in Intelligence Reports

8 0.58970034 211 acl-2011-Liars and Saviors in a Sentiment Annotated Corpus of Comments to Political Debates

9 0.58711451 251 acl-2011-Probabilistic Document Modeling for Syntax Removal in Text Summarization

10 0.58650184 255 acl-2011-Query Snowball: A Co-occurrence-based Approach to Multi-document Summarization for Question Answering

11 0.57380152 201 acl-2011-Learning From Collective Human Behavior to Introduce Diversity in Lexical Choice

12 0.57159644 136 acl-2011-Finding Deceptive Opinion Spam by Any Stretch of the Imagination

13 0.56634355 159 acl-2011-Identifying Noun Product Features that Imply Opinions

14 0.55970091 98 acl-2011-Discovery of Topically Coherent Sentences for Extractive Summarization

15 0.54762906 4 acl-2011-A Class of Submodular Functions for Document Summarization

16 0.53214478 326 acl-2011-Using Bilingual Information for Cross-Language Document Summarization

17 0.52711284 131 acl-2011-Extracting Opinion Expressions and Their Polarities - Exploration of Pipelines and Joint Models

18 0.50901133 84 acl-2011-Contrasting Opposing Views of News Articles on Contentious Issues

19 0.4976103 260 acl-2011-Recognizing Authority in Dialogue with an Integer Linear Programming Constrained Model

20 0.47957402 45 acl-2011-Aspect Ranking: Identifying Important Product Aspects from Online Consumer Reviews


similar papers computed by lda model

lda for this paper:

topicId topicWeight

[(5, 0.05), (15, 0.172), (17, 0.07), (26, 0.013), (31, 0.03), (37, 0.084), (39, 0.037), (41, 0.058), (53, 0.033), (55, 0.021), (59, 0.026), (72, 0.037), (91, 0.032), (96, 0.234)]

similar papers list:

simIndex simValue paperId paperTitle

1 0.9416641 138 acl-2011-French TimeBank: An ISO-TimeML Annotated Reference Corpus

Author: Andre Bittar ; Pascal Amsili ; Pascal Denis ; Laurence Danlos

Abstract: This article presents the main points in the creation of the French TimeBank (Bittar, 2010), a reference corpus annotated according to the ISO-TimeML standard for temporal annotation. A number of improvements were made to the markup language to deal with linguistic phenomena not yet covered by ISO-TimeML, including cross-language modifications and others specific to French. An automatic preannotation system was used to speed up the annotation process. A preliminary evaluation of the methodology adopted for this project yields positive results in terms of data quality and annotation time.

2 0.91340601 287 acl-2011-Structural Topic Model for Latent Topical Structure Analysis

Author: Hongning Wang ; Duo Zhang ; ChengXiang Zhai

Abstract: Topic models have been successfully applied to many document analysis tasks to discover topics embedded in text. However, existing topic models generally cannot capture the latent topical structures in documents. Since languages are intrinsically cohesive and coherent, modeling and discovering latent topical transition structures within documents would be beneficial for many text analysis tasks. In this work, we propose a new topic model, Structural Topic Model, which simultaneously discovers topics and reveals the latent topical structures in text through explicitly modeling topical transitions with a latent first-order Markov chain. Experiment results show that the proposed Structural Topic Model can effectively discover topical structures in text, and the identified structures significantly improve the performance of tasks such as sentence annotation and sentence ordering. ,

same-paper 3 0.8973043 21 acl-2011-A Pilot Study of Opinion Summarization in Conversations

Author: Dong Wang ; Yang Liu

Abstract: This paper presents a pilot study of opinion summarization on conversations. We create a corpus containing extractive and abstractive summaries of speaker’s opinion towards a given topic using 88 telephone conversations. We adopt two methods to perform extractive summarization. The first one is a sentence-ranking method that linearly combines scores measured from different aspects including topic relevance, subjectivity, and sentence importance. The second one is a graph-based method, which incorporates topic and sentiment information, as well as additional information about sentence-to-sentence relations extracted based on dialogue structure. Our evaluation results show that both methods significantly outperform the baseline approach that extracts the longest utterances. In particular, we find that incorporating dialogue structure in the graph-based method contributes to the improved system performance.

4 0.87338442 269 acl-2011-Scaling up Automatic Cross-Lingual Semantic Role Annotation

Author: Lonneke van der Plas ; Paola Merlo ; James Henderson

Abstract: Broad-coverage semantic annotations for training statistical learners are only available for a handful of languages. Previous approaches to cross-lingual transfer of semantic annotations have addressed this problem with encouraging results on a small scale. In this paper, we scale up previous efforts by using an automatic approach to semantic annotation that does not rely on a semantic ontology for the target language. Moreover, we improve the quality of the transferred semantic annotations by using a joint syntacticsemantic parser that learns the correlations between syntax and semantics of the target language and smooths out the errors from automatic transfer. We reach a labelled F-measure for predicates and arguments of only 4% and 9% points, respectively, lower than the upper bound from manual annotations.

5 0.83002985 110 acl-2011-Effective Use of Function Words for Rule Generalization in Forest-Based Translation

Author: Xianchao Wu ; Takuya Matsuzaki ; Jun'ichi Tsujii

Abstract: In the present paper, we propose the effective usage of function words to generate generalized translation rules for forest-based translation. Given aligned forest-string pairs, we extract composed tree-to-string translation rules that account for multiple interpretations of both aligned and unaligned target function words. In order to constrain the exhaustive attachments of function words, we limit to bind them to the nearby syntactic chunks yielded by a target dependency parser. Therefore, the proposed approach can not only capture source-tree-to-target-chunk correspondences but can also use forest structures that compactly encode an exponential number of parse trees to properly generate target function words during decoding. Extensive experiments involving large-scale English-toJapanese translation revealed a significant im- provement of 1.8 points in BLEU score, as compared with a strong forest-to-string baseline system.

6 0.82928371 87 acl-2011-Corpus Expansion for Statistical Machine Translation with Semantic Role Label Substitution Rules

7 0.82844448 327 acl-2011-Using Bilingual Parallel Corpora for Cross-Lingual Textual Entailment

8 0.82795763 225 acl-2011-Monolingual Alignment by Edit Rate Computation on Sentential Paraphrase Pairs

9 0.82772291 72 acl-2011-Collecting Highly Parallel Data for Paraphrase Evaluation

10 0.82756853 308 acl-2011-Towards a Framework for Abstractive Summarization of Multimodal Documents

11 0.8265909 171 acl-2011-Incremental Syntactic Language Models for Phrase-based Translation

12 0.82594955 16 acl-2011-A Joint Sequence Translation Model with Integrated Reordering

13 0.82494551 318 acl-2011-Unsupervised Bilingual Morpheme Segmentation and Alignment with Context-rich Hidden Semi-Markov Models

14 0.82453001 61 acl-2011-Binarized Forest to String Translation

15 0.82284415 37 acl-2011-An Empirical Evaluation of Data-Driven Paraphrase Generation Techniques

16 0.82254612 81 acl-2011-Consistent Translation using Discriminative Learning - A Translation Memory-inspired Approach

17 0.82193226 101 acl-2011-Disentangling Chat with Local Coherence Models

18 0.82186359 221 acl-2011-Model-Based Aligner Combination Using Dual Decomposition

19 0.82180953 310 acl-2011-Translating from Morphologically Complex Languages: A Paraphrase-Based Approach

20 0.82162768 155 acl-2011-Hypothesis Mixture Decoding for Statistical Machine Translation