acl acl2010 acl2010-134 knowledge-graph by maker-knowledge-mining

134 acl-2010-Hierarchical Sequential Learning for Extracting Opinions and Their Attributes


Source: pdf

Author: Yejin Choi ; Claire Cardie

Abstract: Automatic opinion recognition involves a number of related tasks, such as identifying the boundaries of opinion expression, determining their polarity, and determining their intensity. Although much progress has been made in this area, existing research typically treats each of the above tasks in isolation. In this paper, we apply a hierarchical parameter sharing technique using Conditional Random Fields for fine-grained opinion analysis, jointly detecting the boundaries of opinion expressions as well as determining two of their key attributes polarity and intensity. Our experimental results show that our proposed approach improves the performance over a baseline that does not — exploit hierarchical structure among the classes. In addition, we find that the joint approach outperforms a baseline that is based on cascading two separate components.

Reference: text


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 edu l Abstract Automatic opinion recognition involves a number of related tasks, such as identifying the boundaries of opinion expression, determining their polarity, and determining their intensity. [sent-3, score-1.216]

2 Although much progress has been made in this area, existing research typically treats each of the above tasks in isolation. [sent-4, score-0.028]

3 In this paper, we apply a hierarchical parameter sharing technique using Conditional Random Fields for fine-grained opinion analysis, jointly detecting the boundaries of opinion expressions as well as determining two of their key attributes polarity and intensity. [sent-5, score-1.819]

4 Our experimental results show that our proposed approach improves the performance over a baseline that does not — exploit hierarchical structure among the classes. [sent-6, score-0.095]

5 In addition, we find that the joint approach outperforms a baseline that is based on cascading two separate components. [sent-7, score-0.133]

6 1 Introduction Automatic opinion recognition involves a number of related tasks, such as identifying expressions of opinion (e. [sent-8, score-1.097]

7 Most previous work treats each subtask in isolation: opinion expression extraction (i. [sent-18, score-0.625]

8 detecting the boundaries of opinion expressions) and opinion attribute classification (e. [sent-20, score-1.153]

9 determining values for polarity and intensity) are tackled as separate steps in opinion recognition systems. [sent-22, score-0.888]

10 Unfortunately, errors from individual components will propagate in systems with cascaded component architectures, causing performance degradation in the end-toend system (e. [sent-23, score-0.044]

11 (2006)) in our case, in the end-to-end opinion recognition system. [sent-26, score-0.516]

12 In this paper, we apply a hierarchical parameter sharing technique (e. [sent-27, score-0.187]

13 In particular, we aim to jointly identify the boundaries of opinion expressions as well as to determine two of their key attributes polarity and intensity. [sent-32, score-1.041]

14 Experimental results show that our proposed approach improves the performance over the baseline that does not exploit the hierarchical structure among the classes. [sent-33, score-0.095]

15 In addition, we find that the joint approach outperforms a baseline that is based on cascading two separate systems. [sent-34, score-0.133]

16 — — 2 Hierarchical Sequential Learning We define the problem of joint extraction of opinion expressions and their attributes as a sequence tagging task as follows. [sent-35, score-0.832]

17 , 9} are defined as conjunctive values of∈ polarity 9la}b ealrse and intensity labels, as shown in Table 1. [sent-45, score-0.581]

18 Then the conditional probability p(y|x) for linear-chain tChReF cos nisd given as (Lafferty e pt( ayl|. [sent-46, score-0.026]

19 In order to apply a hierarchical parameter sharing technique (e. [sent-50, score-0.187]

20 eN soatme etha cto mthpeorecan be other variations of hierarchical construction. [sent-82, score-0.117]

21 tnfoedanetpficonale raditsyetpicnaorgmautipesoh ne tanoctfhpan larbdameal,neti rnis- For instance, if yi = 1, then 3 λ f(1, x, i) = λOPINION gO Features (OPINION, x, i) + λPOSITIVE + gP(POSITVE, We first introduce definitions of key terms that will be used to describe features. [sent-90, score-0.081]

22 = λN′O-OPINION,OPINION gO′(NO-OPINION, OPINION, x, i) • EXP-POLARITY, EXP-INTENSITY & EXP-SPAN: + λN′O-POLARITY, gP′(NO-POLARITY, NEUTRAL, x, i) Words in a given opinion expression often do + λN′O-INTENSITY, gS′ (NO-INTENSITY, i) not share the same prior-attributes. [sent-93, score-0.553]

23 Such discontinuous distribution of features can make it harder to learn the desired opinion expresThis hierarchical construction of feature and sion boundaries. [sent-94, score-0.674]

24 Therefore, we try to obtain weight vectors allows similar labels to share the expression-level attributes (EXP-POLARITY and same subcomponents of feature and weight vecEXP-INTENSITY) using simple heuristics. [sent-95, score-0.143]

25 The text span with the same expression-level attributes are referred to as EXP-SPAN. [sent-100, score-0.094]

26 1 Per-Token Features Per-token features are defined in the form of gO (α, x, i) , gP (β, x, i) and gS (γ, x, i). [sent-102, score-0.044]

27 Common Per-Token Features Following features are common for all class labels. [sent-104, score-0.067]

28 The notation ⊗ indicates conjunctive operation of tTwhoe nvaoltuaetiso. [sent-105, score-0.04]

29 • OPINION-LEXICON(xi): based on opinion lexicon (Wiebe et al. [sent-109, score-0.491]

30 • PRIOR-POLARITY(xi) ⊗ PRIOR-INTENSITY(xi) • EXP-POLARITY(xi) ⊗ EXP-INTENSITY(xi) • EXP-POLARITY(xi) ⊗ EXP-INTENSITY(xi) ⊗ STEM(xi) • EXP-SPAN(xi): boolean to indicate whether xi is in an EXP-SPAN. [sent-112, score-0.498]

31 • EXP-POLARITY(xi) ⊗ EXP-INTENSITY(xi) ⊗ EXP-SPAN(xi) Polarity Per-Token Features These features are included only for gO (α, x, i) and gP(β, x, i), which are the feature functions corresponding to the polarity-based classes. [sent-114, score-0.044]

32 This feature en∈cod {epso sthiteiv en,u nmebuetrra o,f n positive, neutral, and negative EXP-POLARITY words respectively, in the current sentence. [sent-116, score-0.03]

33 • PRIOR-INTENSITY(xi), EXP-INTENSITY(xi) • STEM(xi) ⊗ EXP-INTENSITY(xi) COUNT-OF-STRONG, COUNT-OF-WEAK: the number of strong and weak EXP-INTENSITY words in the current sentence. [sent-118, score-0.03]

34 • INTENSIFIER(xi): whether xi is an intensifier, such as “extremely”, “highly”, “really”. [sent-119, score-0.498]

35 • STRONGMODAL(xi): whether xi is a strong modal verb, such as “must”, “can”, “will”. [sent-120, score-0.529]

36 • WEAKMODAL(xi): whether xi is a weak modal verb, such as “may”, “could”, “would”. [sent-121, score-0.559]

37 • DIMINISHER(xi): whether xi is a diminisher, such as “little”, “somewhat”, “less”. [sent-122, score-0.498]

38 2 Transition Features Transition features are employed to help with boundary extraction as follows: • ⊗ ⊗ ⊗ ⊗ ⊗ ⊗ Polarity Transition Features Polarity transition features are features that used only for gO′ (α, αˆ , x, i) and gP′(β, x, i). [sent-124, score-0.225]

39 • PART-OF-SPEECH(xi) ⊗ PART-OF-SPEECH(xi+1) EXP-POLARITY(xi) • EXP-POLARITY(xi) ⊗ EXP-POLARITY(xi+1) Intensity Transition Features Intensity transition features are features that used only for gO′ (α, αˆ , x, i) and gS′ (γ, γˆ, x, i). [sent-125, score-0.137]

40 Our gold standard opinion expressions cor1The MPQA corpus can be obtained http://nrrc. [sent-127, score-0.581]

41 271 Method Descriptionr(%)Ppo (s%iti)ve f(%)r(%)N pe(u%tr)al f(%)r(%)N peg(a%ti)ve f(%) JJointialioP n tr wtyiwithlnohO-t uyt HHierarchy31. [sent-131, score-0.084]

42 63 Table 2: Performance of Opinion Extraction with Correct Polarity Attribute at HighMediumLow Method Descriptionr(%) p(%) f(%)r(%) p(%) f(%)r(%) p(%) f(%) JJointialioP n tr wtyiwithlnohO-t uyt HHierarchy27. [sent-149, score-0.084]

43 0 Table 4: Performance of Opinion Extraction respond to direct subjective expression and expressive subjective element (Wiebe et al. [sent-176, score-0.11]

44 2 Our implementation of hierarchical sequential learning is based on the Mallet (McCallum, 2002) code for CRFs. [sent-178, score-0.194]

45 We then combine the results from two separate CRFs by collecting all opinion entities extracted by both sequence taggers. [sent-183, score-0.541]

46 5% of the polarity annotations correspond to both; hence, we merge both into the neutral. [sent-185, score-0.294]

47 Similarly, for gold standard intensity, we merge extremely high into high. [sent-186, score-0.023]

48 [Baseline-2] Joint without Hierarchy: Here we use simple linear-chain CRFs without exploiting the class hierarchy for the opinion recognition task. [sent-189, score-0.633]

49 Joint with Hierarchy: Finally, we test the hierarchical sequential learning approach elaborated in Section 3. [sent-191, score-0.194]

50 1 Evaluation Results We evaluate all experiments at the opinion entity level, i. [sent-193, score-0.491]

51 at the level of each opinion expression rather than at the token level. [sent-195, score-0.553]

52 Table 4 shows the performance of opinion extraction without matching any attribute. [sent-197, score-0.535]

53 That is, an extracted opinion entity is counted as correct if it overlaps4 with a gold standard opinion expression, without checking the correctness of its attributes. [sent-198, score-0.982]

54 Table 2 and 3 show the performance of opinion extraction with the correct polarity and intensity respectively. [sent-199, score-1.076]

55 One might wonder whether the overlap matching scheme could allow a degenerative case where extracting the entire test dataset as one giant opinion expression would yield 100% recall and precision. [sent-202, score-0.632]

56 Because each sentence corresponds to a different test instance in our model, and because some sentences do not contain any opinion expression in the dataset, such degenerative case is not possible in our experiments. [sent-203, score-0.603]

57 272 HIERARCHY performs the best, and the least effective one is BASELINE-1, which cascades two separately trained models. [sent-204, score-0.055]

58 It is interesting that the simple sequential tagging approach even without exploiting the hierarchy (BASELINE-2) performs better than the cascaded approach (BASELINE-1). [sent-205, score-0.277]

59 When evaluating with respect to the polarity attribute, the performance of the negative class is substantially higher than the that of other classes. [sent-206, score-0.324]

60 This is not surprising as there is approximately twice as much data for the negative class. [sent-207, score-0.03]

61 When evaluating with respect to the intensity attribute, the performance of the LOW class is substantially lower than that of other classes. [sent-208, score-0.293]

62 This result reflects the fact that it is inherently harder to distinguish an opinion expression with low intensity from no opinion. [sent-209, score-0.867]

63 In general, we observe that determining correct intensity attributes is a much harder task than determining correct polarity attributes. [sent-210, score-0.827]

64 Remind that neither of these models alone fully solve the joint task of extracting boundaries as well as determining two attributions simultaneously. [sent-219, score-0.214]

65 We conclude from our experiments that the simple joint sequential tagging approach even without exploiting the hierarchy brings a better performance than combining two separately developed systems. [sent-221, score-0.352]

66 In addition, our hierarchical joint sequential learning approach brings a further perfor- mance gain over the simple joint sequential tagging method. [sent-222, score-0.469]

67 5 Related Work Although there have been much research for finegrained opinion analysis (e. [sent-223, score-0.542]

68 The hierarchical parameter sharing technique used in this paper has been previously used by Zhao et al. [sent-234, score-0.187]

69 (2008) employs this technique only to classify sentence-level attributes (polarity and intensity), without involving a much harder task of detecting boundaries of sub-sentential entities. [sent-237, score-0.267]

70 6 Conclusion We applied a hierarchical parameter sharing technique using Conditional Random Fields for finegrained opinion analysis. [sent-238, score-0.729]

71 Our proposed approach jointly extract opinion expressions from unstructured text and determine their attributes polarity and intensity. [sent-239, score-1.005]

72 Empirical results indicate that the simple joint sequential tagging approach even without exploiting the hierarchy brings a better performance than combining two separately developed systems. [sent-240, score-0.352]

73 In addition, we found that the hierarchical joint sequential learning approach improves the performance over the simple joint sequential tagging method. [sent-241, score-0.433]

74 (2005) evaluate only on known words that are in their opinion lexicon. [sent-258, score-0.491]

75 (2005) simplifies the problem by combining neutral opinions and no opinions into the same class, while our system distinguishes the two. [sent-260, score-0.18]

76 Recognizing Contextual Polarity: an exploration of features for phrase-level sentiment analysis. [sent-375, score-0.096]


similar papers computed by tfidf model

tfidf for this paper:

wordName wordTfidf (topN-words)

[('xi', 0.498), ('opinion', 0.491), ('polarity', 0.271), ('intensity', 0.27), ('gp', 0.218), ('wilson', 0.193), ('gs', 0.14), ('wiebe', 0.102), ('sequential', 0.099), ('hierarchical', 0.095), ('attributes', 0.094), ('neutral', 0.092), ('expressions', 0.09), ('attribute', 0.084), ('yi', 0.081), ('go', 0.076), ('descriptionr', 0.075), ('diminisher', 0.075), ('intensifier', 0.075), ('determining', 0.074), ('choi', 0.068), ('hierarchy', 0.065), ('expression', 0.062), ('boundaries', 0.061), ('crfs', 0.058), ('cascading', 0.056), ('zhao', 0.055), ('stem', 0.055), ('sentiment', 0.052), ('cai', 0.051), ('finegrained', 0.051), ('sharing', 0.05), ('degenerative', 0.05), ('ishncrare', 0.05), ('jjointialiop', 0.05), ('positve', 0.05), ('strongmodal', 0.05), ('uyt', 0.05), ('weakmodal', 0.05), ('yty', 0.05), ('joint', 0.05), ('transition', 0.049), ('breck', 0.045), ('features', 0.044), ('extraction', 0.044), ('opinions', 0.044), ('cascaded', 0.044), ('harder', 0.044), ('technique', 0.042), ('kim', 0.041), ('tagging', 0.04), ('conjunctive', 0.04), ('cardie', 0.038), ('lafferty', 0.037), ('brings', 0.036), ('popescu', 0.036), ('mpqa', 0.035), ('hu', 0.034), ('tr', 0.034), ('jointly', 0.034), ('separately', 0.033), ('cunningham', 0.032), ('mallet', 0.032), ('fields', 0.032), ('gate', 0.031), ('modal', 0.031), ('extracts', 0.031), ('hofmann', 0.03), ('negative', 0.03), ('weak', 0.03), ('claire', 0.029), ('extracting', 0.029), ('exploiting', 0.029), ('recognizing', 0.029), ('treats', 0.028), ('vectors', 0.027), ('gi', 0.027), ('separate', 0.027), ('conditional', 0.026), ('detecting', 0.026), ('positive', 0.026), ('recognition', 0.025), ('medium', 0.025), ('unstructured', 0.025), ('subjective', 0.024), ('xp', 0.024), ('finkel', 0.024), ('etzioni', 0.023), ('sequence', 0.023), ('class', 0.023), ('merge', 0.023), ('liu', 0.022), ('strength', 0.022), ('corne', 0.022), ('ychoi', 0.022), ('subcomponents', 0.022), ('cascades', 0.022), ('remind', 0.022), ('ayl', 0.022), ('etha', 0.022)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 1.0000002 134 acl-2010-Hierarchical Sequential Learning for Extracting Opinions and Their Attributes

Author: Yejin Choi ; Claire Cardie

Abstract: Automatic opinion recognition involves a number of related tasks, such as identifying the boundaries of opinion expression, determining their polarity, and determining their intensity. Although much progress has been made in this area, existing research typically treats each of the above tasks in isolation. In this paper, we apply a hierarchical parameter sharing technique using Conditional Random Fields for fine-grained opinion analysis, jointly detecting the boundaries of opinion expressions as well as determining two of their key attributes polarity and intensity. Our experimental results show that our proposed approach improves the performance over a baseline that does not — exploit hierarchical structure among the classes. In addition, we find that the joint approach outperforms a baseline that is based on cascading two separate components.

2 0.4330577 22 acl-2010-A Unified Graph Model for Sentence-Based Opinion Retrieval

Author: Binyang Li ; Lanjun Zhou ; Shi Feng ; Kam-Fai Wong

Abstract: There is a growing research interest in opinion retrieval as on-line users’ opinions are becoming more and more popular in business, social networks, etc. Practically speaking, the goal of opinion retrieval is to retrieve documents, which entail opinions or comments, relevant to a target subject specified by the user’s query. A fundamental challenge in opinion retrieval is information representation. Existing research focuses on document-based approaches and documents are represented by bag-of-word. However, due to loss of contextual information, this representation fails to capture the associative information between an opinion and its corresponding target. It cannot distinguish different degrees of a sentiment word when associated with different targets. This in turn seriously affects opinion retrieval performance. In this paper, we propose a sentence-based approach based on a new information representa- , tion, namely topic-sentiment word pair, to capture intra-sentence contextual information between an opinion and its target. Additionally, we consider inter-sentence information to capture the relationships among the opinions on the same topic. Finally, the two types of information are combined in a unified graph-based model, which can effectively rank the documents. Compared with existing approaches, experimental results on the COAE08 dataset showed that our graph-based model achieved significant improvement. 1

3 0.35527474 208 acl-2010-Sentence and Expression Level Annotation of Opinions in User-Generated Discourse

Author: Cigdem Toprak ; Niklas Jakob ; Iryna Gurevych

Abstract: In this paper, we introduce a corpus of consumer reviews from the rateitall and the eopinions websites annotated with opinion-related information. We present a two-level annotation scheme. In the first stage, the reviews are analyzed at the sentence level for (i) relevancy to a given topic, and (ii) expressing an evaluation about the topic. In the second stage, on-topic sentences containing evaluations about the topic are further investigated at the expression level for pinpointing the properties (semantic orientation, intensity), and the functional components of the evaluations (opinion terms, targets and holders). We discuss the annotation scheme, the inter-annotator agreement for different subtasks and our observations.

4 0.33465701 251 acl-2010-Using Anaphora Resolution to Improve Opinion Target Identification in Movie Reviews

Author: Niklas Jakob ; Iryna Gurevych

Abstract: unkown-abstract

5 0.2159576 123 acl-2010-Generating Focused Topic-Specific Sentiment Lexicons

Author: Valentin Jijkoun ; Maarten de Rijke ; Wouter Weerkamp

Abstract: We present a method for automatically generating focused and accurate topicspecific subjectivity lexicons from a general purpose polarity lexicon that allow users to pin-point subjective on-topic information in a set of relevant documents. We motivate the need for such lexicons in the field of media analysis, describe a bootstrapping method for generating a topic-specific lexicon from a general purpose polarity lexicon, and evaluate the quality of the generated lexicons both manually and using a TREC Blog track test set for opinionated blog post retrieval. Although the generated lexicons can be an order of magnitude more selective than the general purpose lexicon, they maintain, or even improve, the performance of an opin- ion retrieval system.

6 0.144263 209 acl-2010-Sentiment Learning on Product Reviews via Sentiment Ontology Tree

7 0.13127705 141 acl-2010-Identifying Text Polarity Using Random Walks

8 0.09976349 132 acl-2010-Hierarchical Joint Learning: Improving Joint Parsing and Named Entity Recognition with Non-Jointly Labeled Data

9 0.090058893 18 acl-2010-A Study of Information Retrieval Weighting Schemes for Sentiment Analysis

10 0.087887801 42 acl-2010-Automatically Generating Annotator Rationales to Improve Sentiment Classification

11 0.086773895 161 acl-2010-Learning Better Data Representation Using Inference-Driven Metric Learning

12 0.085336655 105 acl-2010-Evaluating Multilanguage-Comparability of Subjectivity Analysis Systems

13 0.083519086 157 acl-2010-Last but Definitely Not Least: On the Role of the Last Sentence in Automatic Polarity-Classification

14 0.073497064 197 acl-2010-Practical Very Large Scale CRFs

15 0.06763681 113 acl-2010-Extraction and Approximation of Numerical Attributes from the Web

16 0.066420719 210 acl-2010-Sentiment Translation through Lexicon Induction

17 0.060260288 255 acl-2010-Viterbi Training for PCFGs: Hardness Results and Competitiveness of Uniform Initialization

18 0.055657145 245 acl-2010-Understanding the Semantic Structure of Noun Phrase Queries

19 0.054975357 256 acl-2010-Vocabulary Choice as an Indicator of Perspective

20 0.054498523 211 acl-2010-Simple, Accurate Parsing with an All-Fragments Grammar


similar papers computed by lsi model

lsi for this paper:

topicId topicWeight

[(0, -0.158), (1, 0.137), (2, -0.243), (3, 0.292), (4, -0.25), (5, 0.06), (6, -0.04), (7, 0.214), (8, 0.053), (9, -0.015), (10, 0.023), (11, -0.063), (12, 0.052), (13, 0.048), (14, 0.011), (15, -0.086), (16, -0.202), (17, 0.174), (18, 0.001), (19, 0.066), (20, 0.044), (21, 0.086), (22, -0.06), (23, 0.086), (24, 0.002), (25, -0.005), (26, -0.017), (27, 0.012), (28, 0.13), (29, 0.031), (30, -0.059), (31, -0.004), (32, 0.045), (33, -0.028), (34, 0.023), (35, 0.115), (36, -0.032), (37, -0.021), (38, 0.028), (39, -0.038), (40, -0.033), (41, -0.052), (42, 0.01), (43, -0.016), (44, -0.017), (45, -0.065), (46, -0.011), (47, 0.08), (48, -0.011), (49, -0.025)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.97883934 134 acl-2010-Hierarchical Sequential Learning for Extracting Opinions and Their Attributes

Author: Yejin Choi ; Claire Cardie

Abstract: Automatic opinion recognition involves a number of related tasks, such as identifying the boundaries of opinion expression, determining their polarity, and determining their intensity. Although much progress has been made in this area, existing research typically treats each of the above tasks in isolation. In this paper, we apply a hierarchical parameter sharing technique using Conditional Random Fields for fine-grained opinion analysis, jointly detecting the boundaries of opinion expressions as well as determining two of their key attributes polarity and intensity. Our experimental results show that our proposed approach improves the performance over a baseline that does not — exploit hierarchical structure among the classes. In addition, we find that the joint approach outperforms a baseline that is based on cascading two separate components.

2 0.90674305 251 acl-2010-Using Anaphora Resolution to Improve Opinion Target Identification in Movie Reviews

Author: Niklas Jakob ; Iryna Gurevych

Abstract: unkown-abstract

3 0.85226434 22 acl-2010-A Unified Graph Model for Sentence-Based Opinion Retrieval

Author: Binyang Li ; Lanjun Zhou ; Shi Feng ; Kam-Fai Wong

Abstract: There is a growing research interest in opinion retrieval as on-line users’ opinions are becoming more and more popular in business, social networks, etc. Practically speaking, the goal of opinion retrieval is to retrieve documents, which entail opinions or comments, relevant to a target subject specified by the user’s query. A fundamental challenge in opinion retrieval is information representation. Existing research focuses on document-based approaches and documents are represented by bag-of-word. However, due to loss of contextual information, this representation fails to capture the associative information between an opinion and its corresponding target. It cannot distinguish different degrees of a sentiment word when associated with different targets. This in turn seriously affects opinion retrieval performance. In this paper, we propose a sentence-based approach based on a new information representa- , tion, namely topic-sentiment word pair, to capture intra-sentence contextual information between an opinion and its target. Additionally, we consider inter-sentence information to capture the relationships among the opinions on the same topic. Finally, the two types of information are combined in a unified graph-based model, which can effectively rank the documents. Compared with existing approaches, experimental results on the COAE08 dataset showed that our graph-based model achieved significant improvement. 1

4 0.84216362 208 acl-2010-Sentence and Expression Level Annotation of Opinions in User-Generated Discourse

Author: Cigdem Toprak ; Niklas Jakob ; Iryna Gurevych

Abstract: In this paper, we introduce a corpus of consumer reviews from the rateitall and the eopinions websites annotated with opinion-related information. We present a two-level annotation scheme. In the first stage, the reviews are analyzed at the sentence level for (i) relevancy to a given topic, and (ii) expressing an evaluation about the topic. In the second stage, on-topic sentences containing evaluations about the topic are further investigated at the expression level for pinpointing the properties (semantic orientation, intensity), and the functional components of the evaluations (opinion terms, targets and holders). We discuss the annotation scheme, the inter-annotator agreement for different subtasks and our observations.

5 0.65568125 123 acl-2010-Generating Focused Topic-Specific Sentiment Lexicons

Author: Valentin Jijkoun ; Maarten de Rijke ; Wouter Weerkamp

Abstract: We present a method for automatically generating focused and accurate topicspecific subjectivity lexicons from a general purpose polarity lexicon that allow users to pin-point subjective on-topic information in a set of relevant documents. We motivate the need for such lexicons in the field of media analysis, describe a bootstrapping method for generating a topic-specific lexicon from a general purpose polarity lexicon, and evaluate the quality of the generated lexicons both manually and using a TREC Blog track test set for opinionated blog post retrieval. Although the generated lexicons can be an order of magnitude more selective than the general purpose lexicon, they maintain, or even improve, the performance of an opin- ion retrieval system.

6 0.39286363 209 acl-2010-Sentiment Learning on Product Reviews via Sentiment Ontology Tree

7 0.34815258 18 acl-2010-A Study of Information Retrieval Weighting Schemes for Sentiment Analysis

8 0.33780134 42 acl-2010-Automatically Generating Annotator Rationales to Improve Sentiment Classification

9 0.32429248 256 acl-2010-Vocabulary Choice as an Indicator of Perspective

10 0.31594497 141 acl-2010-Identifying Text Polarity Using Random Walks

11 0.26724559 140 acl-2010-Identifying Non-Explicit Citing Sentences for Citation-Based Summarization.

12 0.25752982 197 acl-2010-Practical Very Large Scale CRFs

13 0.25440711 105 acl-2010-Evaluating Multilanguage-Comparability of Subjectivity Analysis Systems

14 0.2540063 161 acl-2010-Learning Better Data Representation Using Inference-Driven Metric Learning

15 0.25031945 85 acl-2010-Detecting Experiences from Weblogs

16 0.24263796 111 acl-2010-Extracting Sequences from the Web

17 0.23863141 132 acl-2010-Hierarchical Joint Learning: Improving Joint Parsing and Named Entity Recognition with Non-Jointly Labeled Data

18 0.22703807 122 acl-2010-Generating Fine-Grained Reviews of Songs from Album Reviews

19 0.22377056 159 acl-2010-Learning 5000 Relational Extractors

20 0.22254467 185 acl-2010-Open Information Extraction Using Wikipedia


similar papers computed by lda model

lda for this paper:

topicId topicWeight

[(7, 0.011), (19, 0.246), (25, 0.064), (42, 0.121), (59, 0.077), (72, 0.035), (73, 0.105), (78, 0.025), (83, 0.104), (84, 0.023), (98, 0.077)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.78729403 134 acl-2010-Hierarchical Sequential Learning for Extracting Opinions and Their Attributes

Author: Yejin Choi ; Claire Cardie

Abstract: Automatic opinion recognition involves a number of related tasks, such as identifying the boundaries of opinion expression, determining their polarity, and determining their intensity. Although much progress has been made in this area, existing research typically treats each of the above tasks in isolation. In this paper, we apply a hierarchical parameter sharing technique using Conditional Random Fields for fine-grained opinion analysis, jointly detecting the boundaries of opinion expressions as well as determining two of their key attributes polarity and intensity. Our experimental results show that our proposed approach improves the performance over a baseline that does not — exploit hierarchical structure among the classes. In addition, we find that the joint approach outperforms a baseline that is based on cascading two separate components.

2 0.6895237 39 acl-2010-Automatic Generation of Story Highlights

Author: Kristian Woodsend ; Mirella Lapata

Abstract: In this paper we present a joint content selection and compression model for single-document summarization. The model operates over a phrase-based representation of the source document which we obtain by merging information from PCFG parse trees and dependency graphs. Using an integer linear programming formulation, the model learns to select and combine phrases subject to length, coverage and grammar constraints. We evaluate the approach on the task of generating “story highlights”—a small number of brief, self-contained sentences that allow readers to quickly gather information on news stories. Experimental results show that the model’s output is comparable to human-written highlights in terms of both grammaticality and content.

3 0.64019114 149 acl-2010-Incorporating Extra-Linguistic Information into Reference Resolution in Collaborative Task Dialogue

Author: Ryu Iida ; Syumpei Kobayashi ; Takenobu Tokunaga

Abstract: This paper proposes an approach to reference resolution in situated dialogues by exploiting extra-linguistic information. Recently, investigations of referential behaviours involved in situations in the real world have received increasing attention by researchers (Di Eugenio et al., 2000; Byron, 2005; van Deemter, 2007; Spanger et al., 2009). In order to create an accurate reference resolution model, we need to handle extra-linguistic information as well as textual information examined by existing approaches (Soon et al., 2001 ; Ng and Cardie, 2002, etc.). In this paper, we incorporate extra-linguistic information into an existing corpus-based reference resolution model, and investigate its effects on refer- ence resolution problems within a corpus of Japanese dialogues. The results demonstrate that our proposed model achieves an accuracy of 79.0% for this task.

4 0.62589979 251 acl-2010-Using Anaphora Resolution to Improve Opinion Target Identification in Movie Reviews

Author: Niklas Jakob ; Iryna Gurevych

Abstract: unkown-abstract

5 0.61955202 123 acl-2010-Generating Focused Topic-Specific Sentiment Lexicons

Author: Valentin Jijkoun ; Maarten de Rijke ; Wouter Weerkamp

Abstract: We present a method for automatically generating focused and accurate topicspecific subjectivity lexicons from a general purpose polarity lexicon that allow users to pin-point subjective on-topic information in a set of relevant documents. We motivate the need for such lexicons in the field of media analysis, describe a bootstrapping method for generating a topic-specific lexicon from a general purpose polarity lexicon, and evaluate the quality of the generated lexicons both manually and using a TREC Blog track test set for opinionated blog post retrieval. Although the generated lexicons can be an order of magnitude more selective than the general purpose lexicon, they maintain, or even improve, the performance of an opin- ion retrieval system.

6 0.61776102 150 acl-2010-Inducing Domain-Specific Semantic Class Taggers from (Almost) Nothing

7 0.60802478 208 acl-2010-Sentence and Expression Level Annotation of Opinions in User-Generated Discourse

8 0.60409492 178 acl-2010-Non-Cooperation in Dialogue

9 0.59482318 214 acl-2010-Sparsity in Dependency Grammar Induction

10 0.59409493 231 acl-2010-The Prevalence of Descriptive Referring Expressions in News and Narrative

11 0.58236986 22 acl-2010-A Unified Graph Model for Sentence-Based Opinion Retrieval

12 0.57615292 238 acl-2010-Towards Open-Domain Semantic Role Labeling

13 0.56564337 219 acl-2010-Supervised Noun Phrase Coreference Research: The First Fifteen Years

14 0.56306493 42 acl-2010-Automatically Generating Annotator Rationales to Improve Sentiment Classification

15 0.56178135 230 acl-2010-The Manually Annotated Sub-Corpus: A Community Resource for and by the People

16 0.56162059 118 acl-2010-Fine-Grained Tree-to-String Translation Rule Extraction

17 0.56015468 73 acl-2010-Coreference Resolution with Reconcile

18 0.55744433 85 acl-2010-Detecting Experiences from Weblogs

19 0.55700129 121 acl-2010-Generating Entailment Rules from FrameNet

20 0.55698067 68 acl-2010-Conditional Random Fields for Word Hyphenation