acl acl2011 acl2011-297 knowledge-graph by maker-knowledge-mining

297 acl-2011-That's What She Said: Double Entendre Identification


Source: pdf

Author: Chloe Kiddon ; Yuriy Brun

Abstract: Humor identification is a hard natural language understanding problem. We identify a subproblem — the “that’s what she said” problem with two distinguishing characteristics: (1) use of nouns that are euphemisms for sexually explicit nouns and (2) structure common in the erotic domain. We address this problem in a classification approach that includes features that model those two characteristics. Experiments on web data demonstrate that our approach improves precision by 12% over baseline techniques that use only word-based features. —

Reference: text


Summary: the most important sentenses genereted by tfidf model

sentIndex sentText sentNum sentScore

1 That’s What She Said: Double Entendre Identification Chlo e´ Kiddon and Yuriy Brun Computer Science & Engineering University of Washington Seattle WA 98195-2350 {chloe ,brun}@ cs Abstract Humor identification is a hard natural language understanding problem. [sent-1, score-0.042]

2 We identify a subproblem — the “that’s what she said” problem with two distinguishing characteristics: (1) use of nouns that are euphemisms for sexually explicit nouns and (2) structure common in the erotic domain. [sent-2, score-0.855]

3 Experiments on web data demonstrate that our approach improves precision by 12% over baseline techniques that use only word-based features. [sent-4, score-0.037]

4 The jokes consist of saying “that’s what she said” after someone else utters a statement in a non-sexual context that could also have been used in a sexual context. [sent-7, score-0.127]

5 For example, if Aaron refers to his late-evening basketball practice, saying “I was trying all night, but I just could not get it in! [sent-8, score-0.045]

6 While somewhat juvenile, this joke presents an interesting natural language understanding problem. [sent-10, score-0.049]

7 A “that’s what she said” (TWSS) joke is a type of double entendre. [sent-11, score-0.15]

8 A double entendre, or adianoeta, is an expression that can be understood in two different ways: an innocuous, straightforward way, given the context, and a risqu ´e way that indirectly alludes to a different, indecent context. [sent-12, score-0.101]

9 edu related research has not studied the task of identifying double entendres in text or speech. [sent-15, score-0.134]

10 The task is complex and would require both deep semantic and cultural understanding to recognize the vast array of double entendres. [sent-16, score-0.101]

11 We focus on a subtask of double entendre identification: TWSS recognition. [sent-17, score-0.216]

12 We frame the problem of TWSS recognition as a type of metaphor identification. [sent-19, score-0.1]

13 A metaphor is a figure of speech that creates an analogical mapping between two conceptual domains so that the terminology of one (source) domain can be used to describe situations and objects in the other (target) domain. [sent-20, score-0.168]

14 Usage of the source domain’s terminology in the source domain is literal and is nonliteral in the target domain. [sent-21, score-0.101]

15 Metaphor identification systems seek to differentiate between literal and nonliteral expressions. [sent-22, score-0.117]

16 Some computational approaches to metaphor identification learn selectional preferences of words in multiple domains to help identify nonliteral usage (Mason, 2004; Shutova, 2010). [sent-23, score-0.226]

17 Other approaches train support vector machine (SVM) models on labeled training data to distinguish metaphoric language from literal language (Pasanek and Sculley, 2008). [sent-24, score-0.061]

18 TWSSs also represent mappings between two domains: the innocuous source domain and an erotic target domain. [sent-25, score-0.321]

19 Therefore, we can apply methods from metaphor identification to TWSS identification. [sent-26, score-0.142]

20 In particular, we (1) compare the adjectival selectional preferences of sexually explicit nouns to those of other nouns to determine which nouns may be euphemisms for sexually explicit nouns and (2) Proceedings ofP tohretl 4an9tdh, O Anrneguoanl, M Jueentein 19g- o2f4 t,h 2e0 A1s1s. [sent-27, score-0.965]

21 cc ia2t0io1n1 f Aors Cocoimatpiounta ftoiorn Caolm Lipnugtuaitsiotincasl:s Lhionrgtpuaisptiecrs , pages 89–94, examine the relationship between structures in the erotic domain and nonerotic contexts. [sent-29, score-0.386]

22 We present a novel approach Double Entendre via Noun Transfer (DEviaNT) that applies metaphor identification techniques to solving the double entendre problem and evaluate it on the TWSS problem. [sent-30, score-0.389]

23 DEviaNT classifies individual sentences as either funny if followed by “that’s what she said” or not, which is a type of automatic humor recognition (Mihalcea and Strapparava, 2005; Mihalcea and Pulman, 2007). [sent-31, score-0.095]

24 We argue that in the TWSS domain, high precision is important, while low recall may be tolerated. [sent-32, score-0.037]

25 In experiments on nearly 21K sentences, we find that DEviaNT has 12% higher precision than that of baseline classifiers that use n-gram TWSS models. [sent-33, score-0.059]

26 First, sentences with nouns that are euphemisms for sexually explicit nouns are more likely to be TWSSs. [sent-39, score-0.589]

27 For example, containing the noun “banana” makes a sentence more likely to be a TWSS than containing the noun “door” . [sent-40, score-0.175]

28 Second, TWSSs share common structure with sentences in the erotic domain. [sent-41, score-0.302]

29 Thus, we hypothesize that machine learning with euphemismand structure-based features is a promising approach to solving the TWSS problem. [sent-43, score-0.051]

30 Accordingly, apart from a few basic features that define a TWSS joke (e. [sent-44, score-0.09]

31 , short sentence), all of our approach’s lexical features model a metaphorical mapping to objects and structures in the erotic domain. [sent-46, score-0.328]

32 Part of TWSS identification is recognizing that the source context in which the potential TWSS is uttered is not in an erotic one. [sent-47, score-0.304]

33 If it is, then the mapping to the erotic domain is the identity and the state- ment is not a TWSS. [sent-48, score-0.307]

34 In this paper, we assume all test instances are from nonerotic domains and leave the 90 classification of erotic and nonerotic contexts to future work. [sent-49, score-0.481]

35 First, many domains in which a TWSS classifier could be applied value high precision significantly more than high recall. [sent-51, score-0.06]

36 For example, in a social setting, the cost of saying “that’s what she said” inappropriately is high, whereas the cost of not saying it when it might have been appropriate is negligible. [sent-52, score-0.09]

37 For another example, in automated public tagging of twitter and facebook data, false positives are considered spam and violate usage policies, whereas false negatives go unnoticed. [sent-53, score-0.175]

38 Second, the overwhelming majority of everyday sentences are not TWSSs, making achieving high precision even more difficult. [sent-54, score-0.06]

39 In this paper, we strive specifically to achieve high precision but are willing to sacrifice recall. [sent-55, score-0.037]

40 3 The DEviaNT Approach The TWSS problem has two identifying characteristics: (1) TWSSs are likely to contain nouns that are euphemisms for sexually explicit nouns and (2) TWSSs share common structure with sentences in the erotic domain. [sent-56, score-0.883]

41 Our approach to solving the TWSS problem is centered around an SVM model that uses features designed to model those characteristics. [sent-57, score-0.051]

42 We will use features that build on corpus statistics computed for known erotic words, and their lexical contexts, as described in the rest of this section. [sent-59, score-0.282]

43 1 Data and word classes Let SN be an open set of sexually explicit nouns. [sent-61, score-0.166]

44 We manually approximated SN with a set of 76 nouns that are predominantly used in sexual contexts. [sent-62, score-0.187]

45 We clustered the nouns into 9 categories based on which sexual object, body part, or participant they identify. [sent-63, score-0.166]

46 Let SN− ⊂ SN be the set of sexually explicit nouns that are likely targets sfeotr euphemism. [sent-64, score-0.302]

47 p Wliec td nido not consider euphemisms for people since they rarely, if e? [sent-65, score-0.147]

48 5M sentences from the erotica section of text files . [sent-81, score-0.121]

49 We tagged the erotica corpus with the Stanford Parser (Toutanova and Manning, 2000; Toutanova et al. [sent-85, score-0.098]

50 To make the corpora more generic, we replaced all numbers with the CD tag, all proper nouns with the NNP tag, all nouns ∈ SN with an SN tag, and all nouns ∈ B,P al lw nitohu tnhse ∈NN SN tag. [sent-87, score-0.351]

51 2 Word- and phrase-level analysis We define three functions to measure how closely related a noun, an adjective, and a verb phrase are to the erotica domain. [sent-90, score-0.133]

52 The noun sexiness function NS(n) is a realvalued measure of the maximum similarity a noun n ∈/ SN has to each of the nouns ∈ SN−. [sent-92, score-0.365]

53 For each noun, Nlet h tahse t adjective count vector b SeN the vector of the absolute frequencies of each adjective that modifies the noun in the union of the erotica and the Brown corpora. [sent-93, score-0.266]

54 We define NS(n) to be the maximum cosine similarity, over each noun ∈ SN−, using tmerumm frequency-inverse dvoecru emacehnt n frequency (tf-idf) weights of the nouns’ adjective count vectors. [sent-94, score-0.123]

55 For nouns that occurred fewer that 200 times, occurred fewer than 50 times with adjectives, or were associated with 3 times as many adjectives that never occurred with nouns in SN than adjectives that did, NS(n) = 10−7 (smaller than all recorded similarities). [sent-95, score-0.382]

56 Example nouns with high NS are “rod” and “meat” . [sent-96, score-0.117]

57 The adjective sexiness function AS(a) is a real-valued measure of how likely an adjective a is to modify a noun ∈ SN. [sent-98, score-0.252]

58 i nW see ndteefinncees A Sin( t)h teo e breoti tchae corpus that contain at least one noun ∈ SN. [sent-100, score-0.093]

59 The verb sexiness function VS(v) is a realvalued measure of how much more likely a verb phrase v is to appear in an erotic context than a nonerotic one. [sent-104, score-0.541]

60 Let SE be the set of sentences in the erotica corpus that contain nouns ∈ SN. [sent-105, score-0.253]

61 G Siven a sentence s containing a verb v, the verb phrase v is the contiguous substring of the sentence that con91 tains v and is bordered on each side by the closest noun or one of the set of pronouns {I, you, it, me}. [sent-107, score-0.167]

62 (If unnei othre orn a noun nor none orofn ntoheu pronouns occur on a side of the verb, v itself is an endpoint of v. [sent-108, score-0.097]

63 ) To define VS(v), we approximate the probabilities of v appearing in an erotic and a nonerotic context with counts in SE and SB, respectively. [sent-109, score-0.36]

64 Intuitively, the verb sexiness is a measure of how likely the action described in a sentence could be an action (via some metaphoric mapping) to an action in an erotic context. [sent-113, score-0.47]

65 3 Features DEviaNT uses the following features to identify potential mappings of a sentence s into the erotic domain, organized into two categories: NOUN EUPHEMISMS and STRUCTURAL ELEMENTS. [sent-115, score-0.282]

66 NOUN EUPHEMISMS: • (boolean) does s contain a noun ∈ SN? [sent-116, score-0.093]

67 , • (boolean) ddooeess s ccoonnttaaiinn a noun ∈ BP? [sent-117, score-0.111]

68 , • (boolean) ddooeess s ocontnatianin a a noun n s? [sent-118, score-0.111]

69 u,ch that NS(n) = )1 d0o−e7s, • (real) average NS(n), for all nouns n ∈ s such t(hreaat n ∈/ SerNa g∪e BP, STRUCTURAL ELEMENTS: • (boolean) does s contain a verb that never occurs ilena SE? [sent-119, score-0.206]

70 , • (boolean) does s contain a verb phrase that never occurs eins SE? [sent-120, score-0.089]

71 , • (real) average VS(v) over all verb phrases v ∈ s, • (real) average AS(a) over aallll adjectives a ∈ s, • (boolean) adgoees A s ca)on otvaeinr an adjective a ∈su sc,h t(hboato a never occurs innta a s aennte andcjee s ∈ SE s∪u SB twhaitht a noun ∈ cScNu. [sent-121, score-0.228]

72 (We approximate rth per subject )w tihteh t shueb fjiercstt noun or pronoun. [sent-125, score-0.095]

73 To minimize false positives, while tolerating false negatives, DEviaNT employs the MetaCost metaclassifier (Domingos, 1999), which uses bagging to reclassify the training data to produce a single cost-sensitive classifier. [sent-131, score-0.139]

74 DEviaNT sets the cost of a false positive to be 100 times that of a false negative. [sent-132, score-0.106]

75 DEviaNT explores a particular approach to solving the TWSS problem: recognizing euphemistic and structural relationships between the source domain and an erotic domain. [sent-134, score-0.338]

76 Thus, the goal of our evaluation is not to outperform the baselines in all aspects, but rather to show that by using only euphemism-based and structure-based features, DEviaNT can compete with the baselines, particularly where it matters most, delivering high precision and few false positives. [sent-138, score-0.111]

77 DE92 viaNT’s positive training data are 2001 quoted sentences from tws s stories . [sent-141, score-0.041]

78 DEviaNT’s negative training data are 2001 sentences from three sources (667 each): text s fromlastnight . [sent-143, score-0.04]

79 com/ int imacy (FML), a set of short (1– 2 sentence) user-submitted stories about their love lives; and wikiquote . [sent-145, score-0.044]

80 For testing, we used 262 other TS and 20,700 other TFLN, FML, and WQ sentences (all the data from these sources that were available at the time ofthe experiments). [sent-148, score-0.04]

81 , changing the tag of “i” from the foreign word tag FW to the correct pronoun tag PRP). [sent-152, score-0.093]

82 2 Baselines Our experiments compare DEviaNT to seven other classifiers: (1) a Na¨ ıve Bayes classifier on unigram features, (2) an SVM model trained on unigram features, (3) an SVM model trained on unigram and bigram features, (4–6) MetaCost (Domingos, 1999) (see Section 3. [sent-154, score-0.144]

83 The state-of-the-practice approach to TWSS identification is a na¨ ıve Bayes model trained on a unigram model of instances of twitter tweets, some tagged with #twss (VandenBos, 2011). [sent-157, score-0.106]

84 For completeness, we tested whether adding unigram features to DEviaNT improved its performance but found that it did not. [sent-161, score-0.068]

85 Recall Figure 1: The precision-recall curves for DEviaNT and baseline classifiers on TS, TFLN, FML, and WQ. [sent-162, score-0.044]

86 The best competitor Unigram SVM w/o MetaCost has the maximum precision of 59. [sent-166, score-0.037]

87 Note that the addition of bigram features yields no improvement in (and can hurt) both precision and recall. [sent-170, score-0.057]

88 To qualitatively evaluate DEviaNT, we compared those sentences that DEviaNT, Basic Structure, and Unigram SVM w/o MetaCost are most sure are TWSSs. [sent-171, score-0.046]

89 DEviaNT returned 28 such sentences (all tied for most likely to be a TWSS), 20 of which are true positives. [sent-172, score-0.042]

90 However, 2 of the 8 false positives are in fact TWSSs (despite coming from the negative testing data): “Yes give me all the cream and he’s gone. [sent-173, score-0.099]

91 However, DEviaNT was also able to identify TWSSs that deal with noun euphemisms (e. [sent-177, score-0.225]

92 Note that while DE93 viaNT has a much lower recall than Unigram SVM w/o MetaCost, it accomplishes our goal of delivering high-precision, while tolerating low recall. [sent-182, score-0.054]

93 Note that the DEviaNT’s precision appears low in large because the testing data is predominantly negative. [sent-183, score-0.058]

94 5 Contributions We formally defined the TWSS problem, a subproblem of the double entendre problem. [sent-186, score-0.245]

95 We then identified two characteristics of the TWSS problem (1) TWSSs are likely to contain nouns that are euphemisms for sexually explicit nouns and (2) TWSSs share common structure with sentences in the erotic domain that we used to construct DEviaNT, an approach for TWSS classification. [sent-187, score-0.942]

96 — — DEviaNT identifies euphemism and erotic-domain structure without relying heavily on structural features specific to TWSSs. [sent-188, score-0.056]

97 DEviaNT delivers significantly higher precision than classifiers that use n-gram TWSS models. [sent-189, score-0.059]

98 Our experiments indicate that euphemism- and erotic-domain-structure features contribute to improving the precision of TWSS identification. [sent-190, score-0.057]

99 While significant future work in improving DEviaNT remains, we have identified two characteristics important to the TWSS problem and demonstrated that an approach based on these characteristics has promise. [sent-191, score-0.066]

100 The technique of metaphorical mapping may be generalized to identify other types of double entendres and other forms of humor. [sent-192, score-0.18]


similar papers computed by tfidf model

tfidf for this paper:

wordName wordTfidf (topN-words)

[('twss', 0.589), ('deviant', 0.531), ('erotic', 0.262), ('twsss', 0.196), ('euphemisms', 0.147), ('sexually', 0.131), ('nouns', 0.117), ('entendre', 0.115), ('metacost', 0.115), ('sn', 0.107), ('double', 0.101), ('metaphor', 0.1), ('erotica', 0.098), ('nonerotic', 0.098), ('noun', 0.078), ('said', 0.067), ('sexiness', 0.065), ('false', 0.053), ('boolean', 0.05), ('se', 0.049), ('fml', 0.049), ('joke', 0.049), ('sexual', 0.049), ('tfln', 0.049), ('viant', 0.049), ('svm', 0.048), ('unigram', 0.048), ('saying', 0.045), ('adjective', 0.045), ('toutanova', 0.044), ('ns', 0.044), ('sb', 0.044), ('humor', 0.043), ('nonliteral', 0.043), ('identification', 0.042), ('vs', 0.039), ('precision', 0.037), ('verb', 0.035), ('explicit', 0.035), ('characteristics', 0.033), ('daniels', 0.033), ('ddooeess', 0.033), ('entendres', 0.033), ('innocuous', 0.033), ('jokes', 0.033), ('pasanek', 0.033), ('tolerating', 0.033), ('literal', 0.032), ('adjectives', 0.031), ('solving', 0.031), ('mihalcea', 0.03), ('bp', 0.03), ('positives', 0.03), ('metaphoric', 0.029), ('funny', 0.029), ('wq', 0.029), ('subproblem', 0.029), ('metaphorical', 0.027), ('rigorously', 0.027), ('realvalued', 0.027), ('int', 0.026), ('domain', 0.026), ('tag', 0.024), ('sure', 0.023), ('negatives', 0.023), ('television', 0.023), ('meat', 0.023), ('domains', 0.023), ('sentences', 0.023), ('occurred', 0.022), ('curves', 0.022), ('classifiers', 0.022), ('delivering', 0.021), ('predominantly', 0.021), ('pronoun', 0.021), ('basic', 0.021), ('never', 0.02), ('ts', 0.02), ('features', 0.02), ('action', 0.02), ('mapping', 0.019), ('francis', 0.019), ('domingos', 0.019), ('likely', 0.019), ('pronouns', 0.019), ('occurs', 0.019), ('weka', 0.019), ('structural', 0.019), ('stories', 0.018), ('brown', 0.018), ('let', 0.018), ('selectional', 0.018), ('rada', 0.018), ('structure', 0.017), ('sources', 0.017), ('subject', 0.017), ('bayes', 0.017), ('twitter', 0.016), ('coming', 0.016), ('contain', 0.015)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.99999964 297 acl-2011-That's What She Said: Double Entendre Identification

Author: Chloe Kiddon ; Yuriy Brun

Abstract: Humor identification is a hard natural language understanding problem. We identify a subproblem — the “that’s what she said” problem with two distinguishing characteristics: (1) use of nouns that are euphemisms for sexually explicit nouns and (2) structure common in the erotic domain. We address this problem in a classification approach that includes features that model those two characteristics. Experiments on web data demonstrate that our approach improves precision by 12% over baseline techniques that use only word-based features. —

2 0.049067657 159 acl-2011-Identifying Noun Product Features that Imply Opinions

Author: Lei Zhang ; Bing Liu

Abstract: Identifying domain-dependent opinion words is a key problem in opinion mining and has been studied by several researchers. However, existing work has been focused on adjectives and to some extent verbs. Limited work has been done on nouns and noun phrases. In our work, we used the feature-based opinion mining model, and we found that in some domains nouns and noun phrases that indicate product features may also imply opinions. In many such cases, these nouns are not subjective but objective. Their involved sentences are also objective sentences and imply positive or negative opinions. Identifying such nouns and noun phrases and their polarities is very challenging but critical for effective opinion mining in these domains. To the best of our knowledge, this problem has not been studied in the literature. This paper proposes a method to deal with the problem. Experimental results based on real-life datasets show promising results. 1

3 0.037923329 242 acl-2011-Part-of-Speech Tagging for Twitter: Annotation, Features, and Experiments

Author: Kevin Gimpel ; Nathan Schneider ; Brendan O'Connor ; Dipanjan Das ; Daniel Mills ; Jacob Eisenstein ; Michael Heilman ; Dani Yogatama ; Jeffrey Flanigan ; Noah A. Smith

Abstract: We address the problem of part-of-speech tagging for English data from the popular microblogging service Twitter. We develop a tagset, annotate data, develop features, and report tagging results nearing 90% accuracy. The data and tools have been made available to the research community with the goal of enabling richer text analysis of Twitter and related social media data sets.

4 0.036476713 89 acl-2011-Creative Language Retrieval: A Robust Hybrid of Information Retrieval and Linguistic Creativity

Author: Tony Veale

Abstract: Information retrieval (IR) and figurative language processing (FLP) could scarcely be more different in their treatment of language and meaning. IR views language as an open-ended set of mostly stable signs with which texts can be indexed and retrieved, focusing more on a text’s potential relevance than its potential meaning. In contrast, FLP views language as a system of unstable signs that can be used to talk about the world in creative new ways. There is another key difference: IR is practical, scalable and robust, and in daily use by millions of casual users. FLP is neither scalable nor robust, and not yet practical enough to migrate beyond the lab. This paper thus presents a mutually beneficial hybrid of IR and FLP, one that enriches IR with new operators to enable the non-literal retrieval of creative expressions, and which also transplants FLP into a robust, scalable framework in which practical applications of linguistic creativity can be implemented. 1

5 0.03615978 44 acl-2011-An exponential translation model for target language morphology

Author: Michael Subotin

Abstract: This paper presents an exponential model for translation into highly inflected languages which can be scaled to very large datasets. As in other recent proposals, it predicts targetside phrases and can be conditioned on sourceside context. However, crucially for the task of modeling morphological generalizations, it estimates feature parameters from the entire training set rather than as a collection of separate classifiers. We apply it to English-Czech translation, using a variety of features capturing potential predictors for case, number, and gender, and one of the largest publicly available parallel data sets. We also describe generation and modeling of inflected forms unobserved in training data and decoding procedures for a model with non-local target-side feature dependencies.

6 0.035255078 111 acl-2011-Effects of Noun Phrase Bracketing in Dependency Parsing and Machine Translation

7 0.032812625 229 acl-2011-NULEX: An Open-License Broad Coverage Lexicon

8 0.031491887 198 acl-2011-Latent Semantic Word Sense Induction and Disambiguation

9 0.030367522 127 acl-2011-Exploiting Web-Derived Selectional Preference to Improve Statistical Dependency Parsing

10 0.028739642 208 acl-2011-Lexical Normalisation of Short Text Messages: Makn Sens a #twitter

11 0.028049747 333 acl-2011-Web-Scale Features for Full-Scale Parsing

12 0.02804617 7 acl-2011-A Corpus for Modeling Morpho-Syntactic Agreement in Arabic: Gender, Number and Rationality

13 0.027663736 292 acl-2011-Target-dependent Twitter Sentiment Classification

14 0.027195446 119 acl-2011-Evaluating the Impact of Coder Errors on Active Learning

15 0.026958965 323 acl-2011-Unsupervised Part-of-Speech Tagging with Bilingual Graph-Based Projections

16 0.026684856 157 acl-2011-I Thou Thee, Thou Traitor: Predicting Formal vs. Informal Address in English Literature

17 0.026624793 332 acl-2011-Using Multiple Sources to Construct a Sentiment Sensitive Thesaurus for Cross-Domain Sentiment Classification

18 0.026037637 167 acl-2011-Improving Dependency Parsing with Semantic Classes

19 0.025931373 331 acl-2011-Using Large Monolingual and Bilingual Corpora to Improve Coordination Disambiguation

20 0.025889028 259 acl-2011-Rare Word Translation Extraction from Aligned Comparable Documents


similar papers computed by lsi model

lsi for this paper:

topicId topicWeight

[(0, 0.081), (1, 0.023), (2, -0.011), (3, -0.016), (4, -0.014), (5, 0.002), (6, 0.032), (7, -0.004), (8, -0.005), (9, -0.006), (10, -0.027), (11, -0.02), (12, -0.002), (13, 0.0), (14, -0.004), (15, -0.009), (16, -0.013), (17, 0.01), (18, 0.01), (19, -0.014), (20, 0.004), (21, 0.004), (22, -0.01), (23, 0.009), (24, -0.007), (25, 0.029), (26, 0.047), (27, -0.016), (28, 0.029), (29, -0.006), (30, 0.07), (31, -0.005), (32, -0.016), (33, -0.02), (34, -0.022), (35, 0.03), (36, -0.038), (37, 0.034), (38, -0.007), (39, 0.002), (40, -0.045), (41, 0.034), (42, -0.009), (43, -0.01), (44, 0.035), (45, -0.063), (46, -0.026), (47, 0.043), (48, -0.031), (49, -0.052)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.85514873 297 acl-2011-That's What She Said: Double Entendre Identification

Author: Chloe Kiddon ; Yuriy Brun

Abstract: Humor identification is a hard natural language understanding problem. We identify a subproblem — the “that’s what she said” problem with two distinguishing characteristics: (1) use of nouns that are euphemisms for sexually explicit nouns and (2) structure common in the erotic domain. We address this problem in a classification approach that includes features that model those two characteristics. Experiments on web data demonstrate that our approach improves precision by 12% over baseline techniques that use only word-based features. —

2 0.55857831 136 acl-2011-Finding Deceptive Opinion Spam by Any Stretch of the Imagination

Author: Myle Ott ; Yejin Choi ; Claire Cardie ; Jeffrey T. Hancock

Abstract: Consumers increasingly rate, review and research products online (Jansen, 2010; Litvin et al., 2008). Consequently, websites containing consumer reviews are becoming targets of opinion spam. While recent work has focused primarily on manually identifiable instances of opinion spam, in this work we study deceptive opinion spam—fictitious opinions that have been deliberately written to sound authentic. Integrating work from psychology and computational linguistics, we develop and compare three approaches to detecting deceptive opinion spam, and ultimately develop a classifier that is nearly 90% accurate on our gold-standard opinion spam dataset. Based on feature analysis of our learned models, we additionally make several theoretical contributions, including revealing a relationship between deceptive opinions and imaginative writing.

3 0.54271114 159 acl-2011-Identifying Noun Product Features that Imply Opinions

Author: Lei Zhang ; Bing Liu

Abstract: Identifying domain-dependent opinion words is a key problem in opinion mining and has been studied by several researchers. However, existing work has been focused on adjectives and to some extent verbs. Limited work has been done on nouns and noun phrases. In our work, we used the feature-based opinion mining model, and we found that in some domains nouns and noun phrases that indicate product features may also imply opinions. In many such cases, these nouns are not subjective but objective. Their involved sentences are also objective sentences and imply positive or negative opinions. Identifying such nouns and noun phrases and their polarities is very challenging but critical for effective opinion mining in these domains. To the best of our knowledge, this problem has not been studied in the literature. This paper proposes a method to deal with the problem. Experimental results based on real-life datasets show promising results. 1

4 0.53145468 131 acl-2011-Extracting Opinion Expressions and Their Polarities - Exploration of Pipelines and Joint Models

Author: Richard Johansson ; Alessandro Moschitti

Abstract: We investigate systems that identify opinion expressions and assigns polarities to the extracted expressions. In particular, we demonstrate the benefit of integrating opinion extraction and polarity classification into a joint model using features reflecting the global polarity structure. The model is trained using large-margin structured prediction methods. The system is evaluated on the MPQA opinion corpus, where we compare it to the only previously published end-to-end system for opinion expression extraction and polarity classification. The results show an improvement of between 10 and 15 absolute points in F-measure.

5 0.52077919 130 acl-2011-Extracting Comparative Entities and Predicates from Texts Using Comparative Type Classification

Author: Seon Yang ; Youngjoong Ko

Abstract: The automatic extraction of comparative information is an important text mining problem and an area of increasing interest. In this paper, we study how to build a Korean comparison mining system. Our work is composed of two consecutive tasks: 1) classifying comparative sentences into different types and 2) mining comparative entities and predicates. We perform various experiments to find relevant features and learning techniques. As a result, we achieve outstanding performance enough for practical use. 1

6 0.51802182 323 acl-2011-Unsupervised Part-of-Speech Tagging with Bilingual Graph-Based Projections

7 0.50594038 84 acl-2011-Contrasting Opposing Views of News Articles on Contentious Issues

8 0.5015437 229 acl-2011-NULEX: An Open-License Broad Coverage Lexicon

9 0.49962845 238 acl-2011-P11-2093 k2opt.pdf

10 0.49233598 211 acl-2011-Liars and Saviors in a Sentiment Annotated Corpus of Comments to Political Debates

11 0.48590118 74 acl-2011-Combining Indicators of Allophony

12 0.4782804 306 acl-2011-Towards Style Transformation from Written-Style to Audio-Style

13 0.47673035 109 acl-2011-Effective Measures of Domain Similarity for Parsing

14 0.47281143 179 acl-2011-Is Machine Translation Ripe for Cross-Lingual Sentiment Classification?

15 0.47085795 102 acl-2011-Does Size Matter - How Much Data is Required to Train a REG Algorithm?

16 0.47067371 311 acl-2011-Translationese and Its Dialects

17 0.47058621 331 acl-2011-Using Large Monolingual and Bilingual Corpora to Improve Coordination Disambiguation

18 0.46749547 120 acl-2011-Even the Abstract have Color: Consensus in Word-Colour Associations

19 0.45998234 88 acl-2011-Creating a manually error-tagged and shallow-parsed learner corpus

20 0.45658508 222 acl-2011-Model-Portability Experiments for Textual Temporal Analysis


similar papers computed by lda model

lda for this paper:

topicId topicWeight

[(0, 0.012), (5, 0.026), (17, 0.033), (26, 0.044), (31, 0.026), (37, 0.082), (39, 0.035), (41, 0.053), (46, 0.341), (55, 0.019), (59, 0.025), (72, 0.042), (91, 0.028), (96, 0.088), (97, 0.019), (98, 0.016)]

similar papers list:

simIndex simValue paperId paperTitle

same-paper 1 0.71940529 297 acl-2011-That's What She Said: Double Entendre Identification

Author: Chloe Kiddon ; Yuriy Brun

Abstract: Humor identification is a hard natural language understanding problem. We identify a subproblem — the “that’s what she said” problem with two distinguishing characteristics: (1) use of nouns that are euphemisms for sexually explicit nouns and (2) structure common in the erotic domain. We address this problem in a classification approach that includes features that model those two characteristics. Experiments on web data demonstrate that our approach improves precision by 12% over baseline techniques that use only word-based features. —

2 0.61591375 209 acl-2011-Lexically-Triggered Hidden Markov Models for Clinical Document Coding

Author: Svetlana Kiritchenko ; Colin Cherry

Abstract: The automatic coding of clinical documents is an important task for today’s healthcare providers. Though it can be viewed as multi-label document classification, the coding problem has the interesting property that most code assignments can be supported by a single phrase found in the input document. We propose a Lexically-Triggered Hidden Markov Model (LT-HMM) that leverages these phrases to improve coding accuracy. The LT-HMM works in two stages: first, a lexical match is performed against a term dictionary to collect a set of candidate codes for a document. Next, a discriminative HMM selects the best subset of codes to assign to the document by tagging candidates as present or absent. By confirming codes proposed by a dictionary, the LT-HMM can share features across codes, enabling strong performance even on rare codes. In fact, we are able to recover codes that do not occur in the training set at all. Our approach achieves the best ever performance on the 2007 Medical NLP Challenge test set, with an F-measure of 89.84.

3 0.5534386 238 acl-2011-P11-2093 k2opt.pdf

Author: empty-author

Abstract: We present a pointwise approach to Japanese morphological analysis (MA) that ignores structure information during learning and tagging. Despite the lack of structure, it is able to outperform the current state-of-the-art structured approach for Japanese MA, and achieves accuracy similar to that of structured predictors using the same feature set. We also find that the method is both robust to outof-domain data, and can be easily adapted through the use of a combination of partial annotation and active learning.

4 0.40083846 246 acl-2011-Piggyback: Using Search Engines for Robust Cross-Domain Named Entity Recognition

Author: Stefan Rud ; Massimiliano Ciaramita ; Jens Muller ; Hinrich Schutze

Abstract: We use search engine results to address a particularly difficult cross-domain language processing task, the adaptation of named entity recognition (NER) from news text to web queries. The key novelty of the method is that we submit a token with context to a search engine and use similar contexts in the search results as additional information for correctly classifying the token. We achieve strong gains in NER performance on news, in-domain and out-of-domain, and on web queries.

5 0.39907539 331 acl-2011-Using Large Monolingual and Bilingual Corpora to Improve Coordination Disambiguation

Author: Shane Bergsma ; David Yarowsky ; Kenneth Church

Abstract: Resolving coordination ambiguity is a classic hard problem. This paper looks at coordination disambiguation in complex noun phrases (NPs). Parsers trained on the Penn Treebank are reporting impressive numbers these days, but they don’t do very well on this problem (79%). We explore systems trained using three types of corpora: (1) annotated (e.g. the Penn Treebank), (2) bitexts (e.g. Europarl), and (3) unannotated monolingual (e.g. Google N-grams). Size matters: (1) is a million words, (2) is potentially billions of words and (3) is potentially trillions of words. The unannotated monolingual data is helpful when the ambiguity can be resolved through associations among the lexical items. The bilingual data is helpful when the ambiguity can be resolved by the order of words in the translation. We train separate classifiers with monolingual and bilingual features and iteratively improve them via achieves data and pervised tations. co-training. The co-trained classifier close to 96% accuracy on Treebank makes 20% fewer errors than a susystem trained with Treebank anno-

6 0.39889795 311 acl-2011-Translationese and Its Dialects

7 0.39889395 292 acl-2011-Target-dependent Twitter Sentiment Classification

8 0.39774954 32 acl-2011-Algorithm Selection and Model Adaptation for ESL Correction Tasks

9 0.39755732 126 acl-2011-Exploiting Syntactico-Semantic Structures for Relation Extraction

10 0.3960093 333 acl-2011-Web-Scale Features for Full-Scale Parsing

11 0.39539626 128 acl-2011-Exploring Entity Relations for Named Entity Disambiguation

12 0.39537317 34 acl-2011-An Algorithm for Unsupervised Transliteration Mining with an Application to Word Alignment

13 0.394685 48 acl-2011-Automatic Detection and Correction of Errors in Dependency Treebanks

14 0.39467198 119 acl-2011-Evaluating the Impact of Coder Errors on Active Learning

15 0.39368916 65 acl-2011-Can Document Selection Help Semi-supervised Learning? A Case Study On Event Extraction

16 0.39322233 36 acl-2011-An Efficient Indexer for Large N-Gram Corpora

17 0.39239916 5 acl-2011-A Comparison of Loopy Belief Propagation and Dual Decomposition for Integrated CCG Supertagging and Parsing

18 0.39216256 123 acl-2011-Exact Decoding of Syntactic Translation Models through Lagrangian Relaxation

19 0.39180237 327 acl-2011-Using Bilingual Parallel Corpora for Cross-Lingual Textual Entailment

20 0.39126369 133 acl-2011-Extracting Social Power Relationships from Natural Language