emnlp emnlp2012 knowledge-graph by maker-knowledge-mining

emnlp 2012 knowledge graph


similar papers computed by tfidf model


similar papers computed by lsi model


similar papers computed by lda model


papers list:

1 emnlp-2012-A Bayesian Model for Learning SCFGs with Discontiguous Rules

Author: Abby Levenberg ; Chris Dyer ; Phil Blunsom

Abstract: We describe a nonparametric model and corresponding inference algorithm for learning Synchronous Context Free Grammar derivations for parallel text. The model employs a Pitman-Yor Process prior which uses a novel base distribution over synchronous grammar rules. Through both synthetic grammar induction and statistical machine translation experiments, we show that our model learns complex translational correspondences— including discontiguous, many-to-many alignments—and produces competitive translation results. Further, inference is efficient and we present results on significantly larger corpora than prior work.

2 emnlp-2012-A Beam-Search Decoder for Grammatical Error Correction

Author: Daniel Dahlmeier ; Hwee Tou Ng

Abstract: We present a novel beam-search decoder for grammatical error correction. The decoder iteratively generates new hypothesis corrections from current hypotheses and scores them based on features of grammatical correctness and fluency. These features include scores from discriminative classifiers for specific error categories, such as articles and prepositions. Unlike all previous approaches, our method is able to perform correction of whole sentences with multiple and interacting errors while still taking advantage of powerful existing classifier approaches. Our decoder achieves an F1 correction score significantly higher than all previous published scores on the Helping Our Own (HOO) shared task data set.

3 emnlp-2012-A Coherence Model Based on Syntactic Patterns

Author: Annie Louis ; Ani Nenkova

Abstract: We introduce a model of coherence which captures the intentional discourse structure in text. Our work is based on the hypothesis that syntax provides a proxy for the communicative goal of a sentence and therefore the sequence of sentences in a coherent discourse should exhibit detectable structural patterns. Results show that our method has high discriminating power for separating out coherent and incoherent news articles reaching accuracies of up to 90%. We also show that our syntactic patterns are correlated with manual annotations of intentional structure for academic conference articles and can successfully predict the coherence of abstract, introduction and related work sections of these articles. 59.3 (100.0) Intro 50.3 (100.0) 1166 Rel wk 55.4 (100.0) >= 0.663.8 (67.2)50.8 (71.1)58.6 (75.9) >= 0.7 67.2 (32.0) 54.4 (38.6) 63.3 (52.8) >= 0.8 74.0 (10.0) 51.6 (22.0) 63.0 (25.7) >= 0.9 91.7 (2.0) 30.6 (5.0) 68.1 (7.2) Table 9: Accuracy (% examples) above each confidence level for the conference versus workshop task. These results are shown in Table 9. The proportion of examples under each setting is also indicated. When only examples above 0.6 confidence are examined, the classifier has a higher accuracy of63.8% for abstracts and covers close to 70% of the examples. Similarly, when a cutoff of 0.7 is applied to the confidence for predicting related work sections, we achieve 63.3% accuracy for 53% of examples. So we can consider that 30 to 47% of the examples in the two sections respectively are harder to tell apart. Interestingly however even high confidence predictions on introductions remain incorrect. These results show that our model can successfully distinguish the structure of articles beyond just clearly incoherent permutation examples. 7 Conclusion Our work is the first to develop an unsupervised model for intentional structure and to show that it has good accuracy for coherence prediction and also complements entity and lexical structure of discourse. This result raises interesting questions about how patterns captured by these different coherence metrics vary and how they can be combined usefully for predicting coherence. We plan to explore these ideas in future work. We also want to analyze genre differences to understand if the strength of these coherence dimensions varies with genre. Acknowledgements This work is partially supported by a Google research grant and NSF CAREER 0953445 award. References Regina Barzilay and Mirella Lapata. 2008. Modeling local coherence: An entity-based approach. Computa- tional Linguistics, 34(1): 1–34. Regina Barzilay and Lillian Lee. 2004. Catching the drift: Probabilistic content models, with applications to generation and summarization. In Proceedings of NAACL-HLT, pages 113–120. Xavier Carreras, Michael Collins, and Terry Koo. 2008. Tag, dynamic programming, and the perceptron for efficient, feature-rich parsing. In Proceedings of CoNLL, pages 9–16. Eugene Charniak and Mark Johnson. 2005. Coarse-tofine n-best parsing and maxent discriminative reranking. In Proceedings of ACL, pages 173–180. Jackie C.K. Cheung and Gerald Penn. 2010. Utilizing extra-sentential context for parsing. In Proceedings of EMNLP, pages 23–33. Christelle Cocco, Rapha ¨el Pittier, Fran ¸cois Bavaud, and Aris Xanthos. 2011. Segmentation and clustering of textual sequences: a typological approach. In Proceedings of RANLP, pages 427–433. Michael Collins and Terry Koo. 2005. Discriminative reranking for natural language parsing. Computational Linguistics, 3 1:25–70. Isaac G. Councill, C. Lee Giles, and Min-Yen Kan. 2008. Parscit: An open-source crf reference string parsing package. In Proceedings of LREC, pages 661–667. Micha Elsner and Eugene Charniak. 2008. Coreferenceinspired coherence modeling. In Proceedings of ACLHLT, Short Papers, pages 41–44. Micha Elsner and Eugene Charniak. 2011. Extending the entity grid with entity-specific features. In Proceedings of ACL-HLT, pages 125–129. Micha Elsner, Joseph Austerweil, and Eugene Charniak. 2007. A unified local and global model for discourse coherence. In Proceedings of NAACL-HLT, pages 436–443. Pascale Fung and Grace Ngai. 2006. One story, one flow: Hidden markov story models for multilingual multidocument summarization. ACM Transactions on Speech and Language Processing, 3(2): 1–16. Barbara J. Grosz and Candace L. Sidner. 1986. Attention, intentions, and the structure of discourse. Computational Linguistics, 3(12): 175–204. Yufan Guo, Anna Korhonen, and Thierry Poibeau. 2011. A weakly-supervised approach to argumentative zoning of scientific documents. In Proceedings of EMNLP, pages 273–283. Liang Huang. 2008. Forest reranking: Discriminative parsing with non-local features. In Proceedings of ACL-HLT, pages 586–594, June. 1167 Nikiforos Karamanis, Chris Mellish, Massimo Poesio, and Jon Oberlander. 2009. Evaluating centering for information ordering using corpora. Computational Linguistics, 35(1):29–46. Dan Klein and Christopher D. Manning. 2003. Accurate unlexicalized parsing. In Proceedings of ACL, pages 423–430. Mirella Lapata and Regina Barzilay. 2005. Automatic evaluation of text coherence: Models and representations. In Proceedings of IJCAI. Mirella Lapata. 2003. Probabilistic text structuring: Experiments with sentence ordering. In Proceedings of ACL, pages 545–552. Maria Liakata and Larisa Soldatova. 2008. Guidelines for the annotation of general scientific concepts. JISC Project Report. Maria Liakata, Simone Teufel, Advaith Siddharthan, and Colin Batchelor. 2010. Corpora for the conceptualisation and zoning of scientific papers. In Proceedings of LREC. Ziheng Lin, Min-Yen Kan, and Hwee Tou Ng. 2009. Recognizing implicit discourse relations in the Penn Discourse Treebank. In Proceedings of EMNLP, pages 343–351. Ziheng Lin, Hwee Tou Ng, and Min-Yen Kan. 2011. Automatically evaluating text coherence using discourse relations. In Proceedings of ACL-HLT, pages 997– 1006. Mitchell P. Marcus, Beatrice Santorini, and Mary Ann Marcinkiewicz. 1994. Building a large annotated corpus of english: The penn treebank. Computational Linguistics, 19(2):313–330. Emily Pitler and Ani Nenkova. 2008. Revisiting readability: A unified framework for predicting text quality. In Proceedings of EMNLP, pages 186–195. Dragomir R. Radev, Mark Thomas Joseph, Bryan Gibson, and Pradeep Muthukrishnan. 2009. A Bibliometric and Network Analysis ofthe field of Computational Linguistics. Journal of the American Society for Information Science and Technology. David Reitter, Johanna D. Moore, and Frank Keller. 2006. Priming of Syntactic Rules in Task-Oriented Dialogue and Spontaneous Conversation. In Proceedings of the 28th Annual Conference of the Cognitive Science Society, pages 685–690. Jeffrey C. Reynar and Adwait Ratnaparkhi. 1997. A maximum entropy approach to identifying sentence boundaries. In Proceedings of the fifth conference on Applied natural language processing, pages 16–19. Radu Soricut and Daniel Marcu. 2006. Discourse generation using utility-trained coherence models. In Proceedings of COLING-ACL, pages 803–810. John Swales. 1990. Genre analysis: English in academic and research settings, volume 11. Cambridge University Press. Simone Teufel and Marc Moens. 2000. What’s yours and what’s mine: determining intellectual attribution in scientific text. In Proceedings of EMNLP, pages 9– 17. Simone Teufel, Jean Carletta, and Marc Moens. 1999. An annotation scheme for discourse-level argumentation in research articles. In Proceedings of EACL, pages 110–1 17. Ying Zhao, George Karypis, and Usama Fayyad. 2005. Hierarchical clustering algorithms for document datasets. Data Mining and Knowledge Discovery, 10: 141–168. 1168

4 emnlp-2012-A Comparison of Vector-based Representations for Semantic Composition

Author: William Blacoe ; Mirella Lapata

Abstract: In this paper we address the problem of modeling compositional meaning for phrases and sentences using distributional methods. We experiment with several possible combinations of representation and composition, exhibiting varying degrees of sophistication. Some are shallow while others operate over syntactic structure, rely on parameter learning, or require access to very large corpora. We find that shallow approaches are as good as more computationally intensive alternatives with regards to two particular tests: (1) phrase similarity and (2) paraphrase detection. The sizes of the involved training corpora and the generated vectors are not as important as the fit between the meaning representation and compositional method.

5 emnlp-2012-A Discriminative Model for Query Spelling Correction with Latent Structural SVM

Author: Huizhong Duan ; Yanen Li ; ChengXiang Zhai ; Dan Roth

Abstract: Discriminative training in query spelling correction is difficult due to the complex internal structures of the data. Recent work on query spelling correction suggests a two stage approach a noisy channel model that is used to retrieve a number of candidate corrections, followed by discriminatively trained ranker applied to these candidates. The ranker, however, suffers from the fact the low recall of the first, suboptimal, search stage. This paper proposes to directly optimize the search stage with a discriminative model based on latent structural SVM. In this model, we treat query spelling correction as a multiclass classification problem with structured input and output. The latent structural information is used to model the alignment of words in the spelling correction process. Experiment results show that as a standalone speller, our model outperforms all the baseline systems. It also attains a higher recall compared with the noisy channel model, and can therefore serve as a better filtering stage when combined with a ranker.

6 emnlp-2012-A New Minimally-Supervised Framework for Domain Word Sense Disambiguation

Author: Stefano Faralli ; Roberto Navigli

Abstract: We present a new minimally-supervised framework for performing domain-driven Word Sense Disambiguation (WSD). Glossaries for several domains are iteratively acquired from the Web by means of a bootstrapping technique. The acquired glosses are then used as the sense inventory for fullyunsupervised domain WSD. Our experiments, on new and gold-standard datasets, show that our wide-coverage framework enables highperformance results on dozens of domains at a coarse and fine-grained level.

7 emnlp-2012-A Novel Discriminative Framework for Sentence-Level Discourse Analysis

Author: Shafiq Joty ; Giuseppe Carenini ; Raymond Ng

Abstract: We propose a complete probabilistic discriminative framework for performing sentencelevel discourse analysis. Our framework comprises a discourse segmenter, based on a binary classifier, and a discourse parser, which applies an optimal CKY-like parsing algorithm to probabilities inferred from a Dynamic Conditional Random Field. We show on two corpora that our approach outperforms the state-of-the-art, often by a wide margin.

8 emnlp-2012-A Phrase-Discovering Topic Model Using Hierarchical Pitman-Yor Processes

Author: Robert Lindsey ; William Headden ; Michael Stipicevic

Abstract: Topic models traditionally rely on the bagof-words assumption. In data mining applications, this often results in end-users being presented with inscrutable lists of topical unigrams, single words inferred as representative of their topics. In this article, we present a hierarchical generative probabilistic model of topical phrases. The model simultaneously infers the location, length, and topic of phrases within a corpus and relaxes the bagof-words assumption within phrases by using a hierarchy of Pitman-Yor processes. We use Markov chain Monte Carlo techniques for approximate inference in the model and perform slice sampling to learn its hyperparameters. We show via an experiment on human subjects that our model finds substantially better, more interpretable topical phrases than do competing models.

9 emnlp-2012-A Sequence Labelling Approach to Quote Attribution

Author: Timothy O'Keefe ; Silvia Pareti ; James R. Curran ; Irena Koprinska ; Matthew Honnibal

Abstract: Quote extraction and attribution is the task of automatically extracting quotes from text and attributing each quote to its correct speaker. The present state-of-the-art system uses gold standard information from previous decisions in its features, which, when removed, results in a large drop in performance. We treat the problem as a sequence labelling task, which allows us to incorporate sequence features without using gold standard information. We present results on two new corpora and an augmented version of a third, achieving a new state-of-the-art for systems using only realistic features.

10 emnlp-2012-A Statistical Relational Learning Approach to Identifying Evidence Based Medicine Categories

Author: Mathias Verbeke ; Vincent Van Asch ; Roser Morante ; Paolo Frasconi ; Walter Daelemans ; Luc De Raedt

Abstract: Evidence-based medicine is an approach whereby clinical decisions are supported by the best available findings gained from scientific research. This requires efficient access to such evidence. To this end, abstracts in evidence-based medicine can be labeled using a set of predefined medical categories, the socalled PICO criteria. This paper presents an approach to automatically annotate sentences in medical abstracts with these labels. Since both structural and sequential information are important for this classification task, we use kLog, a new language for statistical relational learning with kernels. Our results show a clear improvement with respect to state-of-the-art systems.

11 emnlp-2012-A Systematic Comparison of Phrase Table Pruning Techniques

Author: Richard Zens ; Daisy Stanton ; Peng Xu

Abstract: When trained on very large parallel corpora, the phrase table component of a machine translation system grows to consume vast computational resources. In this paper, we introduce a novel pruning criterion that places phrase table pruning on a sound theoretical foundation. Systematic experiments on four language pairs under various data conditions show that our principled approach is superior to existing ad hoc pruning methods.

12 emnlp-2012-A Transition-Based System for Joint Part-of-Speech Tagging and Labeled Non-Projective Dependency Parsing

Author: Bernd Bohnet ; Joakim Nivre

Abstract: Most current dependency parsers presuppose that input words have been morphologically disambiguated using a part-of-speech tagger before parsing begins. We present a transitionbased system for joint part-of-speech tagging and labeled dependency parsing with nonprojective trees. Experimental evaluation on Chinese, Czech, English and German shows consistent improvements in both tagging and parsing accuracy when compared to a pipeline system, which lead to improved state-of-theart results for all languages.

13 emnlp-2012-A Unified Approach to Transliteration-based Text Input with Online Spelling Correction

Author: Hisami Suzuki ; Jianfeng Gao

Abstract: This paper presents an integrated, end-to-end approach to online spelling correction for text input. Online spelling correction refers to the spelling correction as you type, as opposed to post-editing. The online scenario is particularly important for languages that routinely use transliteration-based text input methods, such as Chinese and Japanese, because the desired target characters cannot be input at all unless they are in the list of candidates provided by an input method, and spelling errors prevent them from appearing in the list. For example, a user might type suesheng by mistake to mean xuesheng 学生 'student' in Chinese; existing input methods fail to convert this misspelled input to the desired target Chinese characters. In this paper, we propose a unified approach to the problem of spelling correction and transliteration-based character conversion using an approach inspired by the phrasebased statistical machine translation framework. At the phrase (substring) level, k most probable pinyin (Romanized Chinese) corrections are generated using a monotone decoder; at the sentence level, input pinyin strings are directly transliterated into target Chinese characters by a decoder using a loglinear model that refer to the features of both levels. A new method of automatically deriving parallel training data from user keystroke logs is also presented. Experiments on Chinese pinyin conversion show that our integrated method reduces the character error rate by 20% (from 8.9% to 7. 12%) over the previous state-of-the art based on a noisy channel model. 609 1

14 emnlp-2012-A Weakly Supervised Model for Sentence-Level Semantic Orientation Analysis with Multiple Experts

Author: Lizhen Qu ; Rainer Gemulla ; Gerhard Weikum

Abstract: We propose the weakly supervised MultiExperts Model (MEM) for analyzing the semantic orientation of opinions expressed in natural language reviews. In contrast to most prior work, MEM predicts both opinion polarity and opinion strength at the level of individual sentences; such fine-grained analysis helps to understand better why users like or dislike the entity under review. A key challenge in this setting is that it is hard to obtain sentence-level training data for both polarity and strength. For this reason, MEM is weakly supervised: It starts with potentially noisy indicators obtained from coarse-grained training data (i.e., document-level ratings), a small set of diverse base predictors, and, if available, small amounts of fine-grained training data. We integrate these noisy indicators into a unified probabilistic framework using ideas from ensemble learning and graph-based semi-supervised learning. Our experiments indicate that MEM outperforms state-of-the-art methods by a significant margin.

15 emnlp-2012-Active Learning for Imbalanced Sentiment Classification

Author: Shoushan Li ; Shengfeng Ju ; Guodong Zhou ; Xiaojun Li

Abstract: Active learning is a promising way for sentiment classification to reduce the annotation cost. In this paper, we focus on the imbalanced class distribution scenario for sentiment classification, wherein the number of positive samples is quite different from that of negative samples. This scenario posits new challenges to active learning. To address these challenges, we propose a novel active learning approach, named co-selecting, by taking both the imbalanced class distribution issue and uncertainty into account. Specifically, our co-selecting approach employs two feature subspace classifiers to collectively select most informative minority-class samples for manual annotation by leveraging a certainty measurement and an uncertainty measurement, and in the meanwhile, automatically label most informative majority-class samples, to reduce humanannotation efforts. Extensive experiments across four domains demonstrate great potential and effectiveness of our proposed co-selecting approach to active learning for imbalanced sentiment classification. 1

16 emnlp-2012-Aligning Predicates across Monolingual Comparable Texts using Graph-based Clustering

Author: Michael Roth ; Anette Frank

Abstract: Generating coherent discourse is an important aspect in natural language generation. Our aim is to learn factors that constitute coherent discourse from data, with a focus on how to realize predicate-argument structures in a model that exceeds the sentence level. We present an important subtask for this overall goal, in which we align predicates across comparable texts, admitting partial argument structure correspondence. The contribution of this work is two-fold: We first construct a large corpus resource of comparable texts, including an evaluation set with manual predicate alignments. Secondly, we present a novel approach for aligning predicates across comparable texts using graph-based clustering with Mincuts. Our method significantly outperforms other alignment techniques when applied to this novel alignment task, by a margin of at least 6.5 percentage points in F1-score.

17 emnlp-2012-An "AI readability" Formula for French as a Foreign Language

Author: Thomas Francois ; Cedrick Fairon

Abstract: This paper present a new readability formula for French as a foreign language (FFL), which relies on 46 textual features representative of the lexical, syntactic, and semantic levels as well as some of the specificities of the FFL context. We report comparisons between several techniques for feature selection and various learning algorithms. Our best model, based on support vector machines (SVM), significantly outperforms previous FFL formulas. We also found that semantic features behave poorly in our case, in contrast with some previous readability studies on English as a first language.

18 emnlp-2012-An Empirical Investigation of Statistical Significance in NLP

Author: Taylor Berg-Kirkpatrick ; David Burkett ; Dan Klein

Abstract: We investigate two aspects of the empirical behavior of paired significance tests for NLP systems. First, when one system appears to outperform another, how does significance level relate in practice to the magnitude of the gain, to the size of the test set, to the similarity of the systems, and so on? Is it true that for each task there is a gain which roughly implies significance? We explore these issues across a range of NLP tasks using both large collections of past systems’ outputs and variants of single systems. Next, once significance levels are computed, how well does the standard i.i.d. notion of significance hold up in practical settings where future distributions are neither independent nor identically distributed, such as across domains? We explore this question using a range of test set variations for constituency parsing.

19 emnlp-2012-An Entity-Topic Model for Entity Linking

Author: Xianpei Han ; Le Sun

Abstract: Entity Linking (EL) has received considerable attention in recent years. Given many name mentions in a document, the goal of EL is to predict their referent entities in a knowledge base. Traditionally, there have been two distinct directions of EL research: one focusing on the effects of mention’s context compatibility, assuming that “the referent entity of a mention is reflected by its context”; the other dealing with the effects of document’s topic coherence, assuming that “a mention ’s referent entity should be coherent with the document’ ’s main topics”. In this paper, we propose a generative model called entitytopic model, to effectively join the above two complementary directions together. By jointly modeling and exploiting the context compatibility, the topic coherence and the correlation between them, our model can – accurately link all mentions in a document using both the local information (including the words and the mentions in a document) and the global knowledge (including the topic knowledge, the entity context knowledge and the entity name knowledge). Experimental results demonstrate the effectiveness of the proposed model. 1

20 emnlp-2012-Answering Opinion Questions on Products by Exploiting Hierarchical Organization of Consumer Reviews

Author: Jianxing Yu ; Zheng-Jun Zha ; Tat-Seng Chua

Abstract: This paper proposes to generate appropriate answers for opinion questions about products by exploiting the hierarchical organization of consumer reviews. The hierarchy organizes product aspects as nodes following their parent-child relations. For each aspect, the reviews and corresponding opinions on this aspect are stored. We develop a new framework for opinion Questions Answering, which enables accurate question analysis and effective answer generation by making use the hierarchy. In particular, we first identify the (explicit/implicit) product aspects asked in the questions and their sub-aspects by referring to the hierarchy. We then retrieve the corresponding review fragments relevant to the aspects from the hierarchy. In order to gener- ate appropriate answers from the review fragments, we develop a multi-criteria optimization approach for answer generation by simultaneously taking into account review salience, coherence, diversity, and parent-child relations among the aspects. We conduct evaluations on 11 popular products in four domains. The evaluated corpus contains 70,359 consumer reviews and 220 questions on these products. Experimental results demonstrate the effectiveness of our approach.

21 emnlp-2012-Assessment of ESL Learners' Syntactic Competence Based on Similarity Measures

22 emnlp-2012-Automatically Constructing a Normalisation Dictionary for Microblogs

23 emnlp-2012-Besting the Quiz Master: Crowdsourcing Incremental Classification Games

24 emnlp-2012-Biased Representation Learning for Domain Adaptation

25 emnlp-2012-Bilingual Lexicon Extraction from Comparable Corpora Using Label Propagation

26 emnlp-2012-Building a Lightweight Semantic Model for Unsupervised Information Extraction on Short Listings

27 emnlp-2012-Characterizing Stylistic Elements in Syntactic Structure

28 emnlp-2012-Collocation Polarity Disambiguation Using Web-based Pseudo Contexts

29 emnlp-2012-Concurrent Acquisition of Word Meaning and Lexical Categories

30 emnlp-2012-Constructing Task-Specific Taxonomies for Document Collection Browsing

31 emnlp-2012-Cross-Lingual Language Modeling with Syntactic Reordering for Low-Resource Speech Recognition

32 emnlp-2012-Detecting Subgroups in Online Discussions by Modeling Positive and Negative Relations among Participants

33 emnlp-2012-Discovering Diverse and Salient Threads in Document Collections

34 emnlp-2012-Do Neighbours Help? An Exploration of Graph-based Algorithms for Cross-domain Sentiment Classification

35 emnlp-2012-Document-Wide Decoding for Phrase-Based Statistical Machine Translation

36 emnlp-2012-Domain Adaptation for Coreference Resolution: An Adaptive Ensemble Approach

37 emnlp-2012-Dynamic Programming for Higher Order Parsing of Gap-Minding Trees

38 emnlp-2012-Employing Compositional Semantics and Discourse Consistency in Chinese Event Extraction

39 emnlp-2012-Enlarging Paraphrase Collections through Generalization and Instantiation

40 emnlp-2012-Ensemble Semantics for Large-scale Unsupervised Relation Extraction

41 emnlp-2012-Entity based QA Retrieval

42 emnlp-2012-Entropy-based Pruning for Phrase-based Machine Translation

43 emnlp-2012-Exact Sampling and Decoding in High-Order Hidden Markov Models

44 emnlp-2012-Excitatory or Inhibitory: A New Semantic Orientation Extracts Contradiction and Causality from the Web

45 emnlp-2012-Exploiting Chunk-level Features to Improve Phrase Chunking

46 emnlp-2012-Exploiting Reducibility in Unsupervised Dependency Parsing

47 emnlp-2012-Explore Person Specific Evidence in Web Person Name Disambiguation

48 emnlp-2012-Exploring Adaptor Grammars for Native Language Identification

49 emnlp-2012-Exploring Topic Coherence over Many Models and Many Topics

50 emnlp-2012-Extending Machine Translation Evaluation Metrics with Lexical Cohesion to Document Level

51 emnlp-2012-Extracting Opinion Expressions with semi-Markov Conditional Random Fields

52 emnlp-2012-Fast Large-Scale Approximate Graph Construction for NLP

53 emnlp-2012-First Order vs. Higher Order Modification in Distributional Semantics

54 emnlp-2012-Forced Derivation Tree based Model Training to Statistical Machine Translation

55 emnlp-2012-Forest Reranking through Subtree Ranking

56 emnlp-2012-Framework of Automatic Text Summarization Using Reinforcement Learning

57 emnlp-2012-Generalized Higher-Order Dependency Parsing with Cube Pruning

58 emnlp-2012-Generalizing Sub-sentential Paraphrase Acquisition across Original Signal Type of Text Pairs

59 emnlp-2012-Generating Non-Projective Word Order in Statistical Linearization

60 emnlp-2012-Generative Goal-Driven User Simulation for Dialog Management

61 emnlp-2012-Grounded Models of Semantic Representation

62 emnlp-2012-Identifying Constant and Unique Relations by using Time-Series Text

63 emnlp-2012-Identifying Event-related Bursts via Social Media Activities

64 emnlp-2012-Improved Parsing and POS Tagging Using Inter-Sentence Consistency Constraints

65 emnlp-2012-Improving NLP through Marginalization of Hidden Syntactic Structure

66 emnlp-2012-Improving Transition-Based Dependency Parsing with Buffer Transitions

67 emnlp-2012-Inducing a Discriminative Parser to Optimize Machine Translation Reordering

68 emnlp-2012-Iterative Annotation Transformation with Predict-Self Reestimation for Chinese Word Segmentation

69 emnlp-2012-Joining Forces Pays Off: Multilingual Joint Word Sense Disambiguation

70 emnlp-2012-Joint Chinese Word Segmentation, POS Tagging and Parsing

71 emnlp-2012-Joint Entity and Event Coreference Resolution across Documents

72 emnlp-2012-Joint Inference for Event Timeline Construction

73 emnlp-2012-Joint Learning for Coreference Resolution with Markov Logic

74 emnlp-2012-Language Model Rest Costs and Space-Efficient Storage

75 emnlp-2012-Large Scale Decipherment for Out-of-Domain Machine Translation

76 emnlp-2012-Learning-based Multi-Sieve Co-reference Resolution with Knowledge

77 emnlp-2012-Learning Constraints for Consistent Timeline Extraction

78 emnlp-2012-Learning Lexicon Models from Search Logs for Query Expansion

79 emnlp-2012-Learning Syntactic Categories Using Paradigmatic Representations of Word Context

80 emnlp-2012-Learning Verb Inference Rules from Linguistically-Motivated Evidence

81 emnlp-2012-Learning to Map into a Universal POS Tagset

82 emnlp-2012-Left-to-Right Tree-to-String Decoding with Prediction

83 emnlp-2012-Lexical Differences in Autobiographical Narratives from Schizophrenic Patients and Healthy Controls

84 emnlp-2012-Linking Named Entities to Any Database

85 emnlp-2012-Local and Global Context for Supervised and Unsupervised Metonymy Resolution

86 emnlp-2012-Locally Training the Log-Linear Model for SMT

87 emnlp-2012-Lyrics, Music, and Emotions

88 emnlp-2012-Minimal Dependency Length in Realization Ranking

89 emnlp-2012-Mixed Membership Markov Models for Unsupervised Conversation Modeling

90 emnlp-2012-Modelling Sequential Text with an Adaptive Topic Model

91 emnlp-2012-Monte Carlo MCMC: Efficient Inference by Approximate Sampling

92 emnlp-2012-Multi-Domain Learning: When Do Domains Matter?

93 emnlp-2012-Multi-instance Multi-label Learning for Relation Extraction

94 emnlp-2012-Multiple Aspect Summarization Using Integer Linear Programming

95 emnlp-2012-N-gram-based Tense Models for Statistical Machine Translation

96 emnlp-2012-Name Phylogeny: A Generative Model of String Variation

97 emnlp-2012-Natural Language Questions for the Web of Data

98 emnlp-2012-No Noun Phrase Left Behind: Detecting and Typing Unlinkable Entities

99 emnlp-2012-On Amortizing Inference Cost for Structured Prediction

100 emnlp-2012-Open Language Learning for Information Extraction

101 emnlp-2012-Opinion Target Extraction Using Word-Based Translation Model

102 emnlp-2012-Optimising Incremental Dialogue Decisions Using Information Density for Interactive Systems

103 emnlp-2012-PATTY: A Taxonomy of Relational Patterns with Semantic Types

104 emnlp-2012-Parse, Price and Cut-Delayed Column and Row Generation for Graph Based Parsers

105 emnlp-2012-Parser Showdown at the Wall Street Corral: An Empirical Investigation of Error Types in Parser Output

106 emnlp-2012-Part-of-Speech Tagging for Chinese-English Mixed Texts with Dynamic Features

107 emnlp-2012-Polarity Inducing Latent Semantic Analysis

108 emnlp-2012-Probabilistic Finite State Machines for Regression-based MT Evaluation

109 emnlp-2012-Re-training Monolingual Parser Bilingually for Syntactic SMT

110 emnlp-2012-Reading The Web with Learned Syntactic-Semantic Inference Rules

111 emnlp-2012-Regularized Interlingual Projections: Evaluation on Multilingual Transliteration

112 emnlp-2012-Resolving Complex Cases of Definite Pronouns: The Winograd Schema Challenge

113 emnlp-2012-Resolving This-issue Anaphora

114 emnlp-2012-Revisiting the Predictability of Language: Response Completion in Social Media

115 emnlp-2012-SSHLDA: A Semi-Supervised Hierarchical Topic Model

116 emnlp-2012-Semantic Compositionality through Recursive Matrix-Vector Spaces

117 emnlp-2012-Sketch Algorithms for Estimating Point Queries in NLP

118 emnlp-2012-Source Language Adaptation for Resource-Poor Machine Translation

119 emnlp-2012-Spectral Dependency Parsing with Latent Variables

120 emnlp-2012-Streaming Analysis of Discourse Participants

121 emnlp-2012-Supervised Text-based Geolocation Using Language Models on an Adaptive Grid

122 emnlp-2012-Syntactic Surprisal Affects Spoken Word Duration in Conversational Contexts

123 emnlp-2012-Syntactic Transfer Using a Bilingual Lexicon

124 emnlp-2012-Three Dependency-and-Boundary Models for Grammar Induction

125 emnlp-2012-Towards Efficient Named-Entity Rule Induction for Customizability

126 emnlp-2012-Training Factored PCFGs with Expectation Propagation

127 emnlp-2012-Transforming Trees to Improve Syntactic Convergence

128 emnlp-2012-Translation Model Based Cross-Lingual Language Model Adaptation: from Word Models to Phrase Models

129 emnlp-2012-Type-Supervised Hidden Markov Models for Part-of-Speech Tagging with Incomplete Tag Dictionaries

130 emnlp-2012-Unambiguity Regularization for Unsupervised Learning of Probabilistic Grammars

131 emnlp-2012-Unified Dependency Parsing of Chinese Morphological and Syntactic Structures

132 emnlp-2012-Universal Grapheme-to-Phoneme Prediction Over Latin Alphabets

133 emnlp-2012-Unsupervised PCFG Induction for Grounded Language Learning with Highly Ambiguous Supervision

134 emnlp-2012-User Demographics and Language in an Implicit Social Network

135 emnlp-2012-Using Discourse Information for Paraphrase Extraction

136 emnlp-2012-Weakly Supervised Training of Semantic Parsers

137 emnlp-2012-Why Question Answering using Sentiment Analysis and Word Classes

138 emnlp-2012-Wiki-ly Supervised Part-of-Speech Tagging

139 emnlp-2012-Word Salad: Relating Food Prices and Descriptions