emnlp emnlp2010 emnlp2010-77 emnlp2010-77-reference knowledge-graph by maker-knowledge-mining
Source: pdf
Author: Georgiana Dinu ; Mirella Lapata
Abstract: The computation of meaning similarity as operationalized by vector-based models has found widespread use in many tasks ranging from the acquisition of synonyms and paraphrases to word sense disambiguation and textual entailment. Vector-based models are typically directed at representing words in isolation and thus best suited for measuring similarity out of context. In his paper we propose a probabilistic framework for measuring similarity in context. Central to our approach is the intuition that word meaning is represented as a probability distribution over a set of latent senses and is modulated by context. Experimental results on lexical substitution and word similarity show that our algorithm outperforms previously proposed models.
David M. Blei, Andrew Y. Ng, and Michael I. Jordan. 2003. Latent Dirichlet Allocation. Journal of Machine Learning Research, 3:993–1022. Daoud Clarke. 2009. Context-theoretic semantics for natural language: an overview. In Proceedings of the Workshop on Geometrical Models of Natural Language Semantics, pages 112–1 19, Athens, Greece. Scott Deerwester, Susan T. Dumais, George W. Furnas, Thomas, and Richard Harshman. 1990. Indexing by latent semantic analysis. Journal of the American Society for Information Science, 41:391–407. Chris Ding, Tao Li, and Wei Peng. 2008. On the equivalence between non-negative matrix factorization and probabilistic latent semantic indexing. Computational Statistics & Data Analysis, 52(8):3913–3927. Katrin Erk and Sabastian Pad o´. 2008. A structured vector space model for word meaning in context. In Proceedings of the 2008 Conference on Empirical Methods in Natural Language Processing, pages 897–906, Honolulu, Hawaii. Katrin Erk and Sebastian Pad o´. 2010. Exemplar-based models for word meaning in context. In Proceedings of the ACL 2010 Conference Short Papers, pages 92– 97, Uppsala, Sweden. Lev Finkelstein, Evgeniy Gabrilovich, Yossi Matias, Ehud Rivlin, Zach Solan, Gadi Wolfman, and Eytan Ruppin. 2002. Placing search in context: the concept revisited. ACM Transactions on Information Systems, 20(1):1 16–131. Eric Gaussier and Cyril Goutte. 2005. Relation between PLSA and NMF and implications. In Proceedings of the 28th Annual international ACM SIGIR conference on Research and development in information retrieval, pages 601–602, New York, NY. Gregory Grefenstette. 1994. Explorations in Automatic Thesaurus Discovery. Kluwer Academic Publishers. Thomas L. Griffiths and Mark Steyvers. 2004. Finding scientific topics. Proceedings of the National Academy of Sciences, 101(Suppl. 1):5228–5235. Thomas Hofmann. 2001. Unsupervised learning by probabilistic latent semantic analysis. Machine Learn- ing, 41(2): 177–196. Walter Kintsch. 2001. Predication. Cognitive Science, 25: 173–202. Kazuaki Kishida. 2005. Property of average precision and its generalization: An examination of evaluation indicator for information retrieval experiments. NII Technical Report. Thomas K. Landauer and Susan T. Dumais. 1997. A solution to Plato’s problem: The latent semantic analysis theory of acquisition, induction and representation of knowledge. Psychological Review, 104(2):21 1–240. 1171 Daniel D. Lee and H. Sebastian Seung. 2000. Algorithms for non-negative matrix factorization. In NIPS, pages 556–562. Dekang Lin and Patrick Pantel. 2001. Discovery ofinference rules for question answering. Natural Language Engineering, 7(4):342–360. Dekang Lin. 1998. Automatic retrieval and clustering of similar words. In Proceedings of the joint Annual Meeting of the Association for Computational Linguistics and International Conference on Computational Linguistics, pages 768–774, Montr ´eal, Canada. Will Lowe and Scott McDonald. 2000. The direct route: Mediated priming in semantic space. In Proceedings of the 22nd Annual Conference of the Cognitive Sci- ence Society, pages 675–680, Philadelphia, PA. Kevin Lund and Curt Burgess. 1996. Producing high-dimensional semantic spaces from lexical cooccurrence. Behavior Research Methods, Instruments, and Computers, 28:203–208. Diana McCarthy and Roberto Navigli. 2007. SemEval2007 Task 10: English Lexical Substitution Task. In Proceedings of SemEval, pages 48–53, Prague, Czech Republic. Scott McDonald. 2000. Environmental Determinants of Lexical Processing Effort. Ph.D. thesis, University of Edinburgh. Jeff Mitchell and Mirella Lapata. 2008. Vector-based models of semantic composition. In Proceedings of ACL-08: HLT, pages 236–244, Columbus, Ohio. Jeff Mitchell and Mirella Lapata. 2009. Language models based on semantic composition. In Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing, pages 430–439, Suntec, Singapore. Sebastian Pad o´ and Mirella Lapata. 2007. Dependencybased construction of semantic space models. Computational Linguistics, 33(2): 161–199. Ian Porteous, David Newman, Alexander Ihler, Arthur Asuncion, Padhraic Smyth, and Max Welling. 2008. Fast collapsed gibbs sampling for latent Dirichlet allocation. In Proceeding of the 14th ACM SIGKDD inter- national conference on Knowledge discovery and data 569–577, New York, NY. Joseph Reisinger and Raymond J. Mooney. 2010. Multiprototype vector-space models of word meaning. In Human Language Technologies: The 2010 Annual mining, pages Conference of the North American Chapter of the As- for Computational Linguistics, 117, Los Angeles, California. G Salton, A Wang, and C Yang. 1975. A sociation pages 109– vector-space model for information retrieval. Journal of the American Society for Information 18:613–620. Science, Hinrich Schuetze. 1998. Automatic word sense discrimination. Journal of Computational Linguistics, 24:97– 123. Stefan Thater, Georgiana Dinu, and Manfred Pinkal. 2009. Ranking paraphrases in context. In Proceedings of the 2009 Workshop on Applied Textual Inference, pages 44–47, Suntec, Singapore. Stefan Thater, Hagen F ¨urstenau, and Manfred Pinkal. 2010. Contextualizing semantic representations using syntactically enriched vector models. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics, pages 948–957, Uppsala, Sweden. Michael E. Tipping and Chris M. Bishop. 1999. Probabilistic principal component analysis. Journal of the Royal Statistical Society, Series B, 61:611–622. Alexander Yeh. 2000. More accurate tests for the statistical significance of result differences. In Proceedings of the 18th Conference on Computational Linguistics, pages 947–953, Saarbr¨ ucken, Germany. 1172