ABSTRACT
Lexical cooccurrence in textual data is not uniformly random. The statistics inferred from the term-cooccurrence data enable us to model dependencies between terms as graphs, somewhat resembling the way semantic memory is organised in human beings. In this paper we look at cooccurrence patterns to identify topical anchors of a given context. Topical anchors are those terms whose semantics represent the topic of the whole context. This work is based on computing a stationary distribution in the cooccurrence graph. Topical anchors were computed on a set of 100 contexts and were also evaluated by 86 volunteers and the results show that the algorithm correctly identifies the topical anchors around 62% of the time.
- S. Abiteboul, M. Preda, and G. Cobena. Adaptive on-line page importance computation. In WWW '03: Proceedings of the 12th international conference on World Wide Web, 2003. Google ScholarDigital Library
- V. Evans, B. K. Bergen, and J. Zinken. The Cognitive Linguistics Enterprise: An Overview. 2006.Google Scholar
- L. Fu. Neural Networks in Computer Intelligence. McGraw-Hill, Inc., 1994. Google ScholarDigital Library
- S. Mcdonald and M. Ramscar. Testing the distributional hypothesis: The influence of context on judgements of semantic similarity. In Proceedings of the 23rd Annual Conference of the Cognitive Science Society, 2001.Google Scholar
- L. Page, S. Brin, R. Motwani, and T. Winograd. The pagerank citation ranking: Bringing order to the web. Technical report, Stanford Digital Library Technologies Project, 1998.Google Scholar
- M. Sahlgren. Vector-based semantic analysis: Representing word meanings based on random labels. In ESSLI Workshop on Semantic Knowledge Acquistion and Categorization, 2001.Google Scholar
- C. J. van Rijsbergen. A theoritical basis for the use of co--occurrence data in information retrieval. Journal of Documentation, 1977.Google Scholar
Index Terms
- Finding the topical anchors of a context using lexical cooccurrence data
Recommendations
Sentiment classification of blog posts using topical extracts
ADC '12: Proceedings of the Twenty-Third Australasian Database Conference - Volume 124Unlike news stories and product reviews which usually have a strong focus on a single topic, blog posts are often unstructured, and opinions expressed in blog posts do not necessarily correspond to a specific topic. This can lead to unsatisfactory ...
Topical N-Grams: Phrase and Topic Discovery, with an Application to Information Retrieval
ICDM '07: Proceedings of the 2007 Seventh IEEE International Conference on Data MiningMost topic models, such as latent Dirichlet allocation, rely on the bag-of-words assumption. However, word order and phrases are often critical to capturing the meaning of text in many text mining tasks. This paper presents topical n-grams, a topic ...
Affine invariant topic model for generic object recognition
ISNN'10: Proceedings of the 7th international conference on Advances in Neural Networks - Volume Part IIThis paper presents a novel topic model named Affine Invariant Topic Model(AITM) for generic object recognition Abandoning the “bag of words” assumption in traditional topic models, AITM incorporates spatial structure into traditional LDA AITM extends ...
Comments