A Mixture Model with Sharing for Lexical Semantics (2010)
We introduce tiered clustering, a mixture model capable of accounting for varying degrees of shared (context-independent) feature structure, and demonstrate its applicability to inferring distributed representations of word meaning. Common tasks in lexical semantics such as word relatedness or selectional preference can benefit from modeling such structure: Polysemous word usage is often governed by some common background metaphoric usage (e.g. the senses of line or run), and likewise modeling the selectional preference of verbs relies on identifying commonalities shared by their typical arguments. Tiered clustering can also be viewed as a form of soft feature selection, where features that do not contribute meaningfully to the clustering can be excluded. We demonstrate the applicability of tiered clustering, highlighting particular cases where modeling shared structure is beneficial and where it can be detrimental.
In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP-2010), pp. 1173--1182, MIT, Massachusetts, USA, October 9--11 2010.

Slides (PDF)
Raymond J. Mooney Faculty mooney [at] cs utexas edu
Joseph Reisinger Ph.D. Alumni joeraii [at] cs utexas edu