作者: Manaal Faruqui , Yulia Tsvetkov , Dani Yogatama , Chris Dyer , Noah A. Smith
DOI: 10.3115/V1/P15-1144
关键词:
摘要: Current distributed representations of words show little resemblance to theories lexical semantics. The former are dense and uninterpretable, the latter largely based on familiar, discrete classes (e.g., supersenses) relations synonymy hypernymy). We propose methods that transform word vectors into sparse (and optionally binary) vectors. resulting more similar interpretable features typically used in NLP, though they discovered automatically from raw corpora. Because highly sparse, computationally easy work with. Most importantly, we find outperform original benchmark tasks.