Lexical Relation Mining in Neural Word Embeddings

Aishwarya Jadhav, Yifat Amir, Zachary Pardos
2020 Proceedings of the 28th International Conference on Computational Linguistics   unpublished
Work with neural word embeddings and lexical relations has largely focused on confirmatory experiments which use human-curated examples of semantic and syntactic relations to validate against. In this paper, we explore the degree to which lexical relations, such as those found in popular validation sets, can be derived and extended from a variety of neural embeddings using classical clustering methods. We show that the Word2Vec space of word-pairs (i.e., offset vectors) significantly
more » ... ficantly outperforms other more contemporary methods, even in the presence of a large number of noisy offsets. Moreover, we show that via a simple nearest neighbor approach in the offset space, new examples of known relations can be discovered. Our results speak to the amenability of offset vectors from non-contextual neural embeddings to find semantically coherent clusters. This simple approach has implications for the exploration of emergent regularities and their examples, such as emerging trends on social media and their related posts.
doi:10.18653/v1/2020.coling-main.112 fatcat:iaqyb7ag3fhmhcxwr6ivp66pwq