What’s in an Embedding? Analyzing Word Embeddings through Multilingual Evaluation

In this presentation, I wil primarily talk about the results of my paper (with the same name). In addition, I will give a short introduction into word embeddings and how they can be learned.

For more information: My paper for EMNLP 2015 and an overview of the results

Abstract

In the last two years, there has been a surge of word embedding algorithms and research on them. However, evaluation has mostly been carried out on a narrow set of tasks, mainly word similarity/relatedness and word relation similarity and on a single language, namely English. We propose an approach to evaluate embeddings on a variety of languages that also yields insights into the structure of the embedding space by investigating how well word embeddings cluster along different syntactic features. We show that all embedding approaches behave similarly in this task, with dependency-based embeddings performing best. This effect is even more pronounced when generating low dimensional embeddings.

-- ArneKoehn - 20 Nov 2015

This topic: Oberseminar > WebHome > 2015TalkKoehn
Topic revision: 20 Nov 2015, ArneKoehn
 
This site is powered by FoswikiCopyright © by the contributing authors. All material on this collaboration platform is the property of the contributing authors.
Ideas, requests, problems regarding Foswiki? Send feedback