Distributed Representations of Words and Phrases and their Compositionality

Volume: 26, Pages: 3111 - 3119
Published: Dec 5, 2013
Abstract
The recently introduced continuous Skip-gram model is an efficient method for learning high-quality distributed vector representations that capture a large number of precise syntactic and semantic word relationships. In this paper we present several extensions that improve both the quality of the vectors and the training speed. By subsampling of the frequent words we obtain significant speedup and also learn more regular word representations. We...
Paper Details
Title
Distributed Representations of Words and Phrases and their Compositionality
Published Date
Dec 5, 2013
Volume
26
Pages
3111 - 3119
Citation AnalysisPro
  • Scinapse’s Top 10 Citation Journals & Affiliations graph reveals the quality and authenticity of citations received by a paper.
  • Discover whether citations have been inflated due to self-citations, or if citations include institutional bias.