Improving Document Ranking with Dual Word Embeddings
- Eric Nalisnick ,
- Bhaskar Mitra ,
- Nick Craswell ,
- Rich Caruana
WWW'16 |
Published by WWW - World Wide Web Consortium (W3C)
This paper investigates the popular neural word embedding method Word2vec as a source of evidence in document ranking. In contrast to NLP applications of word2vec, which tend to use only the input embeddings, we retain both the input and the output embeddings, allowing us to calculate a different word similarity that may be more suitable for document ranking. We map the query words into the input space and the document words into the output space, and compute a relevance score by aggregating the cosine similarities across all the query-document word pairs. We postulate that the proposed Dual Embedding Space Model (DESM) provides evidence that a document is about a query term, in addition to and complementing the traditional term frequency based approach.