LEARNING WORD RELATEDNESS OVER TIME FOR TEMPORAL RANKING

  • Dinda Sigmawaty Universitas Indonesia
  • Mirna Adriani Universitas Indonesia
Keywords: Information Retrieval, temporal ranking, Dual Embedding Space Model, temporal word embeddings

Abstract

Queries and ranking with temporal aspects gain significant attention in field of Information Retrieval. While searching for articles published over time, the relevant documents usually occur in certain temporal patterns. Given a query that is implicitly time sensitive, we develop a temporal ranking using the important times of query by drawing from the distribution of query trend relatedness over time. We also combine the model with Dual Embedding Space Model (DESM) in the temporal model according to document timestamp. We apply our model using three temporal word embeddings algorithms to learn relatedness of words from news archive in Bahasa Indonesia: (1) QT-W2V-Rank using Word2Vec (2) QT-OW2V-Rank using OrthoTrans-Word2Vec (3) QT-DBE-Rank using Dynamic Bernoulli Embeddings. The highest score was achieved with static word embeddings learned separately over time, called QT-W2V-Rank, which is 66% in average precision and 68% in early precision. Furthermore, studies of different characteristics of temporal topics showed that QT-W2V-Rank is also more effective in capturing temporal patterns such as spikes, periodicity, and seasonality than the baselines.

Published
2019-07-08
How to Cite
Sigmawaty, D., & Adriani, M. (2019). LEARNING WORD RELATEDNESS OVER TIME FOR TEMPORAL RANKING. Jurnal Ilmu Komputer Dan Informasi, 12(2), 91-102. https://doi.org/10.21609/jiki.v12i2.745