LEARNING WORD RELATEDNESS OVER TIME FOR TEMPORAL RANKING

Dinda Sigmawaty, Mirna Adriani

Abstract


Queries and ranking with temporal aspects gain significant attention in field of Information Retrieval. While searching for articles published over time, the relevant documents usually occur in certain temporal patterns. Given a query that is implicitly time sensitive, we develop a temporal ranking using the important times of query by drawing from the distribution of query trend relatedness over time. We also combine the model with Dual Embedding Space Model (DESM) in the temporal model according to document timestamp. We apply our model using three temporal word embeddings algorithms to learn relatedness of words from news archive in Bahasa Indonesia: (1) QT-W2V-Rank using Word2Vec (2) QT-OW2V-Rank using OrthoTrans-Word2Vec (3) QT-DBE-Rank using Dynamic Bernoulli Embeddings. The highest score was achieved with static word embeddings learned separately over time, called QT-W2V-Rank, which is 66% in average precision and 68% in early precision. Furthermore, studies of different characteristics of temporal topics showed that QT-W2V-Rank is also more effective in capturing temporal patterns such as spikes, periodicity, and seasonality than the baselines.


Keywords


Information Retrieval;temporal ranking;Dual Embedding Space Model;temporal word embeddings

Full Text:

PDF


DOI: http://dx.doi.org/10.21609/jiki.v12i2.745

Refbacks

  • There are currently no refbacks.


Copyright © Jurnal Ilmu Komputer dan Informasi. Faculty of Computer Science Universitas Indonesia.

Creative Commons License

This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.

View JIKI Statistic