On the Study of Transformers for Query Suggestion - Sorbonne Université Access content directly
Journal Articles ACM Transactions on Information Systems Year : 2022

On the Study of Transformers for Query Suggestion

Abstract

When conducting a search task, users may find it difficult to articulate their need, even more so when the task is complex. To help them complete their search, search engine usually provide query suggestions. A good query suggestion system requires to model user behavior during the search session. In this article, we study multiple Transformer architectures applied to the query suggestion task and compare them with recurrent neural network (RNN)-based models. We experiment Transformer models with different tokenizers, with different Encoders (large pretrained models or fully trained ones), and with two kinds of architectures (flat or hierarchic). We study the performance and the behaviors of these various models, and observe that Transformer-based models outperform RNN-based ones. We show that while the hierarchical architectures exhibit very good performances for query suggestion, the flat models are more suitable for complex and long search tasks. Finally, we investigate the flat models behavior and demonstrate that they indeed learn to recover the hierarchy of a search session.
Fichier principal
Vignette du fichier
TOIS_2020_HAL.pdf (1.11 Mo) Télécharger le fichier
Origin : Files produced by the author(s)

Dates and versions

hal-03541893 , version 1 (10-02-2022)

Identifiers

Cite

Agnès Mustar, Sylvain Lamprier, Benjamin Piwowarski. On the Study of Transformers for Query Suggestion. ACM Transactions on Information Systems, 2022, 40 (1), pp.18. ⟨10.1145/3470562⟩. ⟨hal-03541893⟩
95 View
638 Download

Altmetric

Share

Gmail Facebook X LinkedIn More