On the Study of Transformers for Query Suggestion - Sorbonne Université
Article Dans Une Revue ACM Transactions on Information Systems Année : 2022

On the Study of Transformers for Query Suggestion

Résumé

When conducting a search task, users may find it difficult to articulate their need, even more so when the task is complex. To help them complete their search, search engine usually provide query suggestions. A good query suggestion system requires to model user behavior during the search session. In this article, we study multiple Transformer architectures applied to the query suggestion task and compare them with recurrent neural network (RNN)-based models. We experiment Transformer models with different tokenizers, with different Encoders (large pretrained models or fully trained ones), and with two kinds of architectures (flat or hierarchic). We study the performance and the behaviors of these various models, and observe that Transformer-based models outperform RNN-based ones. We show that while the hierarchical architectures exhibit very good performances for query suggestion, the flat models are more suitable for complex and long search tasks. Finally, we investigate the flat models behavior and demonstrate that they indeed learn to recover the hierarchy of a search session.
Fichier principal
Vignette du fichier
TOIS_2020_HAL.pdf (1.11 Mo) Télécharger le fichier
Origine Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-03541893 , version 1 (10-02-2022)

Identifiants

Citer

Agnès Mustar, Sylvain Lamprier, Benjamin Piwowarski. On the Study of Transformers for Query Suggestion. ACM Transactions on Information Systems, 2022, 40 (1), pp.18. ⟨10.1145/3470562⟩. ⟨hal-03541893⟩
116 Consultations
853 Téléchargements

Altmetric

Partager

More