MUREL: Multimodal Relational Reasoning for Visual Question Answering - Sorbonne Université Access content directly
Conference Papers Year : 2019

MUREL: Multimodal Relational Reasoning for Visual Question Answering


Multimodal attentional networks are currently state-of-the-art models for Visual Question Answering (VQA) tasks involving real images. Although attention allows to focus on the visual content relevant to the question, this simple mechanism is arguably insufficient to model complex reasoning features required for VQA or other high-level tasks. In this paper, we propose MuRel, a multimodal relational network which is learned end-to-end to reason over real images. Our first contribution is the introduction of the MuRel cell, an atomic reasoning primitive representing interactions between question and image regions by a rich vectorial representation, and modeling region relations with pairwise combinations. Secondly, we incorporate the cell into a full MuRel network, which progressively refines visual and question interactions, and can be leveraged to define visualization schemes finer than mere attention maps. We validate the relevance of our approach with various ablation studies, and show its superiority to attention-based methods on three datasets: VQA 2.0, VQA-CP v2 and TDIUC. Our final MuRel network is competitive to or outperforms state-of-the-art results in this challenging context. Our code is available:
Fichier principal
Vignette du fichier
1902.09487.pdf (4.57 Mo) Télécharger le fichier
Origin : Files produced by the author(s)

Dates and versions

hal-02073649 , version 1 (26-03-2019)



Remi Cadene, Hedi Ben-Younes, Matthieu Cord, Nicolas Thome. MUREL: Multimodal Relational Reasoning for Visual Question Answering. The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Jun 2019, Long Beach, CA, United States. ⟨hal-02073649⟩
235 View
84 Download



Gmail Facebook X LinkedIn More