Joint Embeddings of Scene Graphs and Images - INRIA - Institut National de Recherche en Informatique et en Automatique Accéder directement au contenu
Poster De Conférence Année : 2017

Joint Embeddings of Scene Graphs and Images

Résumé

Multimodal representations of text and images have become popular in recent years. Text however has inherent ambiguities when describing visual scenes, leading to the recent development of datasets with detailed graphical descriptions in the form of scene graphs. We consider the task of joint representation of semantically precise scene graphs and images. We propose models for representing scene graphs and aligning them with images. We investigate methods based on bag-of-words, subpath representations, as well as neural networks. Our investigation proposes and contrasts several models which can address this task and highlights some unique challenges in both designing models and evaluation.
Fichier principal
Vignette du fichier
main.pdf (99.66 Ko) Télécharger le fichier
iclr2017_workshop.bst (26.34 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)
Loading...

Dates et versions

hal-01667777 , version 1 (19-12-2017)

Identifiants

  • HAL Id : hal-01667777 , version 1

Citer

Eugene Belilovsky, Matthew Blaschko, Jamie Ryan Kiros, Raquel Urtasun, Richard Zemel. Joint Embeddings of Scene Graphs and Images. International Conference On Learning Representations - Workshop, 2017, Toulon, France. ⟨hal-01667777⟩
257 Consultations
448 Téléchargements

Partager

Gmail Facebook X LinkedIn More