What you can cram into a single \$&!#* vector: Probing sentence embeddings for linguistic properties - INRIA - Institut National de Recherche en Informatique et en Automatique Accéder directement au contenu
Communication Dans Un Congrès Année : 2018

What you can cram into a single \$&!#* vector: Probing sentence embeddings for linguistic properties

Résumé

Although much effort has recently been devoted to training high-quality sentence embeddings, we still have a poor understanding of what they are capturing. "Downstream" tasks, often based on sentence classification, are commonly used to evaluate the quality of sentence representations. The complexity of the tasks makes it however difficult to infer what kind of information is present in the representations. We introduce here 10 probing tasks designed to capture simple linguistic features of sentences, and we use them to study embeddings generated by three different encoders trained in eight distinct ways, uncovering intriguing properties of both encoders and training methods.

Dates et versions

hal-01898412 , version 1 (18-10-2018)

Identifiants

Citer

Alexis Conneau, German Kruszewski, Guillaume Lample, Loïc Barrault, Marco Baroni. What you can cram into a single \$&!#* vector: Probing sentence embeddings for linguistic properties. ACL 2018 - 56th Annual Meeting of the Association for Computational Linguistics, Jul 2018, Melbourne, Australia. pp.2126-2136. ⟨hal-01898412⟩
214 Consultations
0 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More