Do self-supervised speech models develop human-like perception biases? - INRIA - Institut National de Recherche en Informatique et en Automatique Accéder directement au contenu
Communication Dans Un Congrès Année : 2022

Do self-supervised speech models develop human-like perception biases?

Résumé

Self-supervised models for speech processing form representational spaces without using any external labels. Increasingly, they appear to be a feasible way of at least partially eliminating costly manual annotations, a problem of particular concern for low-resource languages. But what kind of representational spaces do these models construct? Human perception specializes to the sounds of listeners' native languages. Does the same thing happen in self-supervised models? We examine the representational spaces of three kinds of stateof-the-art self-supervised models: wav2vec 2.0, HuBERT and contrastive predictive coding (CPC), and compare them with the perceptual spaces of French-speaking and Englishspeaking human listeners, both globally and taking account of the behavioural differences between the two language groups. We show that the CPC model shows a small native language effect, but that wav2vec 2.0 and Hu-BERT seem to develop a universal speech perception space which is not language specific. A comparison against the predictions of supervised phone recognisers suggests that all three self-supervised models capture relatively finegrained perceptual phenomena, while supervised models are better at capturing coarser, phone-level, effects of listeners' native language, on perception.
Fichier principal
Vignette du fichier
_ACL__Do_self_supervised_speech_models_develop_human_like_perception_biases__.pdf (1.55 Mo) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-03697420 , version 1 (22-06-2022)

Identifiants

Citer

Juliette Millet, Ewan Dunbar. Do self-supervised speech models develop human-like perception biases?. ACL 2022 - 60th Annual Meeting of the Association for Computational Linguistics, May 2022, Dublin, Ireland. pp.7591-7605, ⟨10.18653/v1/2022.acl-long.523⟩. ⟨hal-03697420⟩
44 Consultations
22 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More