The Sensitivity of Language Models and Humans to Winograd Schema Perturbations

Publikation: Bidrag til bog/antologi/rapportKonferencebidrag i proceedingsForskningfagfællebedømt

Dokumenter

Large-scale pretrained language models are the major driving force behind recent improvements in perfromance on the Winograd Schema Challenge, a widely employed test of commonsense reasoning ability. We show, however, with a new diagnostic dataset, that these models are sensitive to linguistic perturbations of the Winograd examples that minimally affect human understanding. Our results highlight interesting differences between humans and language models: language models are more sensitive to number or gender alternations and synonym replacements than humans, and humans are more stable and consistent in their predictions, maintain a much higher absolute performance, and perform better on non-associative instances than associative ones.
OriginalsprogEngelsk
TitelProceedings of the 58th Annual Meeting of the Association for Computational Linguistics
ForlagAssociation for Computational Linguistics
Publikationsdato2020
Sider7590-7604
DOI
StatusUdgivet - 2020
Begivenhed58th Annual Meeting of the Association for Computational Linguistics - Online
Varighed: 5 jul. 202010 jul. 2020

Konference

Konference58th Annual Meeting of the Association for Computational Linguistics
ByOnline
Periode05/07/202010/07/2020

Antal downloads er baseret på statistik fra Google Scholar og www.ku.dk


Ingen data tilgængelig

ID: 258374819