Sie befinden Sich nicht im Netzwerk der Universität Paderborn. Der Zugriff auf elektronische Ressourcen ist gegebenenfalls nur via VPN oder Shibboleth (DFN-AAI) möglich. mehr Informationen...
Ergebnis 2 von 1312
Frontiers in artificial intelligence, 2022-03, Vol.5, p.777963-777963
2022

Details

Autor(en) / Beteiligte
Titel
Comparison of Structural Parsers and Neural Language Models as Surprisal Estimators
Ist Teil von
  • Frontiers in artificial intelligence, 2022-03, Vol.5, p.777963-777963
Ort / Verlag
Switzerland: Frontiers Media S.A
Erscheinungsjahr
2022
Link zum Volltext
Quelle
EZB Electronic Journals Library
Beschreibungen/Notizen
  • Expectation-based theories of sentence processing posit that processing difficulty is determined by predictability in context. While predictability quantified surprisal has gained empirical support, this representation-agnostic measure leaves open the question of how to best approximate the human comprehender's latent probability model. This article first describes an incremental left-corner parser that incorporates information about common linguistic abstractions such as syntactic categories, predicate-argument structure, and morphological rules as a computational-level model of sentence processing. The article then evaluates a variety of structural parsers and deep neural language models as cognitive models of sentence processing by comparing the predictive power of their surprisal estimates on self-paced reading, eye-tracking, and fMRI data collected during real-time language processing. The results show that surprisal estimates from the proposed left-corner processing model deliver comparable and often superior fits to self-paced reading and eye-tracking data when compared to those from neural language models trained on much more data. This may suggest that the strong linguistic generalizations made by the proposed processing model may help predict humanlike processing costs that manifest in latency-based measures, even when the amount of training data is limited. Additionally, experiments using Transformer-based language models sharing the same primary architecture and training data show a surprising negative correlation between parameter count and fit to self-paced reading and eye-tracking data. These findings suggest that large-scale neural language models are making weaker generalizations based on patterns of lexical items rather than stronger, more humanlike generalizations based on linguistic structure.
Sprache
Englisch
Identifikatoren
ISSN: 2624-8212
eISSN: 2624-8212
DOI: 10.3389/frai.2022.777963
Titel-ID: cdi_doaj_primary_oai_doaj_org_article_84b0e7517dd84066aaf65d1928dce260

Weiterführende Literatur

Empfehlungen zum selben Thema automatisch vorgeschlagen von bX