Show simple item record

dc.rights.licenseCC-BY-NC-ND
dc.contributorDong Nguyen, Elize Herrewijnen, Yupei Du
dc.contributor.advisorNguyen, Dong
dc.contributor.authorPieke, Michael
dc.date.accessioned2023-07-22T00:01:48Z
dc.date.available2023-07-22T00:01:48Z
dc.date.issued2023
dc.identifier.urihttps://studenttheses.uu.nl/handle/20.500.12932/44264
dc.description.abstractPre-trained transformers are highly effective across numerous Natural Language Processing (NLP) tasks, yet their ability to generalise to new domains remains a concern due to their tendency to rely on spurious correlations. Consequently, this thesis investigates the impact of token-level human supervision to enhance BERT's generalisation capabilities. Although the benefits of token-level insights have been shown to improve the performance of these models, few studies have examined the effect of these insights to improve generalisation. Consequently, this work explores the potential of human supervision to guide BERT's attention mechanism towards salient features, thereby improving generalisation across domains. Results from experiments in both binary and multi-label classification scenarios demonstrate not only substantial gains in out-of-distribution (OOD) performance in few-shot contexts, but also a closer alignment between the model's attention scores and salient features identified by human annotators. By emphasising the role of human insight in transformer models, this thesis contributes to the ongoing discourse on enhancing performance and explainability in NLP applications.
dc.description.sponsorshipUtrecht University
dc.language.isoEN
dc.subjectThis thesis explores the use of token-level human supervision to increase the generalisability of BERT models
dc.titleBERT, but Better: Improving Robustness using Human Insights
dc.type.contentMaster Thesis
dc.rights.accessrightsOpen Access
dc.subject.courseuuArtificial Intelligence
dc.thesis.id19856


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record