Show simple item record

dc.rights.licenseCC-BY-NC-ND
dc.contributor.advisorPaperno, Denis
dc.contributor.authorKondyurin, Ivan
dc.date.accessioned2022-08-11T00:00:49Z
dc.date.available2022-08-11T00:00:49Z
dc.date.issued2022
dc.identifier.urihttps://studenttheses.uu.nl/handle/20.500.12932/42258
dc.description.abstractAuthorship attribution attempts to establish the author of a particular text. In this work, we examine the capabilities of transformer-based models in the subtype of attribution task referred to as authorship verification, which involves determining whether the texts are created by the same author. A few works have been suggested that applied fine-tuned Transformer models in this field. Such approach is motivated by their excellent performance and adaptability (fine-tuning can be performed on texts of different sizes and genres, and different pre-trained model checkpoints enable switching between languages). However, they are not as transparent as the traditional methods, in which features that quantify the style (stylometric features) are selected to maximize the distance between texts. To tackle this problem, we first implement a model for authorship verification based on BERT architecture and then investigate the way its predictions are made by applying an adapted LIME explainer and proposing an attention-based relevant feature extracting procedure. We then compare the two approaches and analyze their explainability from the causal perspective by input ablation and alteration to verify that they can retrieve the features that have a strong influence on the model predictions. We also describe and classify the extracted features from a linguistic perspective.
dc.description.sponsorshipUtrecht University
dc.language.isoEN
dc.subjectThis project examines capabilities of Transformer models in the task of authorship verification, which involves determining if the texts are created by the same author. We then explore the degree of their explainability by applying two approaches: an adapted LIME explainer and a proposed attention-based relevant feature extracting. We then compare these techniques, analyze their explainability from the causal perspective, and ground them in stylometric theory.
dc.titleExplainability of Transformers for Authorship Attribution
dc.type.contentMaster Thesis
dc.rights.accessrightsOpen Access
dc.subject.keywordsAuthorship attribution; stylometry; transformers; BERT; attention; LIME
dc.subject.courseuuArtificial Intelligence
dc.thesis.id8280


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record