Show simple item record

dc.rights.licenseCC-BY-NC-ND
dc.contributor.advisorPaperno, Denis
dc.contributor.authorGrashoff, Daan
dc.date.accessioned2022-03-01T00:00:29Z
dc.date.available2022-03-01T00:00:29Z
dc.date.issued2022
dc.identifier.urihttps://studenttheses.uu.nl/handle/20.500.12932/533
dc.description.abstractIn this thesis, we studied whether self-attention networks can learn compositional seman- tics using an arithmetic language. The goal of language aims to evaluate the meaning of nested expressions. We find that self-attention networks can learn to evaluate these nested expres- sions by taking shortcuts on less complex expressions or utilizing deeper layers on complex expressions when the nested depth grows. The complexity is in whether expressions are left- (easy) or right-branching (hard) and whether, in the case of right-branching expressions, plus (easy) or minus (complex) operators are used. We find that increasing the number of heads does not always help with more complex expressions, whereas the number of layers does always help to generalize to deeper expressions. Finally, to help with the understanding of what the self-attention networks are doing, we analyzed the attention scores and found exciting patterns such as the numbers attending to the preceding operators and nested sub-expressions attend- ing to preceding operators. These patterns may explain why in less complex expressions, the self-attention networks take shortcuts, but in more complex expressions, this is not possible by the way the self-attention networks try to solve them.
dc.description.sponsorshipUtrecht University
dc.language.isoEN
dc.subjectIn this work, we studied whether self-attention networks can learn compositionality using an arithmetic language, a simple language that allowed us to construct precise compositional structures and test the networks on these.
dc.titleOn how transformers learn to understand and evaluate nested arithmetic expressions
dc.type.contentMaster Thesis
dc.rights.accessrightsOpen Access
dc.subject.keywordsTransformer; interpretability of neural networks;compositionality;embedding depth
dc.subject.courseuuArtificial Intelligence
dc.thesis.id1870


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record