Show simple item record

dc.rights.licenseCC-BY-NC-ND
dc.contributor.advisorPaperno, Denis
dc.contributor.authorDona, Loïs
dc.date.accessioned2023-08-08T00:01:24Z
dc.date.available2023-08-08T00:01:24Z
dc.date.issued2023
dc.identifier.urihttps://studenttheses.uu.nl/handle/20.500.12932/44523
dc.description.abstractThis work builds upon an analogy between human and artificial reasoning. Large Transformer based language models have achieved state-of-the-art in many tasks, and recently have even been able to do this without requiring task specific fine-tuning, by deploying either in-context or zero-shot learning. However, reasoning remains to be a difficult task for these models, especially in a zero-shot setting. On the contrary, humans are good at reasoning without being explicitly trained for it like models are, hinting that properties of human reasoning might be of help to boost the performance of models. We explore this intuition in two ways: (1) using human-like linguistic input for fine-tuning and (2) prompting models to ``imagine", a technique that has shown to help humans reason better. Our results show that our approach was fruitful for reasoning about fantastical scenarios, which is in line with previous research on humans, confirming that making an analogy between human and artificial reasoning can be helpful. This research opens many doors for future research on zero-shot reasoning, also using smaller models, which is a desirable development towards human-like general intelligence.
dc.description.sponsorshipUtrecht University
dc.language.isoEN
dc.subjectIn this work I make an analogy between human reasoning and zero-shot reasoning in GPT-2 and GPT-3. This is done in two ways: (1) using human-like input to fine-tune GPT models and (2) using an adapted version of chain-of-thought prompting, inspired by psychological research on children's reasoning. Results show that these methods, especially the first, can be a promising avenue for using smaller language models in the context of zero-shot reasoning.
dc.titleApplying Characteristics of Human Reasoning to Zero-shot Reasoning in GPT Models
dc.type.contentMaster Thesis
dc.rights.accessrightsOpen Access
dc.subject.courseuuArtificial Intelligence
dc.thesis.id21271


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record