Show simple item record

dc.rights.licenseCC-BY-NC-ND
dc.contributor.advisorAydemir, Başak
dc.contributor.authorChitlangia, Akshay
dc.date.accessioned2025-08-15T00:07:32Z
dc.date.available2025-08-15T00:07:32Z
dc.date.issued2025
dc.identifier.urihttps://studenttheses.uu.nl/handle/20.500.12932/49759
dc.description.abstractIn software development, generating user stories from requirements elicitation interviews is a critical yet time-consuming and subjective task. Traditional methods often rely heavily on human interpretation, which can introduce bias and limit efficiency. This thesis explores the potential of Fine-tuned Large Language Models (LLMs) to automate and enhance this process. To address these challenges, LLMs were fine-tuned using supervised learning on real interview transcripts and further optimized through prompt engineering to generate well-structured user stories. The approach aimed to reduce manual effort while maintaining the accuracy and completeness of the generated user stories. A comprehensive evaluation was conducted using both quantitative and qualitative methods. Quantitative metrics included time taken for generation, carbon emissions estimation, readability scores, semantic similarity, and lexical diversity. Additionally, domain experts assessed the accuracy, completeness, and relevance of the generated user stories through structured questionnaires. Findings show that fine-tuned LLMs can produce user stories that are comparable in quality to those created by human experts, while significantly improving efficiency and consistency. The results demonstrate the practical value of LLMs in supporting requirements engineering tasks and reducing the cognitive load on analysts.
dc.description.sponsorshipUtrecht University
dc.language.isoEN
dc.subjectThe research fine-tunes state-of-the-art LLMs using real interview transcripts to generate user stories in the format “As a [user], I want [goal], so that [benefit]”. It evaluates the output through both quantitative metrics ( readability, semantic similarity, carbon emissions, etc.) and qualitative analysis by industry experts. The findings suggest that fine-tuned LLMs can produce good quality user stories comparable to human-generated ones, improving efficiency and reducing analyst workload
dc.titleAutomated User Story Generation in Requirements Elicitation using Fine-Tuned Large Language Models
dc.type.contentMaster Thesis
dc.rights.accessrightsOpen Access
dc.subject.courseuuBusiness Informatics
dc.thesis.id51726


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record