| dc.rights.license | CC-BY-NC-ND |  | 
| dc.contributor.advisor | Schoot, Rens van de |  | 
| dc.contributor.author | Spedener, Lesley |  | 
| dc.date.accessioned | 2023-09-19T00:00:44Z |  | 
| dc.date.available | 2023-09-19T00:00:44Z |  | 
| dc.date.issued | 2023 |  | 
| dc.identifier.uri | https://studenttheses.uu.nl/handle/20.500.12932/45187 |  | 
| dc.description.abstract | Simulation-based Active Learning (AL) studies have demonstrated the potential of machine
learning methods in reducing manual screening workload in systematic literature reviews. The 
second most used performance metric in this field is Work Saved Over Sampling (WSS), which 
aims to measure the reduction in screening effort. A drawback of the WSS metric, however, is its
sensitivity to dataset class imbalance, which leads to biased performance comparisons across 
datasets. In this light, two main features were added to the state-of-the-art and open-source
simulation software ASReview, which offers a unique infrastructure for testing different AL model 
and feature extractor combinations across datasets. First, the confusion matrix was implemented 
into the ASReview software, which was subsequently used to implement the True Negative Rate 
(TNR), shown to be equal to the normalized WSS (Kusa et al., 2023). These advancements, 
previously absent in the software, represent a step towards achieving a more comprehensive 
understanding of AL performance in SLR tasks. Specifically, the adjustment for class imbalance
facilitates further study of data characteristics related to model performance beyond class 
imbalance. This enhanced understanding enables researchers and practitioners to make more 
informed decisions in selecting and fine tuning AL models, ultimately leading to more efficient 
screening in practice. |  | 
| dc.description.sponsorship | Utrecht University |  | 
| dc.language.iso | EN |  | 
| dc.subject | Implementing and evaluating new performance metrics for AI-assisted systematic reviewing |  | 
| dc.title | Towards Performance Comparability:
An Implementation of New Metrics into the ASReview Active Learning Screening
Prioritization Software for Systematic Literature Reviews |  | 
| dc.type.content | Master Thesis |  | 
| dc.rights.accessrights | Open Access |  | 
| dc.subject.keywords | systematic literature review (SLR), active learning (AL), evaluation, comparability,
Work Saved Over Sampling (WSS), True Negative Rate (TNR), specificity |  | 
| dc.subject.courseuu | Applied Data Science |  | 
| dc.thesis.id | 24501 |  |