Show simple item record

dc.rights.licenseCC-BY-NC-ND
dc.contributor.advisorHardman, Lynda
dc.contributor.authorChen, Yangfan
dc.date.accessioned2025-08-21T00:01:22Z
dc.date.available2025-08-21T00:01:22Z
dc.date.issued2025
dc.identifier.urihttps://studenttheses.uu.nl/handle/20.500.12932/49814
dc.description.abstractGaze interaction in general-purpose GUIs is limited by its inaccuracy and calibration requirements. Smooth pursuit-based methods offer a calibration-free alternative but often demand modified user interfaces. We present ZoomPursuit, a technique that adapts smooth pursuit for unmodified GUIs by repurposing system-level magnification functionality. A baseline evaluation (N=16) established that under simulated calibration drift, ZoomPursuit achieved the accuracy of a manual refinement method with the speed of a simpler automated zoom-based technique. However, the study also revealed a conflict between its imperceptible automation and users' sense of control. A follow-up study (N=7) resolved this with a two-stage model integrating automated correction with manual refinement. This design achieved a final pointing error of 0.10° while dynamically calibrating on the fly from an uncalibrated state. This work details the design evolution of a pursuit-based interaction technique for general GUIs. The final design demonstrates a model that is both technically effective and preserves user agency.
dc.description.sponsorshipUtrecht University
dc.language.isoEN
dc.subjectA Smooth-pursuit based gaze interaction design with 2 user studies to evaluate the design.
dc.titleZoomPursuit: Smooth Pursuit-based Gaze Interaction for General-Purpose GUIs
dc.type.contentMaster Thesis
dc.rights.accessrightsOpen Access
dc.subject.keywordsGaze Interaction; Smooth Pursuit; Human-Computer Interaction; Interaction Design; Eye-tracking
dc.subject.courseuuHuman-Computer Interaction
dc.thesis.id51975


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record