dc.rights.license | CC-BY-NC-ND | |
dc.contributor.advisor | Hardman, Lynda | |
dc.contributor.author | Chen, Yangfan | |
dc.date.accessioned | 2025-08-21T00:01:22Z | |
dc.date.available | 2025-08-21T00:01:22Z | |
dc.date.issued | 2025 | |
dc.identifier.uri | https://studenttheses.uu.nl/handle/20.500.12932/49814 | |
dc.description.abstract | Gaze interaction in general-purpose GUIs is limited by its inaccuracy and calibration requirements. Smooth pursuit-based methods offer a calibration-free alternative but often demand modified user interfaces. We present ZoomPursuit, a technique that adapts smooth pursuit for unmodified GUIs by repurposing system-level magnification functionality.
A baseline evaluation (N=16) established that under simulated calibration drift, ZoomPursuit achieved the accuracy of a manual refinement method with the speed of a simpler automated zoom-based technique. However, the study also revealed a conflict between its imperceptible automation and users' sense of control. A follow-up study (N=7) resolved this with a two-stage model integrating automated correction with manual refinement. This design achieved a final pointing error of 0.10° while dynamically calibrating on the fly from an uncalibrated state.
This work details the design evolution of a pursuit-based interaction technique for general GUIs. The final design demonstrates a model that is both technically effective and preserves user agency. | |
dc.description.sponsorship | Utrecht University | |
dc.language.iso | EN | |
dc.subject | A Smooth-pursuit based gaze interaction design with 2 user studies to evaluate the design. | |
dc.title | ZoomPursuit: Smooth Pursuit-based Gaze Interaction for General-Purpose GUIs | |
dc.type.content | Master Thesis | |
dc.rights.accessrights | Open Access | |
dc.subject.keywords | Gaze Interaction; Smooth Pursuit; Human-Computer Interaction; Interaction Design; Eye-tracking | |
dc.subject.courseuu | Human-Computer Interaction | |
dc.thesis.id | 51975 | |