VisRecall++: Analysing and Predicting Visualisation Recallability from Gaze Behaviour DOI Open Access
Yao Wang, Yue Jiang, Zhiming Hu

и другие.

Proceedings of the ACM on Human-Computer Interaction, Год журнала: 2024, Номер 8(ETRA), С. 1 - 18

Опубликована: Май 20, 2024

Question answering has recently been proposed as a promising means to assess the recallability of information visualisations. However, prior works are yet study link between visually encoding visualisation in memory and recall performance. To fill this gap, we propose VisRecall++ -- novel 40-participant dataset that contains gaze data on 200 visualisations 1,000 questions, including identifying title retrieving values. We measured by asking participants questions after they observed for 10 seconds. Our analyses reveal several insights, such saccade amplitude, number fixations, fixation duration significantly differ high low groups. Finally, GazeRecallNet computational method predict from behaviour outperforms state-of-the-art model RecallNet three other baselines task. Taken together, our results shed light assessing inform future work recallability-based optimisation.

Язык: Английский

Best practices in eye tracking research DOI
Benjamin T. Carter, Steven G. Luke

International Journal of Psychophysiology, Год журнала: 2020, Номер 155, С. 49 - 62

Опубликована: Июнь 3, 2020

Язык: Английский

Процитировано

415

RETRACTED ARTICLE: Eye tracking: empirical foundations for a minimal reporting guideline DOI Creative Commons
Kenneth Holmqvist,

Saga Lee Örbom,

Ignace T. C. Hooge

и другие.

Behavior Research Methods, Год журнала: 2022, Номер 55(1), С. 364 - 416

Опубликована: Апрель 6, 2022

Abstract In this paper, we present a review of how the various aspects any study using an eye tracker (such as instrument, methodology, environment, participant, etc.) affect quality recorded eye-tracking data and obtained eye-movement gaze measures. We take to represent empirical foundation for reporting guidelines involving tracker. compare five existing database 207 published studies. find that vary substantially do not match with actual practices. end by deriving minimal, flexible guideline based on research (Section “An empirically minimal guideline”).

Язык: Английский

Процитировано

139

Eye Tracking in Virtual Reality: a Broad Review of Applications and Challenges DOI Creative Commons
Isayas Berhe Adhanom, Paul R. MacNeilage, Eelke Folmer

и другие.

Virtual Reality, Год журнала: 2023, Номер 27(2), С. 1481 - 1505

Опубликована: Янв. 18, 2023

Eye tracking is becoming increasingly available in head-mounted virtual reality displays with various headsets integrated eye trackers already commercially available. The applications of are highly diversified and span multiple disciplines. As a result, the number peer-reviewed publications that study has surged recent years. We performed broad review to comprehensively search academic literature databases aim assessing extent published research dealing reality, highlighting challenges, limitations areas for future research.

Язык: Английский

Процитировано

106

The fundamentals of eye tracking part 4: Tools for conducting an eye tracking study DOI Creative Commons
Diederick C. Niehorster, Marcus Nyström, Roy S. Hessels

и другие.

Behavior Research Methods, Год журнала: 2025, Номер 57(1)

Опубликована: Янв. 6, 2025

Abstract Researchers using eye tracking are heavily dependent on software and hardware tools to perform their studies, from recording data visualizing it, processing analyzing it. This article provides an overview of available for research trackers discusses considerations make when choosing which adopt one’s study.

Язык: Английский

Процитировано

4

Event-Based Near-Eye Gaze Tracking Beyond 10,000 Hz DOI
Anastasios N. Angelopoulos, Julien Martel, Amit Kohli

и другие.

IEEE Transactions on Visualization and Computer Graphics, Год журнала: 2021, Номер 27(5), С. 2577 - 2586

Опубликована: Март 29, 2021

The cameras in modern gaze-tracking systems suffer from fundamental bandwidth and power limitations, constraining data acquisition speed to 300 Hz realistically. This obstructs the use of mobile eye trackers perform, e.g., low latency predictive rendering, or study quick subtle motions like microsaccades using head-mounted devices wild. Here, we propose a hybrid frame-event-based near-eye gaze tracking system offering update rates beyond 10,000 with an accuracy that matches high-end desktop-mounted commercial when evaluated same conditions. Our system, previewed Figure 1, builds on emerging event simultaneously acquire regularly sampled frames adaptively events. We develop online 2D pupil fitting method updates parametric model every one few Moreover, polynomial regressor for estimating point real time. Using first event-based dataset, demonstrate our achieves accuracies 0.45°-1.75° fields view 45° 98°. With this technology, hope enable new generation ultra-low-latency gaze-contingent rendering display techniques virtual augmented reality.

Язык: Английский

Процитировано

74

Eye-tracking research on sensory and consumer science: A review, pitfalls and future directions DOI
Kosuke Motoki, Toshiki Saito, Takuya Onuma

и другие.

Food Research International, Год журнала: 2021, Номер 145, С. 110389 - 110389

Опубликована: Май 6, 2021

Язык: Английский

Процитировано

73

GazeBase, a large-scale, multi-stimulus, longitudinal eye movement dataset DOI Creative Commons
Henry Griffith, Dillon Lohr,

Evgeny Abdulin

и другие.

Scientific Data, Год журнала: 2021, Номер 8(1)

Опубликована: Июль 16, 2021

This manuscript presents GazeBase, a large-scale longitudinal dataset containing 12,334 monocular eye-movement recordings captured from 322 college-aged participants. Participants completed battery of seven tasks in two contiguous sessions during each round recording, including - (1) fixation task, (2) horizontal saccade (3) random oblique (4) reading (5/6) free viewing cinematic video and (7) gaze-driven gaming task. Nine rounds recording were conducted over 37 month period, with participants subsequent recruited exclusively prior rounds. All data was collected using an EyeLink 1000 eye tracker at 1,000 Hz sampling rate, calibration validation protocol performed before task to ensure quality. Due its large number nature, GazeBase is well suited for exploring research hypotheses movement biometrics, along other applications applying machine learning signal analysis. Classification labels produced by the instrument's real-time parser are provided subset pupil area.

Язык: Английский

Процитировано

60

Eye tracking in virtual reality: Vive pro eye spatial accuracy, precision, and calibration reliability DOI Creative Commons

Immo Schuetz,

Katja Fiehler

Journal of Eye Movement Research, Год журнала: 2022, Номер 15(3)

Опубликована: Сен. 7, 2022

A growing number of virtual reality devices now include eye tracking technology, which can facilitate oculomotor and cognitive research in VR enable use cases like foveated rendering. These applications require different performance, often measured as spatial accuracy precision. While manufacturers report data quality estimates for their devices, these typically represent ideal performance may not reflect real-world quality. Additionally, it is unclear how precision change across sessions within the same participant or between influenced by vision correction. Here, we Vive Pro Eye built-in tracker a range 30 visual degrees horizontally vertically. Participants completed ten measurement over multiple days, allowing to evaluate calibration reliability. Accuracy were highest central gaze decreased with greater eccentricity both axes. Calibration was successful all participants, including those wearing contacts glasses, but glasses yielded significantly lower performance. We further found differences (but precision) two headsets, estimated participants' inter-pupillary distance. Our metrics suggest high reliability serve baseline expected experiments.

Язык: Английский

Процитировано

51

Evaluating the Tobii Pro Glasses 2 and 3 in static and dynamic conditions DOI Creative Commons
V. Onkhar, Dimitra Dodou, Joost de Winter

и другие.

Behavior Research Methods, Год журнала: 2023, Номер unknown

Опубликована: Авг. 7, 2023

Abstract Over the past few decades, there have been significant developments in eye-tracking technology, particularly domain of mobile, head-mounted devices. Nevertheless, questions remain regarding accuracy these eye-trackers during static and dynamic tasks. In light this, we evaluated performance two widely used devices: Tobii Pro Glasses 2 3. A total 36 participants engaged tasks under three dynamicity conditions. “seated with a chinrest” trial, only eyes could be moved; without both head were free to move; walking walked along straight path. During seated trials, participants’ gaze was directed towards dots on wall by means audio instructions, whereas maintained their bullseye while it. Eye-tracker determined using computer vision techniques identify target within scene camera image. The findings showed that 3 outperformed terms trials. Moreover, results suggest employing chinrest case is counterproductive, as it necessitates larger eye eccentricities for fixation, thereby compromising compared not chinrest, which allows movement. Lastly, found who reported higher workload demonstrated poorer accuracy. current may useful design experiments involve eye-trackers.

Язык: Английский

Процитировано

30

Webcam eye tracking close to laboratory standards: Comparing a new webcam-based system and the EyeLink 1000 DOI Creative Commons

Tobiasz Kaduk,

Caspar Goeke,

Holger Finger

и другие.

Behavior Research Methods, Год журнала: 2023, Номер 56(5), С. 5002 - 5022

Опубликована: Окт. 11, 2023

Abstract This paper aims to compare a new webcam-based eye-tracking system, integrated into the Labvanced platform for online experiments, “gold standard” lab-based eye tracker (EyeLink 1000 - SR Research). Specifically, we simultaneously recorded data with both trackers in five different tasks, analyzing their real-time performance. These tasks were subset of standardized test battery trackers, including Large Grid task, Smooth Pursuit movements, viewing natural images, and two Head Movements (roll, yaw). The results show that system achieved an overall accuracy 1.4°, precision 1.1° (standard deviation (SD) across subjects), error about 0.5° larger than EyeLink system. Interestingly, (1.3°) (0.9°) slightly better centrally presented targets, region interest many psychophysical experiments. Remarkably, correlation raw gaze samples between was at 90% task 80% Free View Pursuit. Overall, these put performance roughly on par mobile devices (Ehinger et al. PeerJ , 7 e7086, 2019; Tonsen al., 2020) demonstrate substantial improvement compared existing webcam solutions (Papoutsaki 2017).

Язык: Английский

Процитировано

27