


default search action
ETRA 2021: Stuttgart, Germany - Adjunct Volume
- Andreas Bulling, Anke Huckauf, Hans Gellersen, Daniel Weiskopf, Mihai Bace, Teresa Hirzle, Florian Alt, Thies Pfeiffer, Roman Bednarik, Krzysztof Krejtz, Tanja Blascheck, Michael Burch, Peter Kiefer
, Michael D. Dodd, Bonita Sharif:
ETRA '21 Adjunct: 2021 Symposium on Eye Tracking Research and Applications, Virtual Event, Germany, May 25-27, 2021. ACM 2021, ISBN 978-1-4503-8357-8 - Ryan Lewien:
GazeHelp: Exploring Practical Gaze-assisted Interactions for Graphic Design Tools. 1:1-1:4 - Soon-Gyo Jung, Joni Salminen, Bernard J. Jansen:
Implementing Eye-Tracking for Persona Analytics. 2:1-2:4 - Michael Barz, Sebastian Kapp
, Jochen Kuhn, Daniel Sonntag:
Automatic Recognition and Augmentation of Attended Objects in Real-time using Eye Tracking and a Head-mounted Display. 3:1-3:4 - Yaxiong Lei:
Eye Tracking Calibration on Mobile Devices. 4:1-4:4 - Christos Gkoumas, Andria Shimi
:
The influence of clutter on search-based learning, long-term memory, and memory-guided attention in real-world scenes: an eye-movement research protocol. 5:1-5:3 - Anna Mazurowska:
Climate change overlooked. The role of attitudes and mood regulation in visual attention to global warming. 6:1-6:3 - Katarzyna Wisiecka:
Gaze and Heart Rate Synchronization in Computer-Mediated Collaboration. 7:1-7:3 - Agostino Gibaldi
, Vasha DuTell, Martin S. Banks:
Solving Parallax Error for 3D Eye Tracking. 8:1-8:4 - Vasha DuTell, Agostino Gibaldi
, Giulia Focarelli, Bruno A. Olshausen, Martin S. Banks:
Integrating High Fidelity Eye, Head and World Tracking in a Wearable Device. 9:1-9:4 - Avi M. Aizenman
, Dennis M. Levi:
Fixational stability as a measure for the recovery of visual function in amblyopia. 10:1-10:4 - Markus D. Solbach, John K. Tsotsos
:
Tracking Active Observers in 3D Visuo-Cognitive Tasks. 11:1-11:3 - Daniel Müller, David Mann
:
Algorithmic gaze classification for mobile eye-tracking. 12:1-12:4 - Scott A. Stone
, Quinn A. Boser, T. Riley Dawson, Albert H. Vette, Jacqueline S. Hebert
, Patrick M. Pilarski
, Craig S. Chapman:
Sub-centimeter 3D gaze vector accuracy on real-world tasks: an investigation of eye and motion capture calibration routines. 13:1-13:4 - Bharath Shankar, Christian Sinnott, Kamran Binaee, Mark D. Lescroart
, Paul R. MacNeilage:
Ergonomic Design Development of the Visual Experience Database Headset. 14:1-14:4 - Sanjana Ramanujam, Christian Sinnott, Bharath Shankar, Savannah Jo Halow, Brian Szekely, Paul R. MacNeilage, Kamran Binaee:
VEDBViz: The Visual Experience Database Visualization and Interaction Tool. 15:1-15:4 - Kassia Love, Anca Velisar, Natela M. Shanidze
:
Eye, Robot: Calibration Challenges and Potential Solutions for Wearable Eye Tracking in Individuals with Eccentric Fixation. 16:1-16:3 - Haylie L. Miller, Ian Raphael Zurutuza, Nicholas Fears
, Suleyman Polat, Rodney Nielsen:
Post-processing integration and semi-automated analysis of eye-tracking and motion-capture data obtained in immersive virtual reality environments to measure visuomotor integration. 17:1-17:4 - Kamran Binaee, Christian Sinnott, Kaylie Jacleen Capurro, Paul R. MacNeilage, Mark D. Lescroart
:
Pupil Tracking Under Direct Sunlight. 18:1-18:4 - Arnab Biswas, Kamran Binaee, Kaylie Jacleen Capurro, Mark D. Lescroart
:
Characterizing the Performance of Deep Neural Networks for Eye-Tracking. 19:1-19:4 - Anca Velisar, Natela M. Shanidze
:
Noise in the Machine: Sources of Physical and Computation Error in Eye Tracking with Pupil Core Wearable Eye Tracker: Wearable Eye Tracker Noise in Natural Motion Experiments. 20:1-20:3 - Fiona Bríd Mulvey
, Marek Mikitovic, Mateusz Sadowski, Baosheng James Hou, Nils David Rasamoel, John Paulin Hansen
, Per Bækgaard
:
Gaze Interactive and Attention Aware Low Vision Aids as Future Smart Glasses. 21:1-21:4

manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.