default search action
ETRA 2021: Stuttgart, Germany - Adjunct Volume
- Andreas Bulling, Anke Huckauf, Hans Gellersen, Daniel Weiskopf, Mihai Bace, Teresa Hirzle, Florian Alt, Thies Pfeiffer, Roman Bednarik, Krzysztof Krejtz, Tanja Blascheck, Michael Burch, Peter Kiefer, Michael D. Dodd, Bonita Sharif:
ETRA '21 Adjunct: 2021 Symposium on Eye Tracking Research and Applications, Virtual Event, Germany, May 25-27, 2021. ACM 2021, ISBN 978-1-4503-8357-8 - Ryan Lewien:
GazeHelp: Exploring Practical Gaze-assisted Interactions for Graphic Design Tools. 1:1-1:4 - Soon-Gyo Jung, Joni Salminen, Bernard J. Jansen:
Implementing Eye-Tracking for Persona Analytics. 2:1-2:4 - Michael Barz, Sebastian Kapp, Jochen Kuhn, Daniel Sonntag:
Automatic Recognition and Augmentation of Attended Objects in Real-time using Eye Tracking and a Head-mounted Display. 3:1-3:4 - Yaxiong Lei:
Eye Tracking Calibration on Mobile Devices. 4:1-4:4 - Christos Gkoumas, Andria Shimi:
The influence of clutter on search-based learning, long-term memory, and memory-guided attention in real-world scenes: an eye-movement research protocol. 5:1-5:3 - Anna Mazurowska:
Climate change overlooked. The role of attitudes and mood regulation in visual attention to global warming. 6:1-6:3 - Katarzyna Wisiecka:
Gaze and Heart Rate Synchronization in Computer-Mediated Collaboration. 7:1-7:3 - Agostino Gibaldi, Vasha DuTell, Martin S. Banks:
Solving Parallax Error for 3D Eye Tracking. 8:1-8:4 - Vasha DuTell, Agostino Gibaldi, Giulia Focarelli, Bruno A. Olshausen, Martin S. Banks:
Integrating High Fidelity Eye, Head and World Tracking in a Wearable Device. 9:1-9:4 - Avi M. Aizenman, Dennis M. Levi:
Fixational stability as a measure for the recovery of visual function in amblyopia. 10:1-10:4 - Markus D. Solbach, John K. Tsotsos:
Tracking Active Observers in 3D Visuo-Cognitive Tasks. 11:1-11:3 - Daniel Müller, David Mann:
Algorithmic gaze classification for mobile eye-tracking. 12:1-12:4 - Scott A. Stone, Quinn A. Boser, T. Riley Dawson, Albert H. Vette, Jacqueline S. Hebert, Patrick M. Pilarski, Craig S. Chapman:
Sub-centimeter 3D gaze vector accuracy on real-world tasks: an investigation of eye and motion capture calibration routines. 13:1-13:4 - Bharath Shankar, Christian Sinnott, Kamran Binaee, Mark D. Lescroart, Paul R. MacNeilage:
Ergonomic Design Development of the Visual Experience Database Headset. 14:1-14:4 - Sanjana Ramanujam, Christian Sinnott, Bharath Shankar, Savannah Jo Halow, Brian Szekely, Paul R. MacNeilage, Kamran Binaee:
VEDBViz: The Visual Experience Database Visualization and Interaction Tool. 15:1-15:4 - Kassia Love, Anca Velisar, Natela M. Shanidze:
Eye, Robot: Calibration Challenges and Potential Solutions for Wearable Eye Tracking in Individuals with Eccentric Fixation. 16:1-16:3 - Haylie L. Miller, Ian Raphael Zurutuza, Nicholas Fears, Suleyman Polat, Rodney Nielsen:
Post-processing integration and semi-automated analysis of eye-tracking and motion-capture data obtained in immersive virtual reality environments to measure visuomotor integration. 17:1-17:4 - Kamran Binaee, Christian Sinnott, Kaylie Jacleen Capurro, Paul R. MacNeilage, Mark D. Lescroart:
Pupil Tracking Under Direct Sunlight. 18:1-18:4 - Arnab Biswas, Kamran Binaee, Kaylie Jacleen Capurro, Mark D. Lescroart:
Characterizing the Performance of Deep Neural Networks for Eye-Tracking. 19:1-19:4 - Anca Velisar, Natela M. Shanidze:
Noise in the Machine: Sources of Physical and Computation Error in Eye Tracking with Pupil Core Wearable Eye Tracker: Wearable Eye Tracker Noise in Natural Motion Experiments. 20:1-20:3 - Fiona Bríd Mulvey, Marek Mikitovic, Mateusz Sadowski, Baosheng James Hou, Nils David Rasamoel, John Paulin Hansen, Per Bækgaard:
Gaze Interactive and Attention Aware Low Vision Aids as Future Smart Glasses. 21:1-21:4
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.