default search action
ETRA 2021 - Short Papers
- Andreas Bulling, Anke Huckauf, Hans Gellersen, Daniel Weiskopf, Mihai Bace, Teresa Hirzle, Florian Alt, Thies Pfeiffer, Roman Bednarik, Krzysztof Krejtz, Tanja Blascheck, Michael Burch, Peter Kiefer, Michael D. Dodd, Bonita Sharif:
2021 Symposium on Eye Tracking Research and Applications, ETRA 2020, Virtual Event, Germany, May 25-27, 2021, Short Papers. ACM 2021, ISBN 978-1-4503-8345-5
Applications
- Ryota Nishizono, Naoki Saijo, Makio Kashino:
Synchronization of Spontaneous Eyeblink during Formula Car Driving. 1:1-1:6 - Brendan David-John, Candace E. Peacock, Ting Zhang, T. Scott Murdison, Hrvoje Benko, Tanya R. Jonker:
Towards gaze-based prediction of the intent to interact in virtual reality. 2:1-2:7 - Takashi Hirata, Yutaka Hirata:
Pupillary response reflects vocabulary comprehension. 3:1-3:5 - Kaitlyn M. Roose, Elizabeth S. Veinott:
Understanding Game Roles and Strategy Using a Mixed Methods Approach. 4:1-4:5 - Margarita Vinnikov, Kian Motahari, Louis I. Hamilton, Burcak Ozludil Altin:
Understanding Urban Devotion through the Eyes of an Observer. 5:1-5:6
Methods
- Thomas Berger, Michael Raschke:
Repetition effects in task-driven eye movement analyses after longer time-spans. 6:1-6:6 - Kento Seida, Kentaro Takemura:
Eye Gaze Estimation using Imperceptible Marker Presented on High-Speed Display. 7:1-7:5 - Aayush Kumar Chaudhary, Prashnna K. Gyawali, Linwei Wang, Jeff B. Pelz:
Semi-Supervised Learning for Eye Image Segmentation. 8:1-8:7 - Nathaniel Barbara, Tracey A. Camilleri, Kenneth P. Camilleri:
Modelling of Blink-Related Eyelid-Induced Shunting on the Electrooculogram. 9:1-9:6 - Martin H. U. Prinzler, Christoph Schröder, Sahar Mahdie Klim Al Zaidawi, Gabriel Zachmann, Sebastian Maneth:
Visualizing Prediction Correctness of Eye Tracking Classifiers. 10:1-10:7 - Kara J. Emery, Marina Zannoli, James Warren, Lei Xiao, Sachin S. Talathi:
OpenNEEDS: A Dataset of Gaze, Head, Hand, and Scene Signals During Exploration in Open-Ended VR Environments. 11:1-11:7 - Aayush Kumar Chaudhary, Jeff B. Pelz:
Enhancing the precision of remote eye-tracking using iris velocity estimation. 12:1-12:7
User Interfaces and Interaction
- Zixiong Su, Xinlei Zhang, Naoki Kimura, Jun Rekimoto:
Gaze+Lip: Rapid, Precise and Expressive Interactions Combining Gaze Input and Silent Speech Commands for Hands-free Smart TV Control. 13:1-13:6 - Omair Shahzad Bhatti, Michael Barz, Daniel Sonntag:
EyeLogin - Calibration-free Authentication Method for Public Displays Using Eye Gaze. 14:1-14:7 - Aunnoy K. Mutasim, Anil Ufuk Batmaz, Wolfgang Stuerzlinger:
Pinch, Click, or Dwell: Comparing Different Selection Techniques for Eye-Gaze-Based Pointing in Virtual Reality. 15:1-15:7
Eye Movements and Attention
- Wolfgang Fuhl, Enkelejda Kasneci:
A Multimodal Eye Movement Dataset and a Multimodal Eye Movement Segmentation Analysis. 16:1-16:7 - Thomas C. Kübler, Wolfgang Fuhl, Elena Wagner, Enkelejda Kasneci:
55 Rides: attention annotated head and gaze data during naturalistic driving. 17:1-17:8 - Mostafa Elshamy, Peter Khooshabeh:
Fixation: A universal framework for experimental eye movement research✱. 18:1-18:5 - Takahiro Ueno, Minoru Nakayama:
Estimation of Visual Attention using Microsaccades in response to Vibrations in the Peripheral Field of Vision. 19:1-19:6 - Lorenzo Diana, Patrick Pilastro, Edoardo N. Aiello, Aleksandra K. Eberhard-Moscicka, René M. Müri, Nadia Bolognini:
Saccades, attentional orienting and disengagement: the effects of anodal tDCS over right posterior parietal cortex (PPC) and frontal eye field (FEF). 20:1-20:7
EYESEC: Workshop on Eye-Gaze for Security Applications
- Leon Bernard, Sagar Raina, Blair Taylor, Siddharth Kaza:
Minimizing Cognitive Load in Cyber Learning Materials - An Eye Tracking Study. 21:1-21:6 - Christina P. Katsini, George E. Raptis, Andrew Jian-lan Cen, Nalin Asanka Gamagedara Arachchilage, Lennart E. Nacke:
Eye-GUAna: Higher Gaze-Based Entropy and Increased Password Space in Graphical User Authentication Through Gamification. 22:1-22:7 - Yudai Niitsu, Minoru Nakayama:
Effects of measurement time and presentation size conditions on biometric identification using eye movements. 23:1-23:5
COGAIN Symposium 2021
- Tanya Bafna, Per Bækgaard, John Paulin Hansen:
EyeTell: Tablet-based Calibration-free Eye-typing using Smooth-pursuit movements. 24:1-24:6 - Radiah Rivu, Yasmeen Abdrabou, Yomna Abdelrahman, Ken Pfeuffer, Dagmar Kern, Cornelia Neuert, Daniel Buschek, Florian Alt:
Did you Understand this?: Leveraging Gaze Behavior to Assess Questionnaire Comprehension. 25:1-25:7 - Ken Pfeuffer, Jason Alexander, Hans Gellersen:
Multi-user Gaze-based Interaction Techniques on Collaborative Touchscreens. 26:1-26:7 - Mateusz Dubiel, Minoru Nakayama, Xin Wang:
Combining Oculo-motor Indices to Measure Cognitive Load of Synthetic Speech in Noisy Listening Conditions. 27:1-27:6 - Ard Kastrati, Martyna Plomecka, Roger Wattenhofer, Nicolas Langer:
Using Deep Learning to Classify Saccade Direction from Brain Activity. 28:1-28:6 - Tomomi Okano, Minoru Nakayama:
Feasibility of evaluating temporal changes in cognitive load factors using ocular features. 29:1-29:6
EMIP: Eye Movements in Programming
- Salwa D. Aljehane, Bonita Sharif, Jonathan I. Maletic:
Determining Differences in Reading Behavior Between Experts and Novices by Investigating Eye Movement on Source Code Constructs During a Bug Fixing Task. 30:1-30:6 - Teresa Busjahn, Sascha Tamm:
A Deeper Analysis of AOI Coverage in Code Reading. 31:1-31:7 - Hiroto Harada, Minoru Nakayama:
Estimation of reading ability of program codes using features of eye movements. 32:1-32:5 - Ian R. McChesney, Raymond R. Bond:
Eye Tracking Analysis of Code Layout, Crowding and Dyslexia - An Open Data Set. 33:1-33:6 - Naser Al Madi, Drew T. Guarnera, Bonita Sharif, Jonathan I. Maletic:
EMIP Toolkit: A Python Library for Customized Post-processing of the Eye Movements in Programming Dataset. 34:1-34:6 - Jonas Mucke, Marc Schwarzkopf, Janet Siegmund:
REyeker: Remote Eye Tracker. 35:1-35:5
PLEY: 2nd Eye Tracking in Games and Play Workshop
- Wenyi Lu, Hao He, Alex C. Urban, Joe Griffin:
What the Eyes Can Tell: Analyzing Visual Attention with an Educational Video Game. 36:1-36:7 - George E. Raptis, Christina P. Katsini:
Analyzing Scanpaths From A Field Dependence-Independence Perspective When Playing A Visual Search Game. 37:1-37:7
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.