Abstract
The last twenty years have seen the development of gaze-controlled computer interfaces for augmentative communication and other assistive technology applications. In many applications, the user needs to look at symbols on a virtual on-screen keyboard and maintain gaze to make a selection. Executive control is essential to learning to use gaze-control, affecting the uptake of the technology. Specifically, the user of a gaze-controlled interface must suppress looking for its own sake, the so-called “Midas touch” problem. In a pre-registered study (https://osf.io/2mak4), we tested whether gaze-typing performance depends on executive control and whether learning-dependent plasticity leads to improved executive control as measured using the antisaccade task. Forty-two university students were recruited as participants. After five 30-min training sessions, we found shorter antisaccade latencies in a gaze-control compared to a mouse-control group, and similar error-rates. Subjective workload ratings were also similar across groups, indicating the task in both groups was matched for difficulty. These findings suggest that executive control contributes to gaze-typing performance and leads to learning-induced plasticity.
Original language | English |
---|---|
Article number | 106831 |
Number of pages | 10 |
Journal | Computers in Human Behavior |
Volume | 122 |
Early online date | 16 Apr 2021 |
DOIs | |
Publication status | Published - 30 Sept 2021 |
Externally published | Yes |
Keywords
- human-computer interface
- gaze-typing
- oculomotor plasticity
- inhibitory control
- assistive technology