Optimizing Gaze Typing with Pupil Size for People with Severe Motor Disabilities

Article Preview

Abstract:

Thanks to recent technological advances in the field of eye tracking, eye typing provides means of communication for people with severe disabilities. Typing with gaze using dwell time has been made possible by the development of eye tracking technologies. Recent research indicates that pupil size is viewed as a subtle cue of people is making a decision. Therefore, it may help to infer users’ willing of typing. The present study describes the design process for improving eye typing by adding pupil size index into dwell time triggering. Experimental evaluations showed that the approach was effective; design considerations for such optimization of the gaze typing interfaces are discussed.

You might also be interested in these eBooks

Info:

Periodical:

Pages:

1330-1333

Citation:

Online since:

September 2014

Export:

Price:

Permissions CCC:

Permissions PLS:

Сopyright:

© 2014 Trans Tech Publications Ltd. All Rights Reserved

Share:

Citation:

* - Corresponding Author

[1] Bolt, R. A.: Eyes at the interface. in Proceedings of the 1982 conference on Human factors in computing systems, 1982, 360–362.

Google Scholar

[2] Päivi Majaranta and Kari-Jouko Räihä. 2002. Twenty years of eye typing: systems and design issues. In Proceedings of the 2002 symposium on Eye tracking research & applications (ETRA '02). ACM, New York, NY, USA, 15-22.

DOI: 10.1145/507072.507076

Google Scholar

[3] Abe, K., Ohi, S., and Ohyama, M., An eye-gaze input system using information on eye movement history, In Proceedings of the 4th International Conference on Universal Access in Human-computer Interaction: Ambient Interaction, pp.721-729, (2007).

DOI: 10.1007/978-3-540-73281-5_79

Google Scholar

[4] Areej Al-Wabil, Arwa Al-Issa, Itisam Hazzaa, May Al-Humaimeedi, Lujain Al-Tamimi, Bushra Al-Kadhi. Optimizing Gaze Typing for People with Severe Motor Disabilities: The iWriter Arabic Interface. ASSETS'12, October 22–24, 2012, 261-262.

DOI: 10.1145/2384916.2384983

Google Scholar

[5] Hutchinson, T. F.: Eye-Gaze Computer Interfaces: Computers That Sense Eye Positions on the Display. Computer, 26, 1993, 65–67.

DOI: 10.1109/mc.1993.620436

Google Scholar

[6] Jacob, R. J. K.: What you look at is what you get: eye movement-based interaction techniques. in Proceedings of the SIGCHI conference on Human factors in computing systems: Empowering people (CHI '90), 1990, 11–18.

DOI: 10.1145/97243.97246

Google Scholar

[7] Vilimek, R., Zander. T. O.: BC(eye): combining eye-gaze input with brain-computer interaction. in Proceedings of the 5th International on Conference Universal Access in Human-Computer Interaction. Part II: Intelligent and Ubiquitous Interaction Environments, 5615, 2009, 593–602.

DOI: 10.1007/978-3-642-02710-9_66

Google Scholar

[8] Simpson, H. M., Hale, S. M.: Pupillary changes during a decision making task. Perceptual and Motor Skills, 29(2), 1969, 495–498.

DOI: 10.2466/pms.1969.29.2.495

Google Scholar

[9] Einhauser, W., Koch, C., Carter, O.: Pupil dilation betrays the timing of decisions. Frontiers in Human Neuroscience, 4: 18, (2010).

DOI: 10.3389/fnhum.2010.00018

Google Scholar

[10] Aston-Jones, G., Cohen, J.: An integrative theory of locus coeruleusnorepinephrine function: adaptive gain and optimal performance. Annual Review of Neuroscience, 28, 2005, 403–450.

DOI: 10.1146/annurev.neuro.28.061604.135709

Google Scholar

[11] Chengzhi Feng. Eye-movement Based Human-Computer Interaction. Suzhou University press. 2010. 45-58.

Google Scholar