260 likes | 980 Views
E N D
1. Eye Movement-Based InteractionWhat You Look At Is What You Get (WYLAIWYG) Aulikki Hyrskykari19th January 2000
2. 3/31/2012 2/27 Eye Movement-Based Interaction Eye on/in the interface (2)
Problems and research issues: Technological/HCI issues (2)
Processing the eye movement data (5)
Eye as a control device (2)
Command based gaze interaction (9)
Noncommand gaze interaction (3)
References
Project ideas
3. 3/31/2012 3/27 Constrained interface
between two powerful information processors . Eye on/in the interface
4. 3/31/2012 4/27 Eye on/in the interface Need for keeping the hands free (or the hands can not be used for other reasons)
Increasing number of computer users suffer from RSI (repetitive stress injury)
Eye movements are natural, little conscious effort
Direction of gaze implicitly indicates the focus of attention
5. 3/31/2012 5/27 Problems and Research Issues 1) Technological issues
Usability of the hardware
head mounted systems more reliable but somewhat awkward
floor mounted more comfortable but more constrained
Accuracy - need of calibration
for every user at the beginning of a task
also during the task
Costs of eye tracking (equipment)
6. 3/31/2012 6/27 Problems and Research Issues 2) HCI issues
Need to design and study new interaction techniques
Eyes are a perceptual device, not evolved to a control organ
people are not used to operate things by simply looking at them - if poorly done it could be very annoying
Noisy data - need to refine in order to get useful dialogue information (fixations, input tokens, intentions)
accuracy restricted by biological characteristics of the eye
7. 3/31/2012 7/27 Processing the EM data
8. 3/31/2012 8/27
9. 3/31/2012 9/27 Processing the EM dataScanpaths with fixations identified
10. 3/31/2012 10/27 Processing the EM data (4/5)Input tokens [Siebert00]
The fixations are then turned into input tokens
start of fixation
continuation of fixation (every 50 ms)
end of fixation
failure to locate eye position
entering monitored regions
The tokens formulate eye events
are multiplexed into the event queue stream with other input events
The eye events also carry information of the fixated screen object (using nearest neighbor approach)
11. 3/31/2012 11/27 Processing the EM data Deducing users intentions Objective
to refine the data further on for recognizing the users intentions
to implement a higher level programming interface for gaze aware applications
Eye Interpretation Engine, objective to identify such behaviors as [Edwards98]
the user is reading
just looking around
starts and stops searching for an object (e.g. a button)
wants to select an object
12. 3/31/2012 12/27 Eye as a control device Gaze behavior very different from other devices used for controlling computer (hand, voice, feet)
intentional control of eyes is difficult and stressful, the gaze is easily driven by external events
precise control of eyes difficult
Midas touch problem
Most of the time the eyes are used for obtaining information with no intent to initiate commands
Users are easily afraid of looking at the eye active objects or areas of the window
13. 3/31/2012 13/27 Eye as a control deviceJacobs taxonomy Jacobs taxonomy of possible approaches for using gaze input in the user interface:
14. 3/31/2012 14/27 Even though eye movements are an old research area gaze aware applications practically do not exist
Exception: applications for disabled Command based gaze interaction
15. 3/31/2012 15/27 Command based gaze interaction Applications for disabled
16. 3/31/2012 16/27 Command based gaze interaction Applications for disabled
17. 3/31/2012 17/27 Command based gaze interaction Applications for disabled
18. 3/31/2012 18/27 Command based gaze interaction Selection
19. 3/31/2012 19/27 Command based gaze interaction Selection
20. 3/31/2012 20/27 Command based gaze interaction Selection
21. 3/31/2012 21/27 Command based gaze interaction Selection
22. 3/31/2012 22/27 [Jacob98]
Gaze controlled pull down menus
using dwell time did not work out very well, the time was either too long or too prone to errors
gaze+hardware button worked better
Dragging of objects (with gaze only, with gaze + hardware button)
performed better than most of the other experiments
using the gaze + hardware button felt natural
Command based gaze interaction - menus, dragging, scrolling, window manipulation
23. 3/31/2012 23/27 Noncommand gaze interfaces
Multimodal interfaces head towards task-oriented (and user oriented) interfaces instead of command oriented
In non-command interfaces the computer monitors the users actions instead of waiting users commands [Nielsen93]
In most cases the natural eye movement information could be valuable information for the application
24. 3/31/2012 24/27 iEye -project (University of Tampere, SMI/Germany,GiuntiIlabs/Italy, Conexor/Espoo and University of Nottingham/England, started in January 2000) Noncommand gaze interfaces
25. 3/31/2012 25/27
26. 3/31/2012 26/27 References [Glenstrup95] Glenstrup Arne John , Engell-Nielsen Theo, Eye Controlled Media: Present and Future State. Published as a thesis at the University of Copenhage, Institute of Computer Science. Available in http://www.diku.dk/~panic/eyegaze/article.html
[Jacob98] R.J.K. Jacob, "The Use of Eye Movements in Human-Computer Interaction Techniques: What You Look At is What You Get. ACM Transactions on Information Systems, 9 (3), 152-169, 1991. Also reprinted with commentary in Readings in Intelligent User Interfaces, ed. M.T. Maybury and W. Wahlster, Morgan Kaufmann, San Francisco, 1998, 65-83.
[Nielsen93] Nielsen Jakob, Noncommand interfaces. CACM 36, 4, 1993, 83-99.
[Ohno98] Ohno Takehiko, Features of eye gaze interface for selection tasks. APCHI98, Japan, 1998, 176-181.
[Siebert00] Siebert Linda E. and Jacob Robert J. K., Evaluation of Eye Gaze Interaction, submitted to in the Proc. of ACM CHI 2000. (available in http://www.eecs.tufts.edu/~jacob/papers/chi00.sibert.pdf )
[Staker90] Staker India and Bolt Richard A., A gaze -responsive Self -Disclosing Display. ACM CHI90, 3-9.
[Ware87] Ware Colin and Mikaelian Harutun H., An Evaluation of an Eye Tracker as Device for Computer Input. Proc. ACM CHI'87, 183-188.
[Yarbus67] Yarbus, A. L. (1967). Eye movements during perception of complex objects, in L. A. Riggs, ed., Eye Movements and Vision, Plenum Press, New York, chapter VII, 171-196.
[Zhai99] Zhai Shumin, Morimoto Carlos, and Ihde Steven, Manual and Gaze Input Cascaded (MAGIC) pointing. In Proc. of ACM CHI 1999. 246-253.