1 / 25

Eye Movement-Based Interaction What You Look At Is What You Get WYLAIWYG

breanna
Download Presentation

Eye Movement-Based Interaction What You Look At Is What You Get WYLAIWYG

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


    1. Eye Movement-Based Interaction ”What You Look At Is What You Get ” (WYLAIWYG) Aulikki Hyrskykari 19th January 2000

    2. 3/31/2012 2/27 Eye Movement-Based Interaction Eye on/in the interface (2) Problems and research issues: Technological/HCI issues (2) Processing the eye movement data (5) Eye as a control device (2) Command based gaze interaction (9) Noncommand gaze interaction (3) References Project ideas

    3. 3/31/2012 3/27 Constrained interface between two powerful information processors . Eye on/in the interface

    4. 3/31/2012 4/27 Eye on/in the interface Need for keeping the hands free (or the hands can not be used for other reasons) Increasing number of computer users suffer from RSI (repetitive stress injury) Eye movements are natural, little conscious effort Direction of gaze implicitly indicates the focus of attention

    5. 3/31/2012 5/27 Problems and Research Issues 1) Technological issues Usability of the hardware head mounted systems more reliable but somewhat awkward floor mounted more comfortable but more constrained Accuracy - need of calibration for every user at the beginning of a task also during the task Costs of eye tracking (equipment)

    6. 3/31/2012 6/27 Problems and Research Issues 2) HCI issues Need to design and study new interaction techniques Eyes are a perceptual device, not evolved to a control organ people are not used to operate things by simply looking at them - if poorly done it could be very annoying Noisy data - need to refine in order to get useful dialogue information (fixations, input tokens, intentions) accuracy restricted by biological characteristics of the eye

    7. 3/31/2012 7/27 Processing the EM data

    8. 3/31/2012 8/27

    9. 3/31/2012 9/27 Processing the EM data Scanpaths with fixations identified

    10. 3/31/2012 10/27 Processing the EM data (4/5) Input tokens [Siebert00] The fixations are then turned into input tokens start of fixation continuation of fixation (every 50 ms) end of fixation failure to locate eye position entering monitored regions The tokens formulate eye events are multiplexed into the event queue stream with other input events The eye events also carry information of the fixated screen object (using nearest neighbor approach)

    11. 3/31/2012 11/27 Processing the EM data Deducing user’s intentions Objective to refine the data further on for recognizing the user’s intentions to implement a higher level programming interface for gaze aware applications Eye Interpretation Engine, objective to identify such behaviors as [Edwards98] the user is reading just “looking around” starts and stops searching for an object (e.g. a button) wants to select an object

    12. 3/31/2012 12/27 Eye as a control device Gaze behavior very different from other devices used for controlling computer (hand, voice, feet) intentional control of eyes is difficult and stressful, the gaze is easily driven by external events precise control of eyes difficult “Midas touch” problem Most of the time the eyes are used for obtaining information with no intent to initiate commands Users are easily afraid of looking at the “eye active” objects or areas of the window

    13. 3/31/2012 13/27 Eye as a control device Jacob’s taxonomy Jacob’s taxonomy of possible approaches for using gaze input in the user interface:

    14. 3/31/2012 14/27 Even though eye movements are an old research area gaze aware applications practically do not exist Exception: applications for disabled Command based gaze interaction

    15. 3/31/2012 15/27 Command based gaze interaction Applications for disabled

    16. 3/31/2012 16/27 Command based gaze interaction Applications for disabled

    17. 3/31/2012 17/27 Command based gaze interaction Applications for disabled

    18. 3/31/2012 18/27 Command based gaze interaction Selection

    19. 3/31/2012 19/27 Command based gaze interaction Selection

    20. 3/31/2012 20/27 Command based gaze interaction Selection

    21. 3/31/2012 21/27 Command based gaze interaction Selection

    22. 3/31/2012 22/27 [Jacob98] Gaze controlled pull down menus using dwell time did not work out very well, the time was either too long or too prone to errors gaze+hardware button worked better Dragging of objects (with gaze only, with gaze + hardware button) performed better than most of the other experiments using the gaze + hardware button felt natural Command based gaze interaction - menus, dragging, scrolling, window manipulation

    23. 3/31/2012 23/27 Noncommand gaze interfaces Multimodal interfaces head towards task-oriented (and user oriented) interfaces instead of command oriented In non-command interfaces the computer monitors the user’s actions instead of waiting user’s commands [Nielsen93] In most cases the natural eye movement information could be valuable information for the application

    24. 3/31/2012 24/27 iEye -project (University of Tampere, SMI/Germany, GiuntiIlabs/Italy, Conexor/Espoo and University of Nottingham/England, started in January 2000) Noncommand gaze interfaces

    25. 3/31/2012 25/27

    26. 3/31/2012 26/27 References [Glenstrup95] Glenstrup Arne John , Engell-Nielsen Theo, Eye Controlled Media: Present and Future State. Published as a thesis at the University of Copenhage, Institute of Computer Science. Available in http://www.diku.dk/~panic/eyegaze/article.html [Jacob98] R.J.K. Jacob, "The Use of Eye Movements in Human-Computer Interaction Techniques: What You Look At is What You Get. ACM Transactions on Information Systems, 9 (3), 152-169, 1991. Also reprinted with commentary in Readings in Intelligent User Interfaces, ed. M.T. Maybury and W. Wahlster, Morgan Kaufmann, San Francisco, 1998, 65-83. [Nielsen93] Nielsen Jakob, Noncommand interfaces. CACM 36, 4, 1993, 83-99. [Ohno98] Ohno Takehiko, Features of eye gaze interface for selection tasks. APCHI’98, Japan, 1998, 176-181. [Siebert00] Siebert Linda E. and Jacob Robert J. K.,” Evaluation of Eye Gaze Interaction,” submitted to in the Proc. of ACM CHI 2000. (available in http://www.eecs.tufts.edu/~jacob/papers/chi00.sibert.pdf ) [Staker90] Staker India and Bolt Richard A., A gaze -responsive Self -Disclosing Display. ACM CHI’90, 3-9. [Ware87] Ware Colin and Mikaelian Harutun H., An Evaluation of an Eye Tracker as Device for Computer Input. Proc. ACM CHI'87, 183-188. [Yarbus67] Yarbus, A. L. (1967). Eye movements during perception of complex objects, in L. A. Riggs, ed., Eye Movements and Vision, Plenum Press, New York, chapter VII, 171-196. [Zhai99] Zhai Shumin, Morimoto Carlos, and Ihde Steven, Manual and Gaze Input Cascaded (MAGIC) pointing. In Proc. of ACM CHI 1999. 246-253.

More Related