SlideShare a Scribd company logo
1 of 2
Download to read offline
An Eye on Input:
            Research Challenges in Using the Eye for Computer Input Control
                                                                             I. Scott MacKenzie
                                                                  Dept. of Computer Science and Engineering
                                                                               York University
                                                                     Toronto, Ontario, Canada M3J 1P3
                                                                             mack@cse.yorku.ca

Abstract                                                                                          of the crosshair maps to the user’s point of fixation on the
                                                                                                  computer display. Or does it? Eye tracking systems are generally
The human eye, with the assistance of an eye tracking apparatus,                                  configured to hide the corresponding crosshair on the computer
may serve as an input controller to a computer system. Much                                       display even though the software indeed computes an x-y
like point-select operations with a mouse, the eye can "look-                                     coordinate of the point of fixation. The reason for this is simple.
select", and thereby activate items such as buttons, icons, links,                                The on-screen crosshair is rarely located at the point of fixation,
or text. Applications for accessible computing are particularly                                   or more correctly, at the point of fixation in the user’s mind.
enticing, since the manual ability of disabled users is often                                     Upon seeing an on-screen crosshair that is “near” the actual
lacking or limited. Whether for the able-bodied or the disabled,                                  point of fixation, there is a strong tendency for the user to look
computer control systems using the eye as an input “device”                                       at the crosshair. The result is obvious enough. The ensuing cat-
present numerous research challenges. These involve                                               and-mouse exercise is quickly a source of frustration. So, the
accommodating the innate characteristics of the eye, such as                                      on-screen crosshair is hidden.
movement by saccades, jitter and drift in eye position, the
absence of a simple and intuitive selection method, and the
inability to determine a precise point of fixation through eye
position alone.

Looking

For computer input control, current eye tracking systems
perform poorly in comparison to their manually operated
counterparts, such as mice or joysticks. To improve the design
of eye tracking systems, we might begin by asking: What is a
perfect eye tracker?
                                                                                                                              Figure 1
This question is only worth considering from an engineering                                       What is the problem in the scenario just described? Is this
perspective, since it is not possible to know what meaning the                                    simply a matter of developing a better algorithm? The points of
user ascribes to a visual image. Yet, even from an engineering                                    deviation are well known. The digitized camera image is of
perspective, the challenge is formidable. First, consider the                                     limited resolution. The user’s head moves about slightly. Each
capabilities of a human interacting with a computer, apart from                                   user’s eyes are different in size and shape. There are slight on-
eye tracking. It is relatively straightforward for a computer user                                going movements of the eye, known as jitter and drift. These
to look at, or fixate on, a single pixel on a computer display,                                   issues can be accommodated through improved engineering,
provided the pixel is distinguishable from its neighbors by virtue                                better algorithms, and careful calibration. So, we might expect
of colour or contrast. Furthermore, if in control of an on-screen                                 these challenges to be eventually overcome, such that the
pointer via a mouse, the user can, with a little effort, position the                             position the on-screen crosshair is precisely at the user’s point of
pointer’s hot spot (e.g., the tip of an arrow) on the desired pixel.                              fixation. But, this is not the case. To illustrate what may be the
So a “perfect” eye tracker would, like a mouse, return the x-y                                    most significant obstacle in perfecting eye tracking technology,
coordinate of the gazed-upon pixel. Is such single-pixel                                          consider the sketch in Figure 2.
“looking” possible with an eye tracker, a perfect eye tracker?
The answer is no.

Importantly, the software and algorithms in an eye tracking
system have little to work with other than a digitized camera
image of the user’s face and eyes. An example is shown in
Figure 1. The user’s left eye is expanded showing two glints –
corneal reflections from the eye tracker’s infrared light sources –
and a crosshair. Through some sophisticated software, the center
Copyright © 2010 by the Association for Computing Machinery, Inc.
Permission to make digital or hard copies of part or all of this work for personal or
classroom use is granted without fee provided that copies are not made or distributed
for commercial advantage and that copies bear this notice and the full citation on the
first page. Copyrights for components of this work owned by others than ACM must be                                           Figure 2
honored. Abstracting with credit is permitted. To copy otherwise, to republish, to post on
servers, or to redistribute to lists, requires prior specific permission and/or a fee.            The user’s arm is out-stretched to a distance of 25 inches. His
Request permissions from Permissions Dept, ACM Inc., fax +1 (212) 869-0481 or e-mail              thumb is half an inch wide. The angle of visual arc is about 1.1
permissions@acm.org.
ETRA 2010, Austin, TX, March 22 – 24, 2010.
© 2010 ACM 978-1-60558-994-7/10/0003 $10.00

                                                                                             11
degrees. This exercise is commonly restated in the literature to
demonstrate the approximate diameter of the eye’s foveal image.
The fovea is located near the center of retina and is responsible
for sharp central vision. Although the fovea is only about 1% of
the retina in size, the neural processing associated with the
foveal image engages about 50% of the visual cortex in the
brain. From an eye tracking perspective, the foveal image
represents a “region of uncertainty”. Figure 3 shows the
approximate size of the foveal image superimposed on a typical
full-screen rendering of a web page. Several selectable buttons
fall within the foveal image, making selection difficult.




                                                                                                     Figure 4
                                                                         Accessible Computing

                                                                         Eye tracking is of particular interest in accessible computing,
                                                                         where the user has a disability that limits or excludes manual
                                                                         input. If holding the head steady is difficult, eye tracking will be
                                                                         difficult as well. Yet in such cases, limited eye tracking may be
                                                                         possible, such as detecting simple eye gestures (e.g., look left,
                                                                         look right, etc.). Single-switch scanning keyboards have a long
                                                                         history in accessible computing. A scanning ambiguous
                                                                         keyboard (SAK) is a recent variation that promises higher text
                           Figure 3                                      entry rates than traditional row-column scanning arrangements.
                                                                         Input only requires a single key or switch. The key press signal
Even if it were possible to determine, through careful                   can be provided using the blink detection capability of an eye
engineering and judicious calculation, the exact region on the           tracker. See the Ashtiani and MacKenzie paper elsewhere in the
computer display that is projected onto the fovea, it is not             ETRA 2010 proceedings.
possible to determine a particular point, or pixel, within the
foveal image to which the user is attending. This limitation             Evaluating
cannot be overcome by improvements in eye tracking
technology.                                                              Evaluating the eye working in concert with an eye tracking
                                                                         system requires a methodology that uniquely addresses the
Selecting                                                                characteristics of both the eye and the eye tracking apparatus. As
                                                                         well as reviewing research challenges in looking and selecting,
While many applications of eye tracking systems are directed at          this presentation will summarize and advance the use of
visual search (where is the user looking?), other applications           standardized empirical research methods for evaluating and
take on the additional and role of selecting – using the eye as a        comparing the many new approaches that researchers develop to
computer input control. Holding the point of fixation still for a        improve the ability of eye tracking systems to allow the user
period of time (“dwelling”) is the most common method of                 through, his or her eyes, to control a computer system.
selecting, but other methods are also possible and practical,
depending on the context and application. These include using a
separate manually operated switch or key, or using blinks,
winks, voluntary muscle contractions, and so on.

Regardless of the method of selection, the system must
accommodate the foveal image’s region of uncertainty. The
most obvious solution is simply to design interfaces where the
largest selectable region is larger than the foveal image, as
depicted in Figure 4. Clearly, such a big-button approach
seriously limits the interface options. In view of this, numerous
“smart” target acquisition techniques have been proposed. By
and large, they involve determining which target is the intended
target, given inaccuracies and fluctuations in what the eye
tracking system returns as the user’s point of fixation. The
techniques allow for selection of targets smaller than the foveal
image. Techniques proposed in the literature are numerous and
accompanied with a myriad of terms, such as cursor warping,
grab and hold, target expansion, fisheye lenses, snap on, snap
clutch, nearest neighbor, and so on.



                                                                    12

More Related Content

What's hot

Master thesis 2
Master thesis 2Master thesis 2
Master thesis 2
Jb Soni
 
Human activity recognition
Human activity recognition Human activity recognition
Human activity recognition
srikanthgadam
 
Eye Gaze Tracking With a Web Camera in a Desktop Environment
Eye Gaze Tracking With a Web Camera in a Desktop EnvironmentEye Gaze Tracking With a Web Camera in a Desktop Environment
Eye Gaze Tracking With a Web Camera in a Desktop Environment
1crore projects
 
Schneider.2011.an open source low-cost eye-tracking system for portable real-...
Schneider.2011.an open source low-cost eye-tracking system for portable real-...Schneider.2011.an open source low-cost eye-tracking system for portable real-...
Schneider.2011.an open source low-cost eye-tracking system for portable real-...
mrgazer
 
International Journal of Image Processing (IJIP) Volume (3) Issue (2)
International Journal of Image Processing (IJIP) Volume (3) Issue (2)International Journal of Image Processing (IJIP) Volume (3) Issue (2)
International Journal of Image Processing (IJIP) Volume (3) Issue (2)
CSCJournals
 

What's hot (20)

Saksham presentation
Saksham presentationSaksham presentation
Saksham presentation
 
Kohlbecker Low Latency Combined Eye And Head Tracking System For Teleoperatin...
Kohlbecker Low Latency Combined Eye And Head Tracking System For Teleoperatin...Kohlbecker Low Latency Combined Eye And Head Tracking System For Teleoperatin...
Kohlbecker Low Latency Combined Eye And Head Tracking System For Teleoperatin...
 
Gj3511231126
Gj3511231126Gj3511231126
Gj3511231126
 
Survey on Human Computer interaction for disabled persons
Survey on Human Computer interaction  for disabled persons Survey on Human Computer interaction  for disabled persons
Survey on Human Computer interaction for disabled persons
 
Saksham seminar report
Saksham seminar reportSaksham seminar report
Saksham seminar report
 
Master thesis 2
Master thesis 2Master thesis 2
Master thesis 2
 
Gait analysis report
Gait analysis reportGait analysis report
Gait analysis report
 
Human activity recognition
Human activity recognition Human activity recognition
Human activity recognition
 
Eye Gaze Tracking With a Web Camera in a Desktop Environment
Eye Gaze Tracking With a Web Camera in a Desktop EnvironmentEye Gaze Tracking With a Web Camera in a Desktop Environment
Eye Gaze Tracking With a Web Camera in a Desktop Environment
 
Report face recognition : ArganRecogn
Report face recognition :  ArganRecognReport face recognition :  ArganRecogn
Report face recognition : ArganRecogn
 
Biz model for eye tracking tech
Biz model for eye tracking techBiz model for eye tracking tech
Biz model for eye tracking tech
 
Schneider.2011.an open source low-cost eye-tracking system for portable real-...
Schneider.2011.an open source low-cost eye-tracking system for portable real-...Schneider.2011.an open source low-cost eye-tracking system for portable real-...
Schneider.2011.an open source low-cost eye-tracking system for portable real-...
 
The use of Eye Gaze Gesture Interaction Artificial Intelligence Techniques fo...
The use of Eye Gaze Gesture Interaction Artificial Intelligence Techniques fo...The use of Eye Gaze Gesture Interaction Artificial Intelligence Techniques fo...
The use of Eye Gaze Gesture Interaction Artificial Intelligence Techniques fo...
 
ICCE 2016 NCTU Courses Papers
ICCE 2016 NCTU Courses PapersICCE 2016 NCTU Courses Papers
ICCE 2016 NCTU Courses Papers
 
Eye gaze tracking with a web camera
Eye gaze tracking with a web cameraEye gaze tracking with a web camera
Eye gaze tracking with a web camera
 
International Journal of Image Processing (IJIP) Volume (3) Issue (2)
International Journal of Image Processing (IJIP) Volume (3) Issue (2)International Journal of Image Processing (IJIP) Volume (3) Issue (2)
International Journal of Image Processing (IJIP) Volume (3) Issue (2)
 
Gesture recognition using artificial neural network,a technology for identify...
Gesture recognition using artificial neural network,a technology for identify...Gesture recognition using artificial neural network,a technology for identify...
Gesture recognition using artificial neural network,a technology for identify...
 
IRJET- ATM Security using Machine Learning
IRJET- ATM Security using Machine LearningIRJET- ATM Security using Machine Learning
IRJET- ATM Security using Machine Learning
 
Eye tracking – an innovative monitor
Eye tracking – an innovative monitorEye tracking – an innovative monitor
Eye tracking – an innovative monitor
 
Virtual retinal-display ppt
Virtual retinal-display pptVirtual retinal-display ppt
Virtual retinal-display ppt
 

Similar to Mc Kenzie An Eye On Input Research Challenges In Using The Eye For Computer Input Control

20922174-Blue-Eyes-Technology.pdf
20922174-Blue-Eyes-Technology.pdf20922174-Blue-Eyes-Technology.pdf
20922174-Blue-Eyes-Technology.pdf
Yasmin297583
 

Similar to Mc Kenzie An Eye On Input Research Challenges In Using The Eye For Computer Input Control (20)

F0932733
F0932733F0932733
F0932733
 
IRJET- Sixth Sense Technology in Image Processing
IRJET-  	  Sixth Sense Technology in Image ProcessingIRJET-  	  Sixth Sense Technology in Image Processing
IRJET- Sixth Sense Technology in Image Processing
 
Moshnyaga The Use Of Eye Tracking For Pc Energy Management
Moshnyaga The Use Of Eye Tracking For Pc Energy ManagementMoshnyaga The Use Of Eye Tracking For Pc Energy Management
Moshnyaga The Use Of Eye Tracking For Pc Energy Management
 
Waymo Essay
Waymo EssayWaymo Essay
Waymo Essay
 
Eyetracking
EyetrackingEyetracking
Eyetracking
 
IRJET- Smart Mirror using Eye Gaze Tracking
IRJET- Smart Mirror using Eye Gaze TrackingIRJET- Smart Mirror using Eye Gaze Tracking
IRJET- Smart Mirror using Eye Gaze Tracking
 
Eye tracker based HCI
Eye tracker based HCIEye tracker based HCI
Eye tracker based HCI
 
screen less display documentation
screen less display documentationscreen less display documentation
screen less display documentation
 
Comparative study on computers operated by eyes and brain
Comparative study on computers operated by eyes and brainComparative study on computers operated by eyes and brain
Comparative study on computers operated by eyes and brain
 
Istance Designing Gaze Gestures For Gaming An Investigation Of Performance
Istance Designing Gaze Gestures For Gaming An Investigation Of PerformanceIstance Designing Gaze Gestures For Gaming An Investigation Of Performance
Istance Designing Gaze Gestures For Gaming An Investigation Of Performance
 
IRJET- Navigation and Camera Reading System for Visually Impaired
IRJET- Navigation and Camera Reading System for Visually ImpairedIRJET- Navigation and Camera Reading System for Visually Impaired
IRJET- Navigation and Camera Reading System for Visually Impaired
 
A SMART BRAIN CONTROLLED WHEELCHAIR BASED MICROCONTROLLER SYSTEM
A SMART BRAIN CONTROLLED WHEELCHAIR BASED MICROCONTROLLER SYSTEMA SMART BRAIN CONTROLLED WHEELCHAIR BASED MICROCONTROLLER SYSTEM
A SMART BRAIN CONTROLLED WHEELCHAIR BASED MICROCONTROLLER SYSTEM
 
A SMART BRAIN CONTROLLED WHEELCHAIR BASED MICROCONTROLLER SYSTEM
A SMART BRAIN CONTROLLED WHEELCHAIR BASED MICROCONTROLLER SYSTEMA SMART BRAIN CONTROLLED WHEELCHAIR BASED MICROCONTROLLER SYSTEM
A SMART BRAIN CONTROLLED WHEELCHAIR BASED MICROCONTROLLER SYSTEM
 
IRJET- Igaze-Eye Gaze Direction Evaluation to Operate a Virtual Keyboard for ...
IRJET- Igaze-Eye Gaze Direction Evaluation to Operate a Virtual Keyboard for ...IRJET- Igaze-Eye Gaze Direction Evaluation to Operate a Virtual Keyboard for ...
IRJET- Igaze-Eye Gaze Direction Evaluation to Operate a Virtual Keyboard for ...
 
MOUSE SIMULATION USING NON MAXIMUM SUPPRESSION
MOUSE SIMULATION USING NON MAXIMUM SUPPRESSIONMOUSE SIMULATION USING NON MAXIMUM SUPPRESSION
MOUSE SIMULATION USING NON MAXIMUM SUPPRESSION
 
Cursor Movement with Eyeball
Cursor Movement with EyeballCursor Movement with Eyeball
Cursor Movement with Eyeball
 
20922174-Blue-Eyes-Technology.pdf
20922174-Blue-Eyes-Technology.pdf20922174-Blue-Eyes-Technology.pdf
20922174-Blue-Eyes-Technology.pdf
 
Ryan Match Moving For Area Based Analysis Of Eye Movements In Natural Tasks
Ryan Match Moving For Area Based Analysis Of Eye Movements In Natural TasksRyan Match Moving For Area Based Analysis Of Eye Movements In Natural Tasks
Ryan Match Moving For Area Based Analysis Of Eye Movements In Natural Tasks
 
IRJET- Sign Language Interpreter
IRJET- Sign Language InterpreterIRJET- Sign Language Interpreter
IRJET- Sign Language Interpreter
 
Blue eyes seminar report
Blue eyes seminar reportBlue eyes seminar report
Blue eyes seminar report
 

More from Kalle

More from Kalle (20)

Blignaut Visual Span And Other Parameters For The Generation Of Heatmaps
Blignaut Visual Span And Other Parameters For The Generation Of HeatmapsBlignaut Visual Span And Other Parameters For The Generation Of Heatmaps
Blignaut Visual Span And Other Parameters For The Generation Of Heatmaps
 
Zhang Eye Movement As An Interaction Mechanism For Relevance Feedback In A Co...
Zhang Eye Movement As An Interaction Mechanism For Relevance Feedback In A Co...Zhang Eye Movement As An Interaction Mechanism For Relevance Feedback In A Co...
Zhang Eye Movement As An Interaction Mechanism For Relevance Feedback In A Co...
 
Wastlund What You See Is Where You Go Testing A Gaze Driven Power Wheelchair ...
Wastlund What You See Is Where You Go Testing A Gaze Driven Power Wheelchair ...Wastlund What You See Is Where You Go Testing A Gaze Driven Power Wheelchair ...
Wastlund What You See Is Where You Go Testing A Gaze Driven Power Wheelchair ...
 
Vinnikov Contingency Evaluation Of Gaze Contingent Displays For Real Time Vis...
Vinnikov Contingency Evaluation Of Gaze Contingent Displays For Real Time Vis...Vinnikov Contingency Evaluation Of Gaze Contingent Displays For Real Time Vis...
Vinnikov Contingency Evaluation Of Gaze Contingent Displays For Real Time Vis...
 
Urbina Pies With Ey Es The Limits Of Hierarchical Pie Menus In Gaze Control
Urbina Pies With Ey Es The Limits Of Hierarchical Pie Menus In Gaze ControlUrbina Pies With Ey Es The Limits Of Hierarchical Pie Menus In Gaze Control
Urbina Pies With Ey Es The Limits Of Hierarchical Pie Menus In Gaze Control
 
Urbina Alternatives To Single Character Entry And Dwell Time Selection On Eye...
Urbina Alternatives To Single Character Entry And Dwell Time Selection On Eye...Urbina Alternatives To Single Character Entry And Dwell Time Selection On Eye...
Urbina Alternatives To Single Character Entry And Dwell Time Selection On Eye...
 
Tien Measuring Situation Awareness Of Surgeons In Laparoscopic Training
Tien Measuring Situation Awareness Of Surgeons In Laparoscopic TrainingTien Measuring Situation Awareness Of Surgeons In Laparoscopic Training
Tien Measuring Situation Awareness Of Surgeons In Laparoscopic Training
 
Takemura Estimating 3 D Point Of Regard And Visualizing Gaze Trajectories Und...
Takemura Estimating 3 D Point Of Regard And Visualizing Gaze Trajectories Und...Takemura Estimating 3 D Point Of Regard And Visualizing Gaze Trajectories Und...
Takemura Estimating 3 D Point Of Regard And Visualizing Gaze Trajectories Und...
 
Stevenson Eye Tracking With The Adaptive Optics Scanning Laser Ophthalmoscope
Stevenson Eye Tracking With The Adaptive Optics Scanning Laser OphthalmoscopeStevenson Eye Tracking With The Adaptive Optics Scanning Laser Ophthalmoscope
Stevenson Eye Tracking With The Adaptive Optics Scanning Laser Ophthalmoscope
 
Stellmach Advanced Gaze Visualizations For Three Dimensional Virtual Environm...
Stellmach Advanced Gaze Visualizations For Three Dimensional Virtual Environm...Stellmach Advanced Gaze Visualizations For Three Dimensional Virtual Environm...
Stellmach Advanced Gaze Visualizations For Three Dimensional Virtual Environm...
 
Skovsgaard Small Target Selection With Gaze Alone
Skovsgaard Small Target Selection With Gaze AloneSkovsgaard Small Target Selection With Gaze Alone
Skovsgaard Small Target Selection With Gaze Alone
 
San Agustin Evaluation Of A Low Cost Open Source Gaze Tracker
San Agustin Evaluation Of A Low Cost Open Source Gaze TrackerSan Agustin Evaluation Of A Low Cost Open Source Gaze Tracker
San Agustin Evaluation Of A Low Cost Open Source Gaze Tracker
 
Rosengrant Gaze Scribing In Physics Problem Solving
Rosengrant Gaze Scribing In Physics Problem SolvingRosengrant Gaze Scribing In Physics Problem Solving
Rosengrant Gaze Scribing In Physics Problem Solving
 
Qvarfordt Understanding The Benefits Of Gaze Enhanced Visual Search
Qvarfordt Understanding The Benefits Of Gaze Enhanced Visual SearchQvarfordt Understanding The Benefits Of Gaze Enhanced Visual Search
Qvarfordt Understanding The Benefits Of Gaze Enhanced Visual Search
 
Prats Interpretation Of Geometric Shapes An Eye Movement Study
Prats Interpretation Of Geometric Shapes An Eye Movement StudyPrats Interpretation Of Geometric Shapes An Eye Movement Study
Prats Interpretation Of Geometric Shapes An Eye Movement Study
 
Pontillo Semanti Code Using Content Similarity And Database Driven Matching T...
Pontillo Semanti Code Using Content Similarity And Database Driven Matching T...Pontillo Semanti Code Using Content Similarity And Database Driven Matching T...
Pontillo Semanti Code Using Content Similarity And Database Driven Matching T...
 
Park Quantification Of Aesthetic Viewing Using Eye Tracking Technology The In...
Park Quantification Of Aesthetic Viewing Using Eye Tracking Technology The In...Park Quantification Of Aesthetic Viewing Using Eye Tracking Technology The In...
Park Quantification Of Aesthetic Viewing Using Eye Tracking Technology The In...
 
Palinko Estimating Cognitive Load Using Remote Eye Tracking In A Driving Simu...
Palinko Estimating Cognitive Load Using Remote Eye Tracking In A Driving Simu...Palinko Estimating Cognitive Load Using Remote Eye Tracking In A Driving Simu...
Palinko Estimating Cognitive Load Using Remote Eye Tracking In A Driving Simu...
 
Nakayama Estimation Of Viewers Response For Contextual Understanding Of Tasks...
Nakayama Estimation Of Viewers Response For Contextual Understanding Of Tasks...Nakayama Estimation Of Viewers Response For Contextual Understanding Of Tasks...
Nakayama Estimation Of Viewers Response For Contextual Understanding Of Tasks...
 
Nagamatsu User Calibration Free Gaze Tracking With Estimation Of The Horizont...
Nagamatsu User Calibration Free Gaze Tracking With Estimation Of The Horizont...Nagamatsu User Calibration Free Gaze Tracking With Estimation Of The Horizont...
Nagamatsu User Calibration Free Gaze Tracking With Estimation Of The Horizont...
 

Mc Kenzie An Eye On Input Research Challenges In Using The Eye For Computer Input Control

  • 1. An Eye on Input: Research Challenges in Using the Eye for Computer Input Control I. Scott MacKenzie Dept. of Computer Science and Engineering York University Toronto, Ontario, Canada M3J 1P3 mack@cse.yorku.ca Abstract of the crosshair maps to the user’s point of fixation on the computer display. Or does it? Eye tracking systems are generally The human eye, with the assistance of an eye tracking apparatus, configured to hide the corresponding crosshair on the computer may serve as an input controller to a computer system. Much display even though the software indeed computes an x-y like point-select operations with a mouse, the eye can "look- coordinate of the point of fixation. The reason for this is simple. select", and thereby activate items such as buttons, icons, links, The on-screen crosshair is rarely located at the point of fixation, or text. Applications for accessible computing are particularly or more correctly, at the point of fixation in the user’s mind. enticing, since the manual ability of disabled users is often Upon seeing an on-screen crosshair that is “near” the actual lacking or limited. Whether for the able-bodied or the disabled, point of fixation, there is a strong tendency for the user to look computer control systems using the eye as an input “device” at the crosshair. The result is obvious enough. The ensuing cat- present numerous research challenges. These involve and-mouse exercise is quickly a source of frustration. So, the accommodating the innate characteristics of the eye, such as on-screen crosshair is hidden. movement by saccades, jitter and drift in eye position, the absence of a simple and intuitive selection method, and the inability to determine a precise point of fixation through eye position alone. Looking For computer input control, current eye tracking systems perform poorly in comparison to their manually operated counterparts, such as mice or joysticks. To improve the design of eye tracking systems, we might begin by asking: What is a perfect eye tracker? Figure 1 This question is only worth considering from an engineering What is the problem in the scenario just described? Is this perspective, since it is not possible to know what meaning the simply a matter of developing a better algorithm? The points of user ascribes to a visual image. Yet, even from an engineering deviation are well known. The digitized camera image is of perspective, the challenge is formidable. First, consider the limited resolution. The user’s head moves about slightly. Each capabilities of a human interacting with a computer, apart from user’s eyes are different in size and shape. There are slight on- eye tracking. It is relatively straightforward for a computer user going movements of the eye, known as jitter and drift. These to look at, or fixate on, a single pixel on a computer display, issues can be accommodated through improved engineering, provided the pixel is distinguishable from its neighbors by virtue better algorithms, and careful calibration. So, we might expect of colour or contrast. Furthermore, if in control of an on-screen these challenges to be eventually overcome, such that the pointer via a mouse, the user can, with a little effort, position the position the on-screen crosshair is precisely at the user’s point of pointer’s hot spot (e.g., the tip of an arrow) on the desired pixel. fixation. But, this is not the case. To illustrate what may be the So a “perfect” eye tracker would, like a mouse, return the x-y most significant obstacle in perfecting eye tracking technology, coordinate of the gazed-upon pixel. Is such single-pixel consider the sketch in Figure 2. “looking” possible with an eye tracker, a perfect eye tracker? The answer is no. Importantly, the software and algorithms in an eye tracking system have little to work with other than a digitized camera image of the user’s face and eyes. An example is shown in Figure 1. The user’s left eye is expanded showing two glints – corneal reflections from the eye tracker’s infrared light sources – and a crosshair. Through some sophisticated software, the center Copyright © 2010 by the Association for Computing Machinery, Inc. Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be Figure 2 honored. Abstracting with credit is permitted. To copy otherwise, to republish, to post on servers, or to redistribute to lists, requires prior specific permission and/or a fee. The user’s arm is out-stretched to a distance of 25 inches. His Request permissions from Permissions Dept, ACM Inc., fax +1 (212) 869-0481 or e-mail thumb is half an inch wide. The angle of visual arc is about 1.1 permissions@acm.org. ETRA 2010, Austin, TX, March 22 – 24, 2010. © 2010 ACM 978-1-60558-994-7/10/0003 $10.00 11
  • 2. degrees. This exercise is commonly restated in the literature to demonstrate the approximate diameter of the eye’s foveal image. The fovea is located near the center of retina and is responsible for sharp central vision. Although the fovea is only about 1% of the retina in size, the neural processing associated with the foveal image engages about 50% of the visual cortex in the brain. From an eye tracking perspective, the foveal image represents a “region of uncertainty”. Figure 3 shows the approximate size of the foveal image superimposed on a typical full-screen rendering of a web page. Several selectable buttons fall within the foveal image, making selection difficult. Figure 4 Accessible Computing Eye tracking is of particular interest in accessible computing, where the user has a disability that limits or excludes manual input. If holding the head steady is difficult, eye tracking will be difficult as well. Yet in such cases, limited eye tracking may be possible, such as detecting simple eye gestures (e.g., look left, look right, etc.). Single-switch scanning keyboards have a long history in accessible computing. A scanning ambiguous keyboard (SAK) is a recent variation that promises higher text Figure 3 entry rates than traditional row-column scanning arrangements. Input only requires a single key or switch. The key press signal Even if it were possible to determine, through careful can be provided using the blink detection capability of an eye engineering and judicious calculation, the exact region on the tracker. See the Ashtiani and MacKenzie paper elsewhere in the computer display that is projected onto the fovea, it is not ETRA 2010 proceedings. possible to determine a particular point, or pixel, within the foveal image to which the user is attending. This limitation Evaluating cannot be overcome by improvements in eye tracking technology. Evaluating the eye working in concert with an eye tracking system requires a methodology that uniquely addresses the Selecting characteristics of both the eye and the eye tracking apparatus. As well as reviewing research challenges in looking and selecting, While many applications of eye tracking systems are directed at this presentation will summarize and advance the use of visual search (where is the user looking?), other applications standardized empirical research methods for evaluating and take on the additional and role of selecting – using the eye as a comparing the many new approaches that researchers develop to computer input control. Holding the point of fixation still for a improve the ability of eye tracking systems to allow the user period of time (“dwelling”) is the most common method of through, his or her eyes, to control a computer system. selecting, but other methods are also possible and practical, depending on the context and application. These include using a separate manually operated switch or key, or using blinks, winks, voluntary muscle contractions, and so on. Regardless of the method of selection, the system must accommodate the foveal image’s region of uncertainty. The most obvious solution is simply to design interfaces where the largest selectable region is larger than the foveal image, as depicted in Figure 4. Clearly, such a big-button approach seriously limits the interface options. In view of this, numerous “smart” target acquisition techniques have been proposed. By and large, they involve determining which target is the intended target, given inaccuracies and fluctuations in what the eye tracking system returns as the user’s point of fixation. The techniques allow for selection of targets smaller than the foveal image. Techniques proposed in the literature are numerous and accompanied with a myriad of terms, such as cursor warping, grab and hold, target expansion, fisheye lenses, snap on, snap clutch, nearest neighbor, and so on. 12