The intuitive user interfaces of PCs and PDAs, such as pen display and touch panel, have become widely used in recent times. In this study, we have developed an eye-tracking pen display based on the stereo bright pupil technique. First, the bright pupil camera was developed by examining the arrangement of cameras and LEDs for pen display. Next, the gaze estimation method was proposed for the stereo bright pupil camera, which enables one point calibration. Then, the prototype of the eyetracking pen display was developed. The accuracy of the system was approximately 0.7° on average, which is sufficient for human interaction support. We also developed an eye-tracking tabletop as an application of the proposed stereo bright pupil technique.
2. 3 Reviews of Previous Studies to Decide Ar- right hand and the eyelid. Therefore, we reviewed the arrange-
rangement of Cameras and IR LEDs ments proposed in previous studies again. Some researchers
have proposed camera-LED integrated systems. For example,
Ohno developed a system that involved the use of one camera
As the first step of this study, we analyzed the body motions
and two LEDs [Ohno 2006]. Chen et al. developed a system
involved in using pen display of a right handed user. For this, we
that involved the use of two cameras and two LEDs mounted
used motion capture system (Vicon Motion Systems, Vicon 512)
near the camera centers; in this arrangement, the camera and the
and measured a subject’s body motion; i.e., movement of head,
LED were integrated into one component [Chen 2008]. We can
right shoulder, and arm. As shown in Figure 2, the posture of the
arrange such a system to the left of the pen display; however,
subject and the angle of pen display were assumed to be limited
such a system would be inadequate if the pen display is to be
to 3 cases to avoid hiding the cameras and IR LEDs.
used at various angles. The two cameras should be separated for
the eye tracking pen display system.
We developed a software for analyzing the arrangement. Figure
3 shows the screen shot of the software which draws the results
of measurement of 10 subjects. It can be seen that there is an 4 Stereo Bright Pupil Technique for Pen Dis-
unavailable volume for arranging cameras and IR LEDs at the play
left bottom.
4.1 Bright Pupil Camera
On the basis of the reviews of previous papers, we decided to
Sitting, use the stereo bright pupil technique. We integrated an IR LED
Pen display at the angle of 60° at the center of the camera (POINT GRAY, FFMV-03MTM,
752x480 pixels) lens, as shown in Figure 5; this modified cam-
Markers era is called the bright pupil camera. A 35-mm lens and an IR
filter are attached. We positioned two bright pupil cameras sepa-
Standing, rately to the left of the pen display (Figure 4 (c)). When these
Pen display at the angle of 60 ° cameras are used, the light from the LED reflects on the retina
and a bright pupil can be observed in the camera image.
Standing,
Pen display at the angle of 15 °
Figure 2: Measurement of body motion while using a pen display.
Pen display
Figure 5: Bright pupil camera.
4.2 Eye Model
Unavailable volume
for arranging cameras and IR LEDs Right arm Figure 6 shows the eye model in this study, which is typical in
Figure 3: Arrangement volume of cameras and LEDs. model-based approaches. An eye consists of two balls. It has
two axes: one is the optical axis of the eye that is the geometric
Next, we reviewed previous studies and developed a prototype center line of the eye, and the other is the visual axis that is the
of the system by considering its technical requirements. The 3D line of sight connecting the fovea. These axes intersect at the
gaze-tracking approach was selected for accuracy [Shih et al. center of the corneal curvature. The average of horizontal and
2004; Guestrin et al. 2007; Nagamatsu et al. 2008a]. This ap- vertical angles between the optical and visual axes are 5.5° and
proach involves the use of two cameras and three or four LEDs. 1.0°, respectively [OSAKA, 1993].
Figure 4 (a) shows the arrangement of the system proposed by
Nagamatsu et al. In this study, we first developed a prototype of Cornea
Pupil
the system by positioning the cameras and LEDs: two cameras A Center of Corneal Curvature
are placed to the left of the pen display, and one LED each is Visual Axis Rotation
Center E
placed on the top, left, and bottom frames of the pen display Optical Axis
(Figure 4 (b)). However, even with such an arrangement, stable
eye-tracking cannot be realized due to the obstructions by the B Pupil Fovea
LED Center
Figure 6: Eye model.
Camera 4.3 Image Processing
(a) (b) (c)
By using two bright pupil cameras, the light from one of the
Figure 4: Arrangement of cameras and LEDs. LED reflects at the retina and the camera image of the pupil
166
3. Edge detection
Binarlize
Figure 7: Example of image processing.
Figure 9: Prototype of the eye-tracking pen display.
becomes bright. In addition, there are two reflections of light
sources from the outer surface of the cornea called the first Pur-
kinje image (Figure 7, left). First, we carried out edge detection
in order to detect the position of the pupil. Next, we fitted an
ellipse to the edge, and calculated the pupil center. To detect the
position of the Purkinje image, we trimmed the neighborhood of
the pupil center, and binarlized the image. We considered the
two bright points as a Purkinje image (Figure 7, right). This 60 °
image processing was performed using Open CV 1.0.
4.4 Estimation of the Optical Axis of the Eye 300 mm
Figure 10: Experimental setup.
We estimated the optical axis on the basis of the results of image
processing. We initially calculated the relationship between each can be realized while a user is drawing a line while looking at
pixel on the image plane and the corresponding 3D position by the tip of the pen. The white cross is the estimated point. We can
calibrating the camera. We assumed that the light source and the confirm that the center of the white cross and the tip of the pen
camera center are at the same position. Then, we obtained a is almost the same. We developed this system on an HP xw4600
plane that contains A and B by using the expression Workstation with MS Windows XP. The frame rate was ap-
(C − Β') × (C − P')i( X − C) = 0 , where X is a point on the proximately 10 fps.
plane (Figure 8). One bright pupil camera was used to determine
one plane that contains the optical axis. Therefore, the optical We then evaluated the prototype. Figure 10 shows the experi-
axis can be obtained as the intersection of two planes obtained mental setup. The left part is the eye-tracking pen display and a
using the two cameras. While Chen estimated the optical axis by subject. The minimum distance between the subject and pen
determining Virtual B − A in Figure 8, we determined the exact display was 30 cm. The angle of pen display was 60°. The right
optical axis [Nagamatsu et al. 2010]. After that, the user gazes at LCD is displaying a captured and processed image.
a point on the pen display for calibration. The difference be-
tween optical axis and visual axis is revised by doing this cali- In the experiment, we asked the user to gaze at the marker at the
bration [Nagamatsu et al. 2008b]. The cross point of optical axis left side of the pen display for calibration. We next displayed a
and pen display is estimated as the gaze point. white cross on the pen display, and asked him to gaze at the
center of the white cross for 10 frames. The cross was displayed
Corneal on each of the 128 pixels. Because of the narrow range of view
Surface Virtual B angle and focus of the cameras, the area where a user can move
is limited. 3 students participated in the experiment.
Optical Axis B Pupil Center
Β'' 5.2 Results
Α
Light / Camera Center of Corneal Curvature
P Figure 11 shows the results. The accuracy was average 17.4
C pixels (5.2 mm) on the screen, which means about 0.71°. It was
Purkinje Image Purkinje Image
on Image Plage P' equivalent to Tobii, etc. In other words, the pen display can
recognize 22 horizontal lines.
B'
Image Plane 0 128 256 384 512 640 768 896 1024
0
Figure 8: Estimation of the Optical Axes.
128
5 Evaluation 256 Subject1
Subject2
384
5.1 Method Subject3
512
We integrated the bright pupil camera and pen display (Wacom,
640
DTI-520, 15 inch (380 mm), 1024 × 768 pixels) and developed a
prototype of the eye-tracking pen display. Figure 9 shows a pro- 768
totype of the eye-tracking pen display. Here, the gaze estimation
Figure 11: Result of evaluation experiment.
167
4. a prototype of an eye-tracking tabletop as an application of the
proposed stereo bright pupil technique, and confirmed effective-
ness of the system.
Acknowledgement
This work under our project “Embodied Communication Inter-
face for Mind Connection” has been supported by “New IT In-
Figure 12: Purkinje image on edge of cornea. frastructure for the Information-explosion Era” of MEXT Grant-
in-Aid for Scientific Research on Priority Areas. Also, our pro-
In the case of some subjects, the Purkinje image was reflected ject "Generation and Control Technology of Human-entrained
on the edge of the cornea, and the gaze point could not be cor- Embodied Media" has been supported by CREST (Core Re-
rectly estimated, as shown in Figure 12. However, this problem search for Evolution Science and Technology) of JST (Japan
can be solved by using one or more bright pupil cameras in a Science and Technology Agency).
layout-free arrangement.
References
6 Application
CHEN, J., TONG, Y., GRAY, W., AND JI. Q. 2008. A Robust 3D
The proposed method can be applied to develop various types of Eye Gaze Tracking System using Noise Reduction. In Proceed-
eye-tracking systems. ings of the 2008 symposium on Eye tracking research & appli-
cations, 189–196.
For example, we have developed a prototype of an eye-tracking
tabletop interface as shown in Figure 13. We integrated two GUESTRIN, E. D., AND EIZENMAN, M. 2007. Remote Point-of-
bright pupil cameras and a projector. The image is projected on Gaze Estimation with Free Head Movements Requiring a Sin-
the tabletop. This interface is used to realize both eye-gaze inter- gle-Point Calibration. In Proceedings of the 29th Annual Inter-
action and physical interaction. For example, the red square national Conference of the IEEE EMBS, 4556–4560.
indicates the gaze point on the tabletop. When a user is looking
at a physical pointer on the tabletop, the red square moves ap- NAGAMATSU, T., KAMAHARA, J., IKO, T., AND TANAKA, N. 2008.
propriately, followed by physical movement of the pointer. We One-Point Calibration Gaze Tracking Based on Eyeball Kine-
can enlarge the tabletop interface and extend the interaction area matics Using Stereo Cameras. In Proceedings of the 2008 sym-
to include off-surface areas. posium on Eye tracking research & applications, 95–98.
NAGAMATSU, T., KAMAHARA, J., AND TANAKA, N. 2008. 3D Gaze
Britht Pupil Camera Projector Tracking with Easy Calibration Using stereo Cameras for Ro-
bot and Human Communication. In Proceedings of IEEE RO-
MAN 2008, 59–64.
NAGAMATSU, T., IWAMOTO, Y., KAMAHARA, J., TANAKA, N., AND
User YAMAMOTO, Y. 2010. Gaze Estimation Method based on an
Aspherical Model of the Cornea Surface of Revolution about the
Eye-Gaze
Optical Axis of the Eye. In Proceedings of Eye Tracking Re-
Interaction
search & Applications Symposium ETRA 2010. (to appear).
Tabletop OHNO, T. 2006. One-point calibration gaze tracking method. In
Physical Proceedings of the 2006 symposium on Eye tracking research &
Interaction applications, 34.
Figure 13: Prototype of an eye-tracking tabletop.
OSAKA, R. 1993. Experimental Psychology of Eye Movements
(in Japanese). The University of Nagoya Press, Nagoya, Japan.
In this manner, bright pupil cameras enable flexible arrangement
of cameras, which can lead to the developments of various hu-
SHIH, S.-W., AND LIU, J. 2004. A Novel Approach to 3-D
man-computer interfaces such as pen displays and tabletops as
Gaze Tracking using Stereo Cameras. IEEE Transactions on
well as interaction analysis of laptops.
Systems, Man, and Cybernetics Part B 34, 1, 234–245.
7 Conclusion YAMAMOTO, M., AND WATANABE, T. 2005. Development of an
Embodied Interaction System with InterActor by Speech and
In this study, we have developed an eye-tracking pen display Hand Motion Input. In CD-ROM of the 2005 IEEE International
based on the stereo bright pupil technique. First, the bright pupil Workshop on Robots and Human Interactive Communication,
camera was developed by reviewing and examining the ar- 323–328.
rangement of cameras and LEDs for pen display. Next, the gaze
estimation method was proposed for the bright pupil camera, YAMAMOTO, M., AND WATANABE, T. 2008. Timing Control Ef-
which enables one-point calibration and wide-angle accuracy. fects of Utterance to Communicative Actions on Embodied In-
Then, the prototype of the eye-tracking pen display was devel- teraction with a Robot and CG Character. International Journal
oped. The accuracy of the system was approximately 0.7° on of Human-Computer Interaction 24, 1, 103–112.
average, which is sufficient for pen display. We also developed
168