SlideShare ist ein Scribd-Unternehmen logo
1 von 4
Downloaden Sie, um offline zu lesen
Evaluation of a Low-Cost Open-Source Gaze Tracker
                         Javier San Agustin∗                                  Emilie Mollenbach‡                              Maria Barret§
                         Henrik Skovsgaard†                                 Loughborough University                            Martin Tall¶
                     IT University of Copenhagen                                                                           Dan Witzner Hansen
                                                                                                                           John Paulin Hansen∗∗
                                                                                                                       IT University of Copenhagen


Abstract                                                                                          However, this project is in a very early stage of development, and
                                                                                                  has not been active since 2007.
This paper presents a low-cost gaze tracking system that is based
                                                                                                  In this paper we introduce the ITU Gaze Tracker (available at
on a webcam mounted close to the user’s eye. The performance
                                                                                                  http://www.gazegroup.org/downloads/23-gazetracker), a low-cost
of the gaze tracker was evaluated in an eye-typing task using two
                                                                                                  system that was released as open source in Spring 2009 [San
different typing applications. Participants could type between 3.56
                                                                                                  Agustin et al. 2009], and we present an evaluation of a head-
and 6.78 words per minute, depending on the typing system used.
                                                                                                  mounted setup that employs a webcam with built-in infrared light.
A pilot study to assess the usability of the system was also car-
                                                                                                  Such setup does not require extra light sources, and therefore the
ried out in the home of a user with severe motor impairments. The
                                                                                                  system is very easy to build. However, the system is not head-pose
user successfully typed on a wall-projected interface using his eye
                                                                                                  invariant. The evaluation was carried out using two different eye-
movements.
                                                                                                  typing applications designed to handle a noisy input, GazeTalk and
                                                                                                  StarGazer. A pilot study with a person in the latest stages of ALS
CR Categories: H.5.2 [Input devices and strategies]: Evalua-                                      was also conducted.
tion/Methodology

Keywords: low cost, gaze interaction, gaze typing, off-the-shelf
                                                                                                  2     The ITU Gaze Tracker
components, augmentative and alternative communication
                                                                                                  In order to provide an accessible gaze tracking system, the ITU
                                                                                                  Gaze Tracker was developed. The design considerations and the
1      Introduction                                                                               hardware and software are presented in this section.

The cost of commercial gaze tracking systems often exceeds                                        2.1    Design Considerations
$10.000, preventing access to this technology by many people who
need it for daily communication. Using low-cost components ben-                                   The approach taken to develop the ITU Gaze Tracker aimed to pro-
efits users who cannot afford an expensive commercial system, and                                  vide a fully accessible system, with a set of design requirements:
it gives them a chance to try gaze interaction for the first time with-
out making a big investment. Furthermore, open source solutions                                       1. The gaze tracker should be robust and accurate enough to
might provide a starting point for enthusiasts, who can contribute                                       work with at least one gaze-communication system.
to an active project and adapt the software to fit specific needs.                                      2. Use of low-cost off-the-shelf components. The hardware em-
                                                                                                         ployed should be available in any electronics store or at on-
Commercial gaze tracking systems have been available for more                                            line shops to allow for easy acquisition and replacement. Fur-
than 15 years, but the technology is still a high priced niche. A few                                    thermore, no hardware modifications should be needed.
systems built using off-the-shelf components have been presented.                                     3. The user should be given flexibility to place the different com-
Babcock and Pelz [2004] developed a head-mounted system that                                             ponents (camera, infrared lights, computer display) at various
uses two cameras mounted on a pair of safety glasses, one to track                                       locations to fit specific needs. For instance, mounting the dis-
the eye and the other to record the scene in front of the viewer.                                        play on a wheel chair, moving it to a table or having it acces-
However, building this system requires an in-depth understanding                                         sible in bed should make no difference.
of electronics and the software only runs on the Linux operating
                                                                                                      4. Open-source software. Developing an efficient gaze tracker
system, thereby preventing most potential users from assembling
                                                                                                         from low-cost components is a huge endeavor. No single
their own system. The Opengazer eye tracker uses a web camera to
                                                                                                         group of developers is likely to come up with the ultimate
track the user’s eyes and is rather simple to build [Zieli´ ski 2007].
                                                           n
                                                                                                         solution. Open source allows anybody to improve and modify
     ∗ e-mail:                                                                                           the source code to fit specific needs.
                 javier@itu.dk
     † e-mail:   hhje@itu.dk
     ‡ e-mail:   author@affiliation.com                                                            2.2    Hardware
     § e-mail:   @affiliation.com
    ¶ e-mail:    mtall@itu.dk                                                                     Standard webcams usually have a low resolution and a broad field
       e-mail:   witzner@itu.dk                                                                   of view, but by placing the camera close to the user’s eye we can
    ∗∗ e-mail:   paulin@itu.dk                                                                    obtain images of sufficient quality. Figure 1 shows the gaze tracker
Copyright © 2010 by the Association for Computing Machinery, Inc.
                                                                                                  in two setups, (a) mounted on a piece of inexpensive balsa wood
Permission to make digital or hard copies of part or all of this work for personal or             that the user bites on in order to fix the webcam in front of the eye,
classroom use is granted without fee provided that copies are not made or distributed             and (b) mounted on a flexible arm (e.g., an office lamp).
for commercial advantage and that copies bear this notice and the full citation on the
first page. Copyrights for components of this work owned by others than ACM must be               The setup displayed in Figure 1 employs a Sandberg Nightvision 2
honored. Abstracting with credit is permitted. To copy otherwise, to republish, to post on        webcam, which costs around $20. The camera is equipped with 6
servers, or to redistribute to lists, requires prior specific permission and/or a fee.
Request permissions from Permissions Dept, ACM Inc., fax +1 (212) 869-0481 or e-mail
                                                                                                  built-in infrared LEDs that create a dark pupil effect. The resolution
permissions@acm.org.                                                                              is 640×480 pixels and the frame rate is 30 fps. The weight of the
ETRA 2010, Austin, TX, March 22 – 24, 2010.
© 2010 ACM 978-1-60558-994-7/10/0003 $10.00

                                                                                             77
2.3.3 Gaze Estimation

                                                                             Our flexibility requirements introduce a high uncertainty in the lo-
                                                                             cations of the camera and screen with respect to the user. There-
                                                                             fore, the gaze estimation technique employed must be as generic
                                                                             as possible, and should not assume any specific configuration of the
                                                                             hardware components. The gaze estimation method implemented is
                                                                             an interpolation technique that uses a pair of second-order polyno-
                   (a)                        (b)                            mials (one for each screen axis) [Morimoto et al. 1999]. A 12-point
                                                                             calibration procedure allows to calculate the coefficients that map
Figure 1: Mounting solutions of the ITU Gaze Tracker: (a) on a               the center of the pupil to the gaze coordinates on the screen. The
piece of inexpensive balsa wood; (b) on the arm of an office lamp.            calibration procedure takes approximately 20 seconds.
                                                                             The system was demonstrated at the CHI 2009 Conference. A total
                                                                             of 28 delegates tried out the system mounted on a piece of balsa
camera is 200g. The amount of light shone into the eyes is not               wood. 22 had normal vision or wore contact lenses, and 6 wore
considered to be harmful [Mulvey et al. ].                                   glasses. All 28 were able to get a successful calibration and control
                                                                             the cursor with their eyes.
2.3   Software
                                                                             3     Gaze Typing with the ITU GazeTracker
2.3.1 Software Architecture
                                                                             Text input is a very important task for people who use gaze to com-
The gaze tracking software is developed in C#, and employs
                                                                             municate. The design of typing systems is one of the most well
OpenCV for image processing. There are three main components:
                                                                             researched areas within the field of gaze interaction [Majaranta and
(1) The gaze tracking library, which implements all the methods to
                                                                             R¨ ih¨ 2002]. Gaze typing systems are often evaluated experimen-
                                                                               a a
control a tracker such as extracting eye features, run a calibration
                                                                             tally with standardized performance measurements such as words
procedure, estimate the gaze coordinates and detect the type of eye
                                                                             per minute (WPM) and error rates.
movement (i.e., fixation or saccade); (2) The camera class, respon-
sible for initializing a generic camera and grabbing images that are         The two eye typing applications used in this experiment have dif-
then processed by the gaze-tracking library; and (3) The user in-            ferent approaches to handle noise. GazeTalk [Hansen et al. 2004]
terface, which provides the communication with the gaze-tracking             has 10 large buttons with a hierarchical organization of the alpha-
library to set up the different parameters of the system.                    bet. The user activates the keys by looking at them for a set duration
                                                                             (i.e., dwell time activation). StarGazer [Hansen et al. 2008] is a 3D
2.3.2 Image Processing                                                       interface that uses pan/zoom to enlarge objects of interest, thereby
                                                                             making them easier to select. The full alphabet is visible in the ini-
The ITU Gaze Tracker supports tracking the pupil center and one or           tial state of the system. The user looks at a desired character and
two corneal reflections (i.e., glints). The software has been devel-          zooms towards it. Selections are made when passing through the
oped to work with low-cost webcams with built-in infrared light.             character.
However, initial support for remote setups that make use of more
expensive video cameras is also included in the software.                    3.1   Experimental Procedure
The experiments presented in this paper make use of the webcam-
based setup shown in Figure 1. In this configuration, the software            Seven participants (2 female and 5 male, ranging from 20 to 35
employs the pupil center to estimate gaze. The pupil center is calcu-        years) participated in the experiment. All participants had nor-
lated by thresholding the image and extracting points in the contour         mal or corrected-to-normal vision. Four of them had prior expe-
between pupil and iris. These points are then fitted to an ellipse            rience with gaze interaction. The typing interfaces (GazeTalk and
using a RANSAC procedure that eliminates possible outliers. Fig-             StarGazer) were presented on a 17” monitor with a resolution of
ure 2 shows a screenshot of the application with the setup window            1280×1024. Participants used the webcam-based gaze tracker to
open. The center of the pupil is estimated and a cross-hair is drawn         interact with the typing applications, and, in the conditions where it
on the image. The picture box shows a processed image of the eye             was required, an optical mouse to perform selections. Participants
and gives an indication of the quality of the tracking.                      sat approximately 60 cm away from the screen. Word prediction
                                                                             was deactivated on both systems but the built-in letter prediction in
                                                                             GazeTalk showed the six most likely letters on the main interface.
                                                                             Participants were asked to type a total of 60 sentences (with an av-
                                                                             erage of 25 characters each) as quickly and accurately as possi-
                                                                             ble. The sentences were Danish translations of the phrase set by
                                                                             Mackenzie and Soukoreff [2003]. No participant entered the same
                                                                             sentence more than once. After completing a sentence, the next was
                                                                             presented. A break was offered after each block of five sentences.
                                                                             Before each block the participants were encouraged to adjust the
                                                                             activation time. At the end of each block the subjects were asked
                                                                             to evaluate the physical strain on eyes, mouth/jaw, face, neck and
                                                                             head on a seven-point scale.
Figure 2: Interface of the ITU Gaze Tracker showing a real time              The within-subjects experiment tested the following four condi-
processed picture of the user’s eye and the setup window where gaze          tions repeated over three sessions: StarGazer with zoom activation,
tracking parameters can be set.                                              StarGazer with mouse click activation, GazeTalk with dwell time


                                                                        78
activation and GazeTalk with mouse click activation. A total of 420          In our experiment the final typing speed for GazeTalk with dwell
sentences were typed (7 participants × 3 sessions × 4 conditions             time activation was 6.56 WPM in the third session. This compares
× 5 sentences). The conditions were counterbalanced using a bal-             to the typing speed of 6.26 WPM for equally trained subjects in
anced Latin square design to neutralize learning effects.                    the experiment by Hansen et al. [2004] who used a commercial
                                                                             gaze tracking system. Presumably, the noise tolerance given by
The dependent variables were typing speed, measured in words per             GazeTalk’s big buttons allows to remove any possible difference
minute (WPM), and accuracy, measured as the minimum string dis-              between the commercial gaze trackers and the low-cost ITU Gaze
tance (MSD). The MSD expresses the minimum number of basic                   Tracker. Actually, in the present experiment the participants were a
actions (insertions, deletions and substitutions) between the pro-           bit faster than in the previous experiment from 2004 where a com-
duced sentence and the target sentence [Levenshtein 1966].                   mercial system was used. The slight difference may be explained
                                                                             by a small difference in dwell time: in the experiment by Hansen
3.2   Results                                                                et al. [2004] this was set to 500 ms while the subjects in our exper-
                                                                             iment adjusted it to an average of 403 ms. Text-entry speed in the
Sentences with an MSD value above 6 (49 out of the total of                  present experiment is much lower than in other gaze typing studies,
420) were considered outliers and removed prior to analysis. One-            where typing speeds of 15 WPM are reported (e.g., [Tuisku et al.
way ANOVAs were conducted to compare the effects of the differ-              2008]). However, the low-cost gaze tracker would not work well
ent typing methods (StarGazer zoom, StarGazer mouse, GazeTalk                with the small on-screen keyboard buttons used in the other exper-
dwell, GazeTalk mouse).                                                      iments, which were all done with commercial systems.
The results obtained for WPM and MSD are shown in Figure 3.                  After calibration of the mouth-mounted gaze tracker, even the
The grand mean for WPM was 5.02. There was a significant effect               slightest head movement introduced a large offset relative to the
of typing method on WPM, F(3, 18) = 30.67, p < 0.05. Both of                 screen. Since the camera is mounted on the user’s head, the two will
the GazeTalk conditions showed significantly higher WPM than the              move together. Therefore, a relative movement of the head/camera
StarGazer conditions in an LSD post-hoc analysis. GazeTalk with              with respect to the screen will take place under head movements.
click activation had a mean WPM = 6.78 (SD = 2.72), and a mean               As a result, when the user maintains his gaze on the same point on
WPM = 5.62 (SD = 1.98) with dwell activation. StarGazer with                 the screen and moves the head, the estimated gaze coordinates will
click had a mean WPM = 3.91 (SD = 1.63), and a mean WPM =                    be affected. The participants in our experiments got used to this
3.56 (SD = 1.58) with zoom activation. The grand mean for MSD                behavior rather quickly and started to take advantage of the ability
errors was 1.61% of all characters typed. Typing method did not              to adjust for offsets by head movements.
have a significant effect on MSD error, F (3, 18) = 0.56, p > 0.05.
                                                                             Carrying the camera and biting the balsa wood felt uncomfortable,
                                                                             at least in the beginning. Eventually, it just came to rest in the bite
                                                                             marks and then it did not feel strenuous. However, it prevents the
                                                                             user from talking and it looks quite strange. Hence, in the current
                                                                             version we do not consider mouth mounting to be feasible in social
                                                                             settings outside the private home or the experimental lab. However,
                                                                             reduced size and weight of future webcams might make it possible
                                                                             to mount them firmly on for example eyeglass frames.

                                                                             4    Case Study
                                                                             The main objective of the case study was to see if the low-cost gaze
                                                                             tracker was flexible enough to accommodate individual setups. The
                                                                             system was set up in the living room of Birger Bergmann Jeppesen,
                                                                             a person who has been living with ALS for more than 10 years
Figure 3: Words per minute and MSD error for each typing method.             and has been gaze typing for more than 5 years. The interface was
Error bars show the standard error of the mean.                              projected onto a wall and the camera was placed on an arm close
                                                                             to his eye. Birger’s wheelchair was placed in front of the wall,
The average settings chosen by the participants in the last session          approximately 3 meters away. Figure 4 shows the setup.
were 1375 ms for StarGazer with zoom activation and 403 ms for               The trial revealed that completing a calibration was consistently
GazeTalk with dwell time activation.                                         achievable and subsequently Birger was able to interact with
Participants reported biting the piece of balsa wood to be quite un-         GazeTalk and produce text. The typing speed was very low - less
pleasant after the first session, with an average value of 2.4 on a           than a word per minute. However, it must be noted that although
scale from 1 (“most uncomfortable”) to 7 (“no strain at all”). How-          Birger can be considered an expert user of GazeTalk, he is in a late
ever, by session 3 this rating had gone up to 4.6, indicating only           stage of ALS and does not have good control of his eye movements
moderate discomfort.                                                         anymore. Furthermore, he no longer has the ability to blink, and the
                                                                             LEDs mounted on the camera would dry out his eyes after a while.
                                                                             Therefore, they needed to be moistened by the helper a few times.
3.3   Discussion of the Results
                                                                             When asked after the trial what he thought the system could be
Gaze typing with the low-cost gaze tracker was possible in both              used for, he answered “For poor people”, referring to the fact that
of the noise tolerant interfaces. The relatively small amount of er-         this approach would be beneficial for users who cannot afford an
rors in the final sentences indicates that participants controlled the        expensive system. His wife, who was present during the trial, was
applications properly and were able to correct their mistakes ade-           very enthusiastic about the setup. She liked that her husband could
quately. This suggests that a low-cost, webcam-based gaze tracker            now be part of conversations around the living room table and ev-
holds potential for gaze interaction with interfaces that have rather        erybody would be able to see what he was typing, thereby eliminat-
large targets, as is the case of GazeTalk and StarGazer.                     ing the need for a synthetic voice. Furthermore, the disassociation


                                                                        79
As of January 2009, more than 3500 people have downloaded the
                                                                               software, showing a high interest in the technology. Furthermore,
                                                                               more than 200 people have taken part in on-line discussions on how
                                                                               to use the software, and the community is collaborating in the de-
                                                                               velopment of the new version. Thanks to the simplicity of the hard-
                                                                               ware setup and usage of low-cost components, we expect more en-
                                                                               thusiasts to become interested in the project.

                                                                               Acknowledgments
                                                                               This work was supported by the COGAIN European Network of
                                                                               Excellence, funded under the FP6/IST program of the European
                                                                               Commission. Thanks to the anonymous people who have con-
                                                                               tributed to the further development of the source code.

                                                                               References
                                                                               BABCOCK , J. S., AND P ELZ , J. B. 2004. Building a lightweight
Figure 4: The typing application (GazeTalk) is projected onto the
                                                                                 eyetracking headgear. In Proceedings of the 2004 symposium
wall. Birger sits on his wheelchair approximately 3 meters away.
                                                                                 on Eye tracking research & applications, ACM, San Antonio,
                                                                                 Texas, 109–114.

of the camera with respect to the screen gives freedom in regard               H ANSEN , J. P., T ORNING , K., J OHANSEN , A. S., I TOH , K., AND
to screen format, thereby allowing a wall projection or a regular                 AOKI , H. 2004. Gaze typing compared with input by head and
monitor to be controlled with the eyes.                                           hand. In Proceedings of the 2004 symposium on Eye tracking
                                                                                  research & applications, ACM, San Antonio, Texas, 131–138.
This initial user trial highlighted some issues regarding the setup.
First, care is needed when placing the camera in a fixed position               H ANSEN , D. W., S KOVSGAARD , H. H. T., H ANSEN , J. P., AND
so that it does not block the field of view of the user. Second, the               M OLLENBACH , E. 2008. Noise tolerant selection by gaze-
camera was sensitive to light; drawing the blinds helped increase                 controlled pan and zoom in 3D. In Proceedings of the 2008
the stability. There is still a need for more robustness and flexibil-             symposium on Eye tracking research & applications, ACM, Sa-
ity but this case study demonstrated the potential of low-cost gaze               vannah, Georgia, 205–212.
trackers.                                                                      L EVENSHTEIN , V. 1966. Binary codes capable of correcting dele-
                                                                                  tions, insertions, and reversals. Soviet Physics Doklady 10, 6,
5    Conclusions and Future Work                                                  707–710.
                                                                               M AC K ENZIE , I. S., AND S OUKOREFF , R. W. 2003. Phrase sets
A low-cost, open-source gaze tracker has been described and eval-                for evaluating text entry techniques. In CHI ’03 extended ab-
uated in this paper. The system offers a high flexibility regarding               stracts on Human factors in computing systems, ACM, Ft. Laud-
both the software, which can be easily extended by modifying the                 erdale, Florida, USA, 754–755.
source code, and the hardware, which can be composed of different
cameras and light sources.                                                                              ¨ ¨
                                                                               M AJARANTA , P., AND R AIH A , K. 2002. Twenty years of eye
                                                                                 typing: systems and design issues. In Proceedings of the 2002
The results obtained in the evaluation indicate that a gaze tracking             symposium on Eye tracking research & applications, ACM, New
system built from inexpensive components can be robust and accu-                 Orleans, Louisiana, 15–22.
rate enough to make gaze typing possible for people who cannot
afford the more expensive commercial systems. The user interface               M ORIMOTO , C., KOONS , D., A MIT, A., F LICKNER , M., AND
needs to have big buttons, which will reduce the productivity, but               Z HAI , S. 1999. Keeping an eye for HCI. In Proceedings of the
typing speeds at more than 5 words per minute are found to be                    XII Symposium on Computer Graphics and Image Processing,
achievable when the typing interface is tolerant to noise.                       1999., 171–176.
The main limitation of the webcam-based system is the lack of                  M ULVEY, F., V ILLANUEVA , A., S LINEY, D., L ANGE , R.,
head-pose invariance. However, some participants would often use                 C OTMORE , S., AND D ONEGAN , M.           Exploration of
this as a feature and compensate for calibration offsets by making               safety issues in eyetracking. In COGAIN Deliverable 5.4.
small adjustments of their head position at each selection. Never-               http://www.cogain.org/w/images/c/c8/COGAIN-D5.4.pdf.
theless, in normal everyday usage it might be difficult to keep the
                                                                               S AN AGUSTIN , J., S KOVSGAARD , H., H ANSEN , J. P., AND
head still for long periods of time.
                                                                                  H ANSEN , D. W. 2009. Low-cost gaze interaction: ready to
One situation where the head-pose invariance does not play a role                 deliver the promises. In Proceedings of the 27th international
is when the user is not able to move the body, for example due to                 conference extended abstracts on Human factors in computing
a locked-in syndrome. The webcam can be mounted on the user’s                     systems, ACM, Boston, MA, USA, 4453–4458.
wheelchair or bed and placed close to the eye. Once the calibra-                                                                        ¨ ¨
                                                                               T UISKU , O., M AJARANTA , P., I SOKOSKI , P., AND R AIH A , K.
tion procedure is completed, the accuracy is maintained due to the                2008. Now dasher! dash away!: longitudinal study of fast text
immobility of the user. The pilot study described in this paper in-               entry by eye gaze. In Proceedings of the 2008 symposium on Eye
dicates that this is a feasible solution; however, a more thorough                tracking research & applications, ACM, Savannah, Georgia.
experiment will be carried out in the busy environment of a hospi-
tal or care home to see if the user will in fact not move and to see if               ´
                                                                               Z IELI NSKI , P., 2007. Opengazer: opensource gaze tracker for ordi-
it is realistic to assume that the camera stays fixed.                             nary webcams. http://www.inference.phy.cam.ac.uk/opengazer/.


                                                                          80

Weitere ähnliche Inhalte

Andere mochten auch

Sentencia an rev sal en ggaa
Sentencia an rev sal en ggaaSentencia an rev sal en ggaa
Sentencia an rev sal en ggaa
oscargaliza
 
Process Consulting Perspective
Process Consulting PerspectiveProcess Consulting Perspective
Process Consulting Perspective
Vibhanshu Sharma
 
Правила и Условия Программы
Правила и Условия ПрограммыПравила и Условия Программы
Правила и Условия Программы
AeroSvit Airlines
 
חטיבת חיר בהתקדמות ורדיפה
חטיבת חיר בהתקדמות ורדיפהחטיבת חיר בהתקדמות ורדיפה
חטיבת חיר בהתקדמות ורדיפה
haimkarel
 
Acta santiago barbanza
Acta santiago barbanzaActa santiago barbanza
Acta santiago barbanza
oscargaliza
 
הכשרת המפקד הנמוך
הכשרת המפקד הנמוךהכשרת המפקד הנמוך
הכשרת המפקד הנמוך
haimkarel
 
Anexo ás normas, calendario previo (aprobado)
Anexo ás normas, calendario previo (aprobado)Anexo ás normas, calendario previo (aprobado)
Anexo ás normas, calendario previo (aprobado)
oscargaliza
 
Balanced Scorecard and Strategy Execution
Balanced Scorecard and Strategy ExecutionBalanced Scorecard and Strategy Execution
Balanced Scorecard and Strategy Execution
Shaji Bhaskaran
 

Andere mochten auch (20)

Sentencia an rev sal en ggaa
Sentencia an rev sal en ggaaSentencia an rev sal en ggaa
Sentencia an rev sal en ggaa
 
Process Consulting Perspective
Process Consulting PerspectiveProcess Consulting Perspective
Process Consulting Perspective
 
ภูมิศาสตร์ของทวีปยุโรป 2.4
ภูมิศาสตร์ของทวีปยุโรป 2.4ภูมิศาสตร์ของทวีปยุโรป 2.4
ภูมิศาสตร์ของทวีปยุโรป 2.4
 
Αγωγή Υγείας - Εφηβικές συζητήσεις και προβληματισμοί - Επίλυση κρίσεων
Αγωγή Υγείας - Εφηβικές συζητήσεις και προβληματισμοί - Επίλυση κρίσεωνΑγωγή Υγείας - Εφηβικές συζητήσεις και προβληματισμοί - Επίλυση κρίσεων
Αγωγή Υγείας - Εφηβικές συζητήσεις και προβληματισμοί - Επίλυση κρίσεων
 
Правила и Условия Программы
Правила и Условия ПрограммыПравила и Условия Программы
Правила и Условия Программы
 
China
ChinaChina
China
 
חטיבת חיר בהתקדמות ורדיפה
חטיבת חיר בהתקדמות ורדיפהחטיבת חיר בהתקדמות ורדיפה
חטיבת חיר בהתקדמות ורדיפה
 
Acta santiago barbanza
Acta santiago barbanzaActa santiago barbanza
Acta santiago barbanza
 
הכשרת המפקד הנמוך
הכשרת המפקד הנמוךהכשרת המפקד הנמוך
הכשרת המפקד הנמוך
 
Statby school 2553_m3_1057012007
Statby school 2553_m3_1057012007Statby school 2553_m3_1057012007
Statby school 2553_m3_1057012007
 
Statby school 2555_m3_1057012007
Statby school 2555_m3_1057012007Statby school 2555_m3_1057012007
Statby school 2555_m3_1057012007
 
สถาปนากรุงรัตนโกสินทร์
สถาปนากรุงรัตนโกสินทร์สถาปนากรุงรัตนโกสินทร์
สถาปนากรุงรัตนโกสินทร์
 
Log
LogLog
Log
 
Anexo ás normas, calendario previo (aprobado)
Anexo ás normas, calendario previo (aprobado)Anexo ás normas, calendario previo (aprobado)
Anexo ás normas, calendario previo (aprobado)
 
Power Poritพุทธประวัติ
Power PoritพุทธประวัติPower Poritพุทธประวัติ
Power Poritพุทธประวัติ
 
Hennessey An Open Source Eye Gaze Interface Expanding The Adoption Of Eye Gaz...
Hennessey An Open Source Eye Gaze Interface Expanding The Adoption Of Eye Gaz...Hennessey An Open Source Eye Gaze Interface Expanding The Adoption Of Eye Gaz...
Hennessey An Open Source Eye Gaze Interface Expanding The Adoption Of Eye Gaz...
 
Grindinger Group Wise Similarity And Classification Of Aggregate Scanpaths
Grindinger Group Wise Similarity And Classification Of Aggregate ScanpathsGrindinger Group Wise Similarity And Classification Of Aggregate Scanpaths
Grindinger Group Wise Similarity And Classification Of Aggregate Scanpaths
 
Exploring Italy
Exploring ItalyExploring Italy
Exploring Italy
 
Boonmak
BoonmakBoonmak
Boonmak
 
Balanced Scorecard and Strategy Execution
Balanced Scorecard and Strategy ExecutionBalanced Scorecard and Strategy Execution
Balanced Scorecard and Strategy Execution
 

Ähnlich wie San Agustin Evaluation Of A Low Cost Open Source Gaze Tracker

Schneider.2011.an open source low-cost eye-tracking system for portable real-...
Schneider.2011.an open source low-cost eye-tracking system for portable real-...Schneider.2011.an open source low-cost eye-tracking system for portable real-...
Schneider.2011.an open source low-cost eye-tracking system for portable real-...
mrgazer
 
Van der kamp.2011.gaze and voice controlled drawing
Van der kamp.2011.gaze and voice controlled drawingVan der kamp.2011.gaze and voice controlled drawing
Van der kamp.2011.gaze and voice controlled drawing
mrgazer
 
Remote sensing and control of an irrigation system using a distributed wirele...
Remote sensing and control of an irrigation system using a distributed wirele...Remote sensing and control of an irrigation system using a distributed wirele...
Remote sensing and control of an irrigation system using a distributed wirele...
nithinreddykaithi
 

Ähnlich wie San Agustin Evaluation Of A Low Cost Open Source Gaze Tracker (20)

Schneider.2011.an open source low-cost eye-tracking system for portable real-...
Schneider.2011.an open source low-cost eye-tracking system for portable real-...Schneider.2011.an open source low-cost eye-tracking system for portable real-...
Schneider.2011.an open source low-cost eye-tracking system for portable real-...
 
Prediction of Wireless Sensor Network and Attack using Machine Learning Techn...
Prediction of Wireless Sensor Network and Attack using Machine Learning Techn...Prediction of Wireless Sensor Network and Attack using Machine Learning Techn...
Prediction of Wireless Sensor Network and Attack using Machine Learning Techn...
 
Object Detection and Localization for Visually Impaired People using CNN
Object Detection and Localization for Visually Impaired People using CNNObject Detection and Localization for Visually Impaired People using CNN
Object Detection and Localization for Visually Impaired People using CNN
 
Advanced Intelligent Video Surveillance System In Elevators By Using OpenCV
Advanced Intelligent Video Surveillance System In Elevators By Using OpenCVAdvanced Intelligent Video Surveillance System In Elevators By Using OpenCV
Advanced Intelligent Video Surveillance System In Elevators By Using OpenCV
 
IRJETDevice for Location Finder and Text Reader for Visually Impaired People
IRJETDevice for Location Finder and Text Reader for Visually Impaired PeopleIRJETDevice for Location Finder and Text Reader for Visually Impaired People
IRJETDevice for Location Finder and Text Reader for Visually Impaired People
 
IRJET- Device for Location Finder and Text Reader for Visually Impaired P...
IRJET-  	  Device for Location Finder and Text Reader for Visually Impaired P...IRJET-  	  Device for Location Finder and Text Reader for Visually Impaired P...
IRJET- Device for Location Finder and Text Reader for Visually Impaired P...
 
Drishyam - Virtual Eye for Blind
Drishyam - Virtual Eye for BlindDrishyam - Virtual Eye for Blind
Drishyam - Virtual Eye for Blind
 
Dog Breed Prediction System (Web)
Dog Breed Prediction System (Web)Dog Breed Prediction System (Web)
Dog Breed Prediction System (Web)
 
Sanjaya: A Blind Assistance System
Sanjaya: A Blind Assistance SystemSanjaya: A Blind Assistance System
Sanjaya: A Blind Assistance System
 
Wireless Network Intrinsic Secrecy
Wireless Network Intrinsic SecrecyWireless Network Intrinsic Secrecy
Wireless Network Intrinsic Secrecy
 
Motion Detection System for Security Using IoT- Survey
Motion Detection System for Security Using IoT- SurveyMotion Detection System for Security Using IoT- Survey
Motion Detection System for Security Using IoT- Survey
 
Object and Currency Detection for the Visually Impaired
Object and Currency Detection for the Visually ImpairedObject and Currency Detection for the Visually Impaired
Object and Currency Detection for the Visually Impaired
 
IRJET-Voice Assisted Blind Stick using Ultrasonic Sensor
IRJET-Voice Assisted Blind Stick using Ultrasonic SensorIRJET-Voice Assisted Blind Stick using Ultrasonic Sensor
IRJET-Voice Assisted Blind Stick using Ultrasonic Sensor
 
Van der kamp.2011.gaze and voice controlled drawing
Van der kamp.2011.gaze and voice controlled drawingVan der kamp.2011.gaze and voice controlled drawing
Van der kamp.2011.gaze and voice controlled drawing
 
INDOOR AND OUTDOOR NAVIGATION ASSISTANCE SYSTEM FOR VISUALLY IMPAIRED PEOPLE ...
INDOOR AND OUTDOOR NAVIGATION ASSISTANCE SYSTEM FOR VISUALLY IMPAIRED PEOPLE ...INDOOR AND OUTDOOR NAVIGATION ASSISTANCE SYSTEM FOR VISUALLY IMPAIRED PEOPLE ...
INDOOR AND OUTDOOR NAVIGATION ASSISTANCE SYSTEM FOR VISUALLY IMPAIRED PEOPLE ...
 
Remote sensing and control of an irrigation system using a distributed wirele...
Remote sensing and control of an irrigation system using a distributed wirele...Remote sensing and control of an irrigation system using a distributed wirele...
Remote sensing and control of an irrigation system using a distributed wirele...
 
Advance Intelligent Video Surveillance System Using OpenCV
Advance Intelligent Video Surveillance System Using OpenCVAdvance Intelligent Video Surveillance System Using OpenCV
Advance Intelligent Video Surveillance System Using OpenCV
 
Control the computer with your eyes
Control the computer with your eyesControl the computer with your eyes
Control the computer with your eyes
 
2012 ieee projects software engineering @ Seabirds ( Trichy, Chennai, Pondich...
2012 ieee projects software engineering @ Seabirds ( Trichy, Chennai, Pondich...2012 ieee projects software engineering @ Seabirds ( Trichy, Chennai, Pondich...
2012 ieee projects software engineering @ Seabirds ( Trichy, Chennai, Pondich...
 
Object Detection Bot
Object Detection BotObject Detection Bot
Object Detection Bot
 

Mehr von Kalle

Mehr von Kalle (20)

Blignaut Visual Span And Other Parameters For The Generation Of Heatmaps
Blignaut Visual Span And Other Parameters For The Generation Of HeatmapsBlignaut Visual Span And Other Parameters For The Generation Of Heatmaps
Blignaut Visual Span And Other Parameters For The Generation Of Heatmaps
 
Zhang Eye Movement As An Interaction Mechanism For Relevance Feedback In A Co...
Zhang Eye Movement As An Interaction Mechanism For Relevance Feedback In A Co...Zhang Eye Movement As An Interaction Mechanism For Relevance Feedback In A Co...
Zhang Eye Movement As An Interaction Mechanism For Relevance Feedback In A Co...
 
Yamamoto Development Of Eye Tracking Pen Display Based On Stereo Bright Pupil...
Yamamoto Development Of Eye Tracking Pen Display Based On Stereo Bright Pupil...Yamamoto Development Of Eye Tracking Pen Display Based On Stereo Bright Pupil...
Yamamoto Development Of Eye Tracking Pen Display Based On Stereo Bright Pupil...
 
Wastlund What You See Is Where You Go Testing A Gaze Driven Power Wheelchair ...
Wastlund What You See Is Where You Go Testing A Gaze Driven Power Wheelchair ...Wastlund What You See Is Where You Go Testing A Gaze Driven Power Wheelchair ...
Wastlund What You See Is Where You Go Testing A Gaze Driven Power Wheelchair ...
 
Vinnikov Contingency Evaluation Of Gaze Contingent Displays For Real Time Vis...
Vinnikov Contingency Evaluation Of Gaze Contingent Displays For Real Time Vis...Vinnikov Contingency Evaluation Of Gaze Contingent Displays For Real Time Vis...
Vinnikov Contingency Evaluation Of Gaze Contingent Displays For Real Time Vis...
 
Urbina Pies With Ey Es The Limits Of Hierarchical Pie Menus In Gaze Control
Urbina Pies With Ey Es The Limits Of Hierarchical Pie Menus In Gaze ControlUrbina Pies With Ey Es The Limits Of Hierarchical Pie Menus In Gaze Control
Urbina Pies With Ey Es The Limits Of Hierarchical Pie Menus In Gaze Control
 
Urbina Alternatives To Single Character Entry And Dwell Time Selection On Eye...
Urbina Alternatives To Single Character Entry And Dwell Time Selection On Eye...Urbina Alternatives To Single Character Entry And Dwell Time Selection On Eye...
Urbina Alternatives To Single Character Entry And Dwell Time Selection On Eye...
 
Tien Measuring Situation Awareness Of Surgeons In Laparoscopic Training
Tien Measuring Situation Awareness Of Surgeons In Laparoscopic TrainingTien Measuring Situation Awareness Of Surgeons In Laparoscopic Training
Tien Measuring Situation Awareness Of Surgeons In Laparoscopic Training
 
Takemura Estimating 3 D Point Of Regard And Visualizing Gaze Trajectories Und...
Takemura Estimating 3 D Point Of Regard And Visualizing Gaze Trajectories Und...Takemura Estimating 3 D Point Of Regard And Visualizing Gaze Trajectories Und...
Takemura Estimating 3 D Point Of Regard And Visualizing Gaze Trajectories Und...
 
Stevenson Eye Tracking With The Adaptive Optics Scanning Laser Ophthalmoscope
Stevenson Eye Tracking With The Adaptive Optics Scanning Laser OphthalmoscopeStevenson Eye Tracking With The Adaptive Optics Scanning Laser Ophthalmoscope
Stevenson Eye Tracking With The Adaptive Optics Scanning Laser Ophthalmoscope
 
Stellmach Advanced Gaze Visualizations For Three Dimensional Virtual Environm...
Stellmach Advanced Gaze Visualizations For Three Dimensional Virtual Environm...Stellmach Advanced Gaze Visualizations For Three Dimensional Virtual Environm...
Stellmach Advanced Gaze Visualizations For Three Dimensional Virtual Environm...
 
Skovsgaard Small Target Selection With Gaze Alone
Skovsgaard Small Target Selection With Gaze AloneSkovsgaard Small Target Selection With Gaze Alone
Skovsgaard Small Target Selection With Gaze Alone
 
Ryan Match Moving For Area Based Analysis Of Eye Movements In Natural Tasks
Ryan Match Moving For Area Based Analysis Of Eye Movements In Natural TasksRyan Match Moving For Area Based Analysis Of Eye Movements In Natural Tasks
Ryan Match Moving For Area Based Analysis Of Eye Movements In Natural Tasks
 
Rosengrant Gaze Scribing In Physics Problem Solving
Rosengrant Gaze Scribing In Physics Problem SolvingRosengrant Gaze Scribing In Physics Problem Solving
Rosengrant Gaze Scribing In Physics Problem Solving
 
Qvarfordt Understanding The Benefits Of Gaze Enhanced Visual Search
Qvarfordt Understanding The Benefits Of Gaze Enhanced Visual SearchQvarfordt Understanding The Benefits Of Gaze Enhanced Visual Search
Qvarfordt Understanding The Benefits Of Gaze Enhanced Visual Search
 
Prats Interpretation Of Geometric Shapes An Eye Movement Study
Prats Interpretation Of Geometric Shapes An Eye Movement StudyPrats Interpretation Of Geometric Shapes An Eye Movement Study
Prats Interpretation Of Geometric Shapes An Eye Movement Study
 
Porta Ce Cursor A Contextual Eye Cursor For General Pointing In Windows Envir...
Porta Ce Cursor A Contextual Eye Cursor For General Pointing In Windows Envir...Porta Ce Cursor A Contextual Eye Cursor For General Pointing In Windows Envir...
Porta Ce Cursor A Contextual Eye Cursor For General Pointing In Windows Envir...
 
Pontillo Semanti Code Using Content Similarity And Database Driven Matching T...
Pontillo Semanti Code Using Content Similarity And Database Driven Matching T...Pontillo Semanti Code Using Content Similarity And Database Driven Matching T...
Pontillo Semanti Code Using Content Similarity And Database Driven Matching T...
 
Park Quantification Of Aesthetic Viewing Using Eye Tracking Technology The In...
Park Quantification Of Aesthetic Viewing Using Eye Tracking Technology The In...Park Quantification Of Aesthetic Viewing Using Eye Tracking Technology The In...
Park Quantification Of Aesthetic Viewing Using Eye Tracking Technology The In...
 
Palinko Estimating Cognitive Load Using Remote Eye Tracking In A Driving Simu...
Palinko Estimating Cognitive Load Using Remote Eye Tracking In A Driving Simu...Palinko Estimating Cognitive Load Using Remote Eye Tracking In A Driving Simu...
Palinko Estimating Cognitive Load Using Remote Eye Tracking In A Driving Simu...
 

San Agustin Evaluation Of A Low Cost Open Source Gaze Tracker

  • 1. Evaluation of a Low-Cost Open-Source Gaze Tracker Javier San Agustin∗ Emilie Mollenbach‡ Maria Barret§ Henrik Skovsgaard† Loughborough University Martin Tall¶ IT University of Copenhagen Dan Witzner Hansen John Paulin Hansen∗∗ IT University of Copenhagen Abstract However, this project is in a very early stage of development, and has not been active since 2007. This paper presents a low-cost gaze tracking system that is based In this paper we introduce the ITU Gaze Tracker (available at on a webcam mounted close to the user’s eye. The performance http://www.gazegroup.org/downloads/23-gazetracker), a low-cost of the gaze tracker was evaluated in an eye-typing task using two system that was released as open source in Spring 2009 [San different typing applications. Participants could type between 3.56 Agustin et al. 2009], and we present an evaluation of a head- and 6.78 words per minute, depending on the typing system used. mounted setup that employs a webcam with built-in infrared light. A pilot study to assess the usability of the system was also car- Such setup does not require extra light sources, and therefore the ried out in the home of a user with severe motor impairments. The system is very easy to build. However, the system is not head-pose user successfully typed on a wall-projected interface using his eye invariant. The evaluation was carried out using two different eye- movements. typing applications designed to handle a noisy input, GazeTalk and StarGazer. A pilot study with a person in the latest stages of ALS CR Categories: H.5.2 [Input devices and strategies]: Evalua- was also conducted. tion/Methodology Keywords: low cost, gaze interaction, gaze typing, off-the-shelf 2 The ITU Gaze Tracker components, augmentative and alternative communication In order to provide an accessible gaze tracking system, the ITU Gaze Tracker was developed. The design considerations and the 1 Introduction hardware and software are presented in this section. The cost of commercial gaze tracking systems often exceeds 2.1 Design Considerations $10.000, preventing access to this technology by many people who need it for daily communication. Using low-cost components ben- The approach taken to develop the ITU Gaze Tracker aimed to pro- efits users who cannot afford an expensive commercial system, and vide a fully accessible system, with a set of design requirements: it gives them a chance to try gaze interaction for the first time with- out making a big investment. Furthermore, open source solutions 1. The gaze tracker should be robust and accurate enough to might provide a starting point for enthusiasts, who can contribute work with at least one gaze-communication system. to an active project and adapt the software to fit specific needs. 2. Use of low-cost off-the-shelf components. The hardware em- ployed should be available in any electronics store or at on- Commercial gaze tracking systems have been available for more line shops to allow for easy acquisition and replacement. Fur- than 15 years, but the technology is still a high priced niche. A few thermore, no hardware modifications should be needed. systems built using off-the-shelf components have been presented. 3. The user should be given flexibility to place the different com- Babcock and Pelz [2004] developed a head-mounted system that ponents (camera, infrared lights, computer display) at various uses two cameras mounted on a pair of safety glasses, one to track locations to fit specific needs. For instance, mounting the dis- the eye and the other to record the scene in front of the viewer. play on a wheel chair, moving it to a table or having it acces- However, building this system requires an in-depth understanding sible in bed should make no difference. of electronics and the software only runs on the Linux operating 4. Open-source software. Developing an efficient gaze tracker system, thereby preventing most potential users from assembling from low-cost components is a huge endeavor. No single their own system. The Opengazer eye tracker uses a web camera to group of developers is likely to come up with the ultimate track the user’s eyes and is rather simple to build [Zieli´ ski 2007]. n solution. Open source allows anybody to improve and modify ∗ e-mail: the source code to fit specific needs. javier@itu.dk † e-mail: hhje@itu.dk ‡ e-mail: author@affiliation.com 2.2 Hardware § e-mail: @affiliation.com ¶ e-mail: mtall@itu.dk Standard webcams usually have a low resolution and a broad field e-mail: witzner@itu.dk of view, but by placing the camera close to the user’s eye we can ∗∗ e-mail: paulin@itu.dk obtain images of sufficient quality. Figure 1 shows the gaze tracker Copyright © 2010 by the Association for Computing Machinery, Inc. in two setups, (a) mounted on a piece of inexpensive balsa wood Permission to make digital or hard copies of part or all of this work for personal or that the user bites on in order to fix the webcam in front of the eye, classroom use is granted without fee provided that copies are not made or distributed and (b) mounted on a flexible arm (e.g., an office lamp). for commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be The setup displayed in Figure 1 employs a Sandberg Nightvision 2 honored. Abstracting with credit is permitted. To copy otherwise, to republish, to post on webcam, which costs around $20. The camera is equipped with 6 servers, or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from Permissions Dept, ACM Inc., fax +1 (212) 869-0481 or e-mail built-in infrared LEDs that create a dark pupil effect. The resolution permissions@acm.org. is 640×480 pixels and the frame rate is 30 fps. The weight of the ETRA 2010, Austin, TX, March 22 – 24, 2010. © 2010 ACM 978-1-60558-994-7/10/0003 $10.00 77
  • 2. 2.3.3 Gaze Estimation Our flexibility requirements introduce a high uncertainty in the lo- cations of the camera and screen with respect to the user. There- fore, the gaze estimation technique employed must be as generic as possible, and should not assume any specific configuration of the hardware components. The gaze estimation method implemented is an interpolation technique that uses a pair of second-order polyno- (a) (b) mials (one for each screen axis) [Morimoto et al. 1999]. A 12-point calibration procedure allows to calculate the coefficients that map Figure 1: Mounting solutions of the ITU Gaze Tracker: (a) on a the center of the pupil to the gaze coordinates on the screen. The piece of inexpensive balsa wood; (b) on the arm of an office lamp. calibration procedure takes approximately 20 seconds. The system was demonstrated at the CHI 2009 Conference. A total of 28 delegates tried out the system mounted on a piece of balsa camera is 200g. The amount of light shone into the eyes is not wood. 22 had normal vision or wore contact lenses, and 6 wore considered to be harmful [Mulvey et al. ]. glasses. All 28 were able to get a successful calibration and control the cursor with their eyes. 2.3 Software 3 Gaze Typing with the ITU GazeTracker 2.3.1 Software Architecture Text input is a very important task for people who use gaze to com- The gaze tracking software is developed in C#, and employs municate. The design of typing systems is one of the most well OpenCV for image processing. There are three main components: researched areas within the field of gaze interaction [Majaranta and (1) The gaze tracking library, which implements all the methods to R¨ ih¨ 2002]. Gaze typing systems are often evaluated experimen- a a control a tracker such as extracting eye features, run a calibration tally with standardized performance measurements such as words procedure, estimate the gaze coordinates and detect the type of eye per minute (WPM) and error rates. movement (i.e., fixation or saccade); (2) The camera class, respon- sible for initializing a generic camera and grabbing images that are The two eye typing applications used in this experiment have dif- then processed by the gaze-tracking library; and (3) The user in- ferent approaches to handle noise. GazeTalk [Hansen et al. 2004] terface, which provides the communication with the gaze-tracking has 10 large buttons with a hierarchical organization of the alpha- library to set up the different parameters of the system. bet. The user activates the keys by looking at them for a set duration (i.e., dwell time activation). StarGazer [Hansen et al. 2008] is a 3D 2.3.2 Image Processing interface that uses pan/zoom to enlarge objects of interest, thereby making them easier to select. The full alphabet is visible in the ini- The ITU Gaze Tracker supports tracking the pupil center and one or tial state of the system. The user looks at a desired character and two corneal reflections (i.e., glints). The software has been devel- zooms towards it. Selections are made when passing through the oped to work with low-cost webcams with built-in infrared light. character. However, initial support for remote setups that make use of more expensive video cameras is also included in the software. 3.1 Experimental Procedure The experiments presented in this paper make use of the webcam- based setup shown in Figure 1. In this configuration, the software Seven participants (2 female and 5 male, ranging from 20 to 35 employs the pupil center to estimate gaze. The pupil center is calcu- years) participated in the experiment. All participants had nor- lated by thresholding the image and extracting points in the contour mal or corrected-to-normal vision. Four of them had prior expe- between pupil and iris. These points are then fitted to an ellipse rience with gaze interaction. The typing interfaces (GazeTalk and using a RANSAC procedure that eliminates possible outliers. Fig- StarGazer) were presented on a 17” monitor with a resolution of ure 2 shows a screenshot of the application with the setup window 1280×1024. Participants used the webcam-based gaze tracker to open. The center of the pupil is estimated and a cross-hair is drawn interact with the typing applications, and, in the conditions where it on the image. The picture box shows a processed image of the eye was required, an optical mouse to perform selections. Participants and gives an indication of the quality of the tracking. sat approximately 60 cm away from the screen. Word prediction was deactivated on both systems but the built-in letter prediction in GazeTalk showed the six most likely letters on the main interface. Participants were asked to type a total of 60 sentences (with an av- erage of 25 characters each) as quickly and accurately as possi- ble. The sentences were Danish translations of the phrase set by Mackenzie and Soukoreff [2003]. No participant entered the same sentence more than once. After completing a sentence, the next was presented. A break was offered after each block of five sentences. Before each block the participants were encouraged to adjust the activation time. At the end of each block the subjects were asked to evaluate the physical strain on eyes, mouth/jaw, face, neck and head on a seven-point scale. Figure 2: Interface of the ITU Gaze Tracker showing a real time The within-subjects experiment tested the following four condi- processed picture of the user’s eye and the setup window where gaze tions repeated over three sessions: StarGazer with zoom activation, tracking parameters can be set. StarGazer with mouse click activation, GazeTalk with dwell time 78
  • 3. activation and GazeTalk with mouse click activation. A total of 420 In our experiment the final typing speed for GazeTalk with dwell sentences were typed (7 participants × 3 sessions × 4 conditions time activation was 6.56 WPM in the third session. This compares × 5 sentences). The conditions were counterbalanced using a bal- to the typing speed of 6.26 WPM for equally trained subjects in anced Latin square design to neutralize learning effects. the experiment by Hansen et al. [2004] who used a commercial gaze tracking system. Presumably, the noise tolerance given by The dependent variables were typing speed, measured in words per GazeTalk’s big buttons allows to remove any possible difference minute (WPM), and accuracy, measured as the minimum string dis- between the commercial gaze trackers and the low-cost ITU Gaze tance (MSD). The MSD expresses the minimum number of basic Tracker. Actually, in the present experiment the participants were a actions (insertions, deletions and substitutions) between the pro- bit faster than in the previous experiment from 2004 where a com- duced sentence and the target sentence [Levenshtein 1966]. mercial system was used. The slight difference may be explained by a small difference in dwell time: in the experiment by Hansen 3.2 Results et al. [2004] this was set to 500 ms while the subjects in our exper- iment adjusted it to an average of 403 ms. Text-entry speed in the Sentences with an MSD value above 6 (49 out of the total of present experiment is much lower than in other gaze typing studies, 420) were considered outliers and removed prior to analysis. One- where typing speeds of 15 WPM are reported (e.g., [Tuisku et al. way ANOVAs were conducted to compare the effects of the differ- 2008]). However, the low-cost gaze tracker would not work well ent typing methods (StarGazer zoom, StarGazer mouse, GazeTalk with the small on-screen keyboard buttons used in the other exper- dwell, GazeTalk mouse). iments, which were all done with commercial systems. The results obtained for WPM and MSD are shown in Figure 3. After calibration of the mouth-mounted gaze tracker, even the The grand mean for WPM was 5.02. There was a significant effect slightest head movement introduced a large offset relative to the of typing method on WPM, F(3, 18) = 30.67, p < 0.05. Both of screen. Since the camera is mounted on the user’s head, the two will the GazeTalk conditions showed significantly higher WPM than the move together. Therefore, a relative movement of the head/camera StarGazer conditions in an LSD post-hoc analysis. GazeTalk with with respect to the screen will take place under head movements. click activation had a mean WPM = 6.78 (SD = 2.72), and a mean As a result, when the user maintains his gaze on the same point on WPM = 5.62 (SD = 1.98) with dwell activation. StarGazer with the screen and moves the head, the estimated gaze coordinates will click had a mean WPM = 3.91 (SD = 1.63), and a mean WPM = be affected. The participants in our experiments got used to this 3.56 (SD = 1.58) with zoom activation. The grand mean for MSD behavior rather quickly and started to take advantage of the ability errors was 1.61% of all characters typed. Typing method did not to adjust for offsets by head movements. have a significant effect on MSD error, F (3, 18) = 0.56, p > 0.05. Carrying the camera and biting the balsa wood felt uncomfortable, at least in the beginning. Eventually, it just came to rest in the bite marks and then it did not feel strenuous. However, it prevents the user from talking and it looks quite strange. Hence, in the current version we do not consider mouth mounting to be feasible in social settings outside the private home or the experimental lab. However, reduced size and weight of future webcams might make it possible to mount them firmly on for example eyeglass frames. 4 Case Study The main objective of the case study was to see if the low-cost gaze tracker was flexible enough to accommodate individual setups. The system was set up in the living room of Birger Bergmann Jeppesen, a person who has been living with ALS for more than 10 years Figure 3: Words per minute and MSD error for each typing method. and has been gaze typing for more than 5 years. The interface was Error bars show the standard error of the mean. projected onto a wall and the camera was placed on an arm close to his eye. Birger’s wheelchair was placed in front of the wall, The average settings chosen by the participants in the last session approximately 3 meters away. Figure 4 shows the setup. were 1375 ms for StarGazer with zoom activation and 403 ms for The trial revealed that completing a calibration was consistently GazeTalk with dwell time activation. achievable and subsequently Birger was able to interact with Participants reported biting the piece of balsa wood to be quite un- GazeTalk and produce text. The typing speed was very low - less pleasant after the first session, with an average value of 2.4 on a than a word per minute. However, it must be noted that although scale from 1 (“most uncomfortable”) to 7 (“no strain at all”). How- Birger can be considered an expert user of GazeTalk, he is in a late ever, by session 3 this rating had gone up to 4.6, indicating only stage of ALS and does not have good control of his eye movements moderate discomfort. anymore. Furthermore, he no longer has the ability to blink, and the LEDs mounted on the camera would dry out his eyes after a while. Therefore, they needed to be moistened by the helper a few times. 3.3 Discussion of the Results When asked after the trial what he thought the system could be Gaze typing with the low-cost gaze tracker was possible in both used for, he answered “For poor people”, referring to the fact that of the noise tolerant interfaces. The relatively small amount of er- this approach would be beneficial for users who cannot afford an rors in the final sentences indicates that participants controlled the expensive system. His wife, who was present during the trial, was applications properly and were able to correct their mistakes ade- very enthusiastic about the setup. She liked that her husband could quately. This suggests that a low-cost, webcam-based gaze tracker now be part of conversations around the living room table and ev- holds potential for gaze interaction with interfaces that have rather erybody would be able to see what he was typing, thereby eliminat- large targets, as is the case of GazeTalk and StarGazer. ing the need for a synthetic voice. Furthermore, the disassociation 79
  • 4. As of January 2009, more than 3500 people have downloaded the software, showing a high interest in the technology. Furthermore, more than 200 people have taken part in on-line discussions on how to use the software, and the community is collaborating in the de- velopment of the new version. Thanks to the simplicity of the hard- ware setup and usage of low-cost components, we expect more en- thusiasts to become interested in the project. Acknowledgments This work was supported by the COGAIN European Network of Excellence, funded under the FP6/IST program of the European Commission. Thanks to the anonymous people who have con- tributed to the further development of the source code. References BABCOCK , J. S., AND P ELZ , J. B. 2004. Building a lightweight Figure 4: The typing application (GazeTalk) is projected onto the eyetracking headgear. In Proceedings of the 2004 symposium wall. Birger sits on his wheelchair approximately 3 meters away. on Eye tracking research & applications, ACM, San Antonio, Texas, 109–114. of the camera with respect to the screen gives freedom in regard H ANSEN , J. P., T ORNING , K., J OHANSEN , A. S., I TOH , K., AND to screen format, thereby allowing a wall projection or a regular AOKI , H. 2004. Gaze typing compared with input by head and monitor to be controlled with the eyes. hand. In Proceedings of the 2004 symposium on Eye tracking research & applications, ACM, San Antonio, Texas, 131–138. This initial user trial highlighted some issues regarding the setup. First, care is needed when placing the camera in a fixed position H ANSEN , D. W., S KOVSGAARD , H. H. T., H ANSEN , J. P., AND so that it does not block the field of view of the user. Second, the M OLLENBACH , E. 2008. Noise tolerant selection by gaze- camera was sensitive to light; drawing the blinds helped increase controlled pan and zoom in 3D. In Proceedings of the 2008 the stability. There is still a need for more robustness and flexibil- symposium on Eye tracking research & applications, ACM, Sa- ity but this case study demonstrated the potential of low-cost gaze vannah, Georgia, 205–212. trackers. L EVENSHTEIN , V. 1966. Binary codes capable of correcting dele- tions, insertions, and reversals. Soviet Physics Doklady 10, 6, 5 Conclusions and Future Work 707–710. M AC K ENZIE , I. S., AND S OUKOREFF , R. W. 2003. Phrase sets A low-cost, open-source gaze tracker has been described and eval- for evaluating text entry techniques. In CHI ’03 extended ab- uated in this paper. The system offers a high flexibility regarding stracts on Human factors in computing systems, ACM, Ft. Laud- both the software, which can be easily extended by modifying the erdale, Florida, USA, 754–755. source code, and the hardware, which can be composed of different cameras and light sources. ¨ ¨ M AJARANTA , P., AND R AIH A , K. 2002. Twenty years of eye typing: systems and design issues. In Proceedings of the 2002 The results obtained in the evaluation indicate that a gaze tracking symposium on Eye tracking research & applications, ACM, New system built from inexpensive components can be robust and accu- Orleans, Louisiana, 15–22. rate enough to make gaze typing possible for people who cannot afford the more expensive commercial systems. The user interface M ORIMOTO , C., KOONS , D., A MIT, A., F LICKNER , M., AND needs to have big buttons, which will reduce the productivity, but Z HAI , S. 1999. Keeping an eye for HCI. In Proceedings of the typing speeds at more than 5 words per minute are found to be XII Symposium on Computer Graphics and Image Processing, achievable when the typing interface is tolerant to noise. 1999., 171–176. The main limitation of the webcam-based system is the lack of M ULVEY, F., V ILLANUEVA , A., S LINEY, D., L ANGE , R., head-pose invariance. However, some participants would often use C OTMORE , S., AND D ONEGAN , M. Exploration of this as a feature and compensate for calibration offsets by making safety issues in eyetracking. In COGAIN Deliverable 5.4. small adjustments of their head position at each selection. Never- http://www.cogain.org/w/images/c/c8/COGAIN-D5.4.pdf. theless, in normal everyday usage it might be difficult to keep the S AN AGUSTIN , J., S KOVSGAARD , H., H ANSEN , J. P., AND head still for long periods of time. H ANSEN , D. W. 2009. Low-cost gaze interaction: ready to One situation where the head-pose invariance does not play a role deliver the promises. In Proceedings of the 27th international is when the user is not able to move the body, for example due to conference extended abstracts on Human factors in computing a locked-in syndrome. The webcam can be mounted on the user’s systems, ACM, Boston, MA, USA, 4453–4458. wheelchair or bed and placed close to the eye. Once the calibra- ¨ ¨ T UISKU , O., M AJARANTA , P., I SOKOSKI , P., AND R AIH A , K. tion procedure is completed, the accuracy is maintained due to the 2008. Now dasher! dash away!: longitudinal study of fast text immobility of the user. The pilot study described in this paper in- entry by eye gaze. In Proceedings of the 2008 symposium on Eye dicates that this is a feasible solution; however, a more thorough tracking research & applications, ACM, Savannah, Georgia. experiment will be carried out in the busy environment of a hospi- tal or care home to see if the user will in fact not move and to see if ´ Z IELI NSKI , P., 2007. Opengazer: opensource gaze tracker for ordi- it is realistic to assume that the camera stays fixed. nary webcams. http://www.inference.phy.cam.ac.uk/opengazer/. 80