Eeg Signal Processing For Eye Tracking
Eeg Signal Processing For Eye Tracking
ABSTRACT 2. BACKGROUND
Head-mounted Video-Oculography (VOG) eye tracking is vi- 2.1. Eye Movement and Tracking
sually intrusive due to a camera in the peripheral view. Elec-
trooculography (EOG) eye tracking is socially intrusive be- The eye tracking problem can be defined as estimating a per-
cause of face-mounted electrodes. In this work we explore son’s focus of foveal attention [3, 4, 5] and eye movement
Electroencephalography (EEG) eye tracking from less intru- types such as saccades (large and rapid movements), smooth
sive wireless cap scalp-based electrodes. Classification algo- pursuits (slow movements), fixation (little movement) and
rithms to detect eye movement and the focus of foveal at- blinking.
tention are proposed and evaluated on data from a matched Eye tracking techniques can be classed as either Video-
dataset of VOG and 16-channel EEG. The algorithms utilise oculography (VOG) or non-optical. Electrooculography
EOG artefacts and the brain’s steady state visually evoked (EOG) [6] is a non-optical method for eye movement sensing
potential (SSVEP) response while viewing flickering stim- based on changes measured in the steady corneoretinal po-
ulus. We demonstrate improved performance by extracting tential difference. The potential changes are measured with a
features from source signals estimated by Independent Com- set of electrodes mounted around the eyes. If the eye moves
ponent Analysis (ICA) rather than the traditional band-pass from its central position, the positive pole of the eye (cornea)
preprocessed EEG channels. The work envisages eye track- moves to one of the electrodes while the negative pole (retina)
ing technologies that utilise non-facially intrusive EEG brain approaches the opposite electrode. The dipole orientation
sensing via wireless dry contact scalp based electrodes. change effects the EOG signal amplitude. Although EOG
can estimate 2◦ visual angle [5], it can be cumbersome and
Index Terms— ICA, SSVEP, VOG, eye tracking, visual uncomfortable, normally limited to laboratory experiments
attention and they be employed in daily life and mobile circumstances.
VOG records an eye image sequence with high resolu-
tion cameras. They rely on the software processing system
1. INTRODUCTION for obtaining the eye position [5, 7]. Cameras are either head
mounted or remote. VOG is more comfortable and less intru-
The prevalent method for eye tracking is a camera pointed at sive than EOG. However, remote eye trackers limit the visual
the eye - Video-Oculuography (VOG). Eye image capture is field and mobility. Head-mounted trackers require users to
inherently noisy due to the sensor technology and the unstable wear devices which potentially limit physical activity.
physical space between the camera and eye from occlusion,
movement and varying light conditions; state-of-the-art head
mounted cameras have 0.5◦ visual error and remote systems 2.2. Electroencephalography
2◦ visual error [1]. The variation of the surface potential distribution over the
Brain signal processing via Electroencephalography scalp reflects the brain’s functional activities which are
(EEG) provides two additional eye movement information recorded by placing a set of electrodes (sensors) on the scalp
sources- Oculomuscular movements via Electrooculography and measuring the voltage differences between electrode
(EOG), and the brain’s response to flickering stimuli - the pairs. The resulting data is called the Electroencephalogra-
Steady-State Visual Evoked Potential (SSVEP) [2]. In this phy (EEG) [8, 9].
work, these two information sources are combined to develop EEG signals are contaminated by artefacts rendering them
an EEG-based eye tracking solution. less usable. There are two categories of artefact- environ-
Section 2 reviews eye tracking methods and EEG. Sec- mental and biological [10]. Environmental artefacts originate
tion 3 describes the proposed EEG signal processing method. from outside of the body such as 50Hz or 60Hz power line
Section 4 reports the evaluation including details of the noise. Biological artefacts are electrical activities originate
EEG/VOG dataset collected specifically for this study. Sec- from non-cerebral origins such as muscle artefacts (facial and
tion 5 discusses future work. corporeal) and EOG.
4.2. Classifiers
Eight kNN Classifier pairs (labelled C1 − 8 in the result
Table 1) are trained on the EEG session data following the
process outline in section 3. Each pair consists one classifier
trained on EEG pre-processed channels and one classifier
trained on source components estimated by ICA. C1 and C2
classify long and short saccade directions respectively from
Fig. 1. Display to user over time (left to right). Row (a) Pos-
EOG features (8 class problem). C3 and C4 classify FoA
sible stimuli locations; (b) Elicit short saccade; (c) Elicit fix-
fixations and pursuit movement respectively from SSVEP
ation in the display centre; (d) Elicit a long saccade; (e) Elicit
features (2 class problem). C5 and C6 repeat this but for
fixation at the display edge; (f) Possible stimuli locations (g)
SSVEP detection for a specific flickering frequency (4 class
Static SSVEP stimulation for fixations where stimulus is fixed
problem assuming 3 potential FoA flickering at different fre-
at the screen centre; (h) SSVEP stimulation for pursuit where
quencies). C7 and C8 classifiers classify pursuit direction
stimulus moves from the screen centre to the sides.
with and without SSVEP present (2 class problem).
The saccade and pursuit direction detection performance
The 30Hz VOG data is up-sampled to match the 128Hz EEG is assessed by the Root Mean Square Error (RMSE) of the
data by linear interpolation; this approximates the saccades’ difference between the detected angle and the real angle asso-
trajectory in VOG. ciated with the stimuli’s position:
v
u
u1 X N
4.1. Sessions recorded
RM SE = t (θt − θ̂t )2 (1)
N t=1
Fixation, Saccade and Pursuit direction detection from EOG:
Fixations and horizontal, vertical and diagonal saccades are
elicited by visual cue. The EEG collected is used to detect and ,where N is the time length, and θt , θ̂t are the real and esti-
classify pursuit and saccade direction from the EOG features mated angular stimuli position at time t, respectively.
as outlined in section 3. Refer to Fig. 1, users are presented
with a black square display and there are 9 possible stimulus 4.3. Results
locations around the screen edge. Sessions start with a white
fixation cross appearing. To elicit short saccades, the cross Table 1 summarises the classification performance results ob-
appears in the centre. To elicit long saccades, the cross ap- tained through different classifiers (C1-C2). The classifiers’
pears at the edge. After a random time between 500ms and accuracy and RMSE before and after applying ICA are re-
900ms, the stimulus is presented at one of the nine edge lo- ported. As its evident from the results, the all classifiers’
cations. The cross disappears at a random interval 500ms to accuracy is improved and the between-person variation de-
900ms after stimulus presentation which provides the cue for creased significantly if features are extracted from ICA com-
users to make a saccade to the stimulus location. Stimuli are ponents (Fig. 2).
presented for a further 1s after the cue. Each user performs For C1, short saccade direction classification accuracy is
112 cued fixations and 112 cued saccades which lasts approx- almost 74% with a 14% improvement if features are extracted
imately 11 minutes. from ICA components. C2 benefits about 2% more from ap-
Focus of foveal Attention (FoA) during fixations and pur- plying ICA than C1, result in 89% accuracy for long saccade
suit movements: Fixations and smooth pursuits are elicited by direction classification. Considering RMSE, similar trend is
flickering stimuli. The EEG collected is used to detect and obtained when comparing C1 and C2 (Fig. 2).
classify FoA fixations and pursuits direction from features For C3 and C4, FoA estimation (i.e. SSVEP detection
outlined in section 3. Similar to the previous session, refer during fixations and pursuits respectively) accuracy again is
to Fig. 1, the screen is a black square. Again, sessions start shown to be improved (e.g. C3 by 6% to 97.90%) with ICA
with a white fixation cross appearing. After 5s the stimulus - a component features. Similarity between results suggests that
10x10 checkerboard - appears flickering at one of three differ- the SSVEP response during pursuit movements is no harder
Fig. 2. , (a) RMSE for saccadic eye movement direction detection; C1: Short saccade; C2: Long saccades, grey bars represent
RMSE when features are extracted from original EEG channels, black bars represents RMSE when features are extracted from
ICA estimated sources; (b) SSVEP frequency classification accuracy;(c) Pursuit direction detection; C7: without SSVEP, C8:
with SSVEP. The error bars represent standard errors.
to detect than fixations despite the potential for more EEG and evaluated. Feature extraction from EEG source compo-
signal contamination due to EOG. nents estimated by ICA results in better performance com-
For C5 and C6, FoA estimation for SSVEP given 3 po- pared to feature extraction from pre-processed EEG channels;
tential stimuli flickering has 96% accuracy for pursuit move- notably between-person variation reduces. This work is the
ments and fixations with features extracted from ICA compo- first we are aware that uses source components for SSVEP
nents. classification rather than EEG channels. Future work will
Finally, results from C7 and C8 demonstrate that the pres- refine the eye movement type classification in the presence
ence of SSVEP in the visual field does not hinder the classi- of other EEG artefacts such as body movement. The VOG
fication of pursuit direction with both classifiers show similar data captured in this study affords the opportunity for hybrid
results. VOG/EEG multimodal eye tracking. The modulation of real-
Overall, eye movement direction classification accuracy scenes with flickering stimuli via augmented reality technolo-
is best in C8 (i.e. 97.57%) with the least RMSE (i.e. 1.63◦ ) gies to estimate focus of foveal visual attention will poten-
where there is pursuit eye movement rather than saccadic eye tially lead to VOG-free eye tracking, without the social and
movement. physical intrusion of face mounted electrodes.
5. DISCUSSION 6. REFERENCES
A signal processing scheme for extracting eye tracking data [1] Dan Witzner Hansen and Qiang Ji, “In the eye of the be-
from wireless portable scalp-based EEG has been proposed holder: A survey of models for eyes and gaze,” Pattern
[8] Hans Berger, “Über das elektrenkephalogramm des [19] Danhua Zhu, Jordi Bieger, Gary Garcia Molina, and
menschen,” European Archives of Psychiatry and Clin- Ronald M Aarts, “A survey of stimulation methods used
ical Neuroscience, vol. 87, no. 1, pp. 527–570, 1929. in ssvep-based bcis,” Computational intelligence and
neuroscience, vol. 2010, pp. 1, 2010.
[9] Saeid Sanei and Jonathon A Chambers, EEG signal pro-
cessing, Wiley. com, 2008.