Keywords

1 Introduction

Eyes are among the important sensory organs that we receive outside information. The eye is not only an important channel to obtain information, but also send information through the human behavior. Through the location of the pupil we can get the human eye line of sight. As early as the beginning of the 20th century, scholars tried to analyze people’s psychological activities through the record of the eye movement trajectory, pupil size, gaze time and other information [1], but the observation method is relatively simple, mainly relying on manual recording. So its accuracy and reliability are difficult to be guaranteed. It is so difficult to accurately reflect the true trajectory of eye movement. The appearance of eye tracker provides an effective means for the psychologist to record the information of eye movements. It can be used to explore the visual signals in a variety of situations, and to observe the relationship between eye movements and human psychological activities [2]. Eye tracker is an instrument based on eye tracking technology to collect eye movement trajectory. With the development of computer, electronic information and infrared technology, eye tracker has been further developed in recent years [3].

With the development of artificial intelligence technology, the research of eye movements can explore the visual information processing and control system, because of its wide application value and business prospects in psychology [4], such as medical diagnosis [5], military research [6], commercial advertising evaluation [7] and driving the vehicle [8] and other fields. Eye tracking technology has made rapid development in recent years. The researchers extract and record the user’s eye movement data, and then extract the important information by analyzing these data. Eye tracking technology can also be used in education field. We can use the learner’s visual psychology to simulate a realistic learning environment, and interact with learners. Eye tracking technology can also be used in an e-learning system by taking appropriate interventions when the user does not focus on the active content on the computer screen for a long time or is concerned with the irrelevant area of the learning task.

2 Eye Tracking Technology

Nowadays, the eye tracking technology is based on the interactive electronic products, which is divided into wearable and non-contact eye movements. The basic principle of eye tracking technology is the use of image processing technology to extract the eye image directly or indirectly using the camera, by recording the position of the infrared light spot reflected by the human cornea and the pupil, so as to achieve the purpose of recording the change of the line of sight. The eye movement system identifies and determines the vector change between the pupil center and the cornea reflex point by analyzing the eye video information collected from the camera. Thereby determining the change of gaze point (Fig. 1).

Fig. 1.
figure 1

Non-contact and wearable eye tracker.

2.1 Face Detection and Eye Positioning

At present, face detection classifiers are mostly based on haar features by using Adaboost learning algorithm training. In this paper, the face detection algorithm is also based on Haar feature and AdaBoost classifier, but combined with prior knowledge of facial feature distribution, to locate the eye area, and the human face region and human eye region are positioned in the image (Fig. 2).

Fig. 2.
figure 2

Face recognition and human eye positioning.

2.2 Pupil Localization and Infrared Spot Detection

Due to eye movement or uneven illumination and other external factors, the pupil area will appear scattered noise information, and these scattered noise on the pupil and cornea reflection high light points will cause errors. Median filtering has some effect on filtering out such noises. So filtering and denoising methods are used as pre-processing of the image.

  1. (1)

    Infrared spot center

    In this paper, the gaze tracking technology is mainly through the near infrared light source in the human cornea to produce the corresponding corneal reflex spot as a reference, combined with the human eye pupil center point, and then through the geometric model for spatial mapping, we get the human eye gaze point.

Step1: Calculate the image gradient.

Due to the complexity of the eye image, we estimate the region of interest (ROI) of the eye, since the gray value of the spot in the region of interest is higher than the gray value of the surrounding pixel, which gives us an idea of locating the infrared spot coordinates. In order to obtain the central position of the spot, it is necessary to divide the spot from the eye images. Before the traditional image segmentation, it is usually necessary to perform edge detection by using the feature image. The traditional edge detection operator has the Sobel operator, Prewitt operator and Roberts operator [9]. The following is a description of the Prewitt operator.

$$ \left[ {\begin{array}{*{20}c} { - 1} & { - 1} & { - 1} \\ 0 & 0 & 0 \\ 1 & 1 & 1 \\ \end{array} } \right],\left[ {\begin{array}{*{20}c} { - 1} & 0 & 1 \\ { - 1} & 0 & 1 \\ { - 1} & 0 & 1 \\ \end{array} } \right] $$
(2)

In the determination of the center of the spot, more stable edge-based fitting method are based on geometric features, through the circle and other geometric features to determine the center of the spot. There are usually grayscale centroid methods:

$$ x_{c} = \frac{{\sum {x_{n} I_{n} } }}{{\sum {I_{n} } }},y_{c} = \frac{{\sum {y_{n} I_{n} } }}{{\sum {I_{n} } }} $$
(3)

Where \( \left( {x_{c} ,y_{c} } \right) \) is the center of the detected spot and \( I_{n} \) is the pixel gray value of the ROI region. So, if our ROI area is small enough, we can use the centroid method to quickly determine the center of the spot. In order to minimize the ROI, we improved the basis of the Prewitt operator, through the gradient of the image, to determine the approximate edge position and to determine the maximum gradient value of the pixel.

$$ g_{x} = \frac{I(x + 1,y) - I(x - 1,y)}{2},g_{x} = \frac{I(x,y + 1) - I(x,y - 1)}{2} $$
(4)

By using the Eq. (5) to determine the gradient matrix of the ROI region, we can ignore the gradient values at the edges of the image based on our prior knowledge of the pupil position. Finally, we determine the maximum point of the gradient point coordinates p, according to the image pixel size, the rectangular area where the spot position is determined with the p point as the center.

Step2: Threshold processing.

After we get the interested area, in order to eliminate the impact of other image areas, we are not interested in for the next image operation, We will get the whole human eye image through the threshold for image segmentation.

$$ I_{n} = \left\{ \begin{aligned} I_{n} ,I_{n} \in ROI \hfill \\ \,\,0,I_{n} \notin ROI \hfill \\ \end{aligned} \right. $$
(5)

Then, we focus on the segmentation of ROI images. In this paper, we study and improve the traditional maximum interclass variance (OTSU) threshold method.

  1. (2)

    Pupil center positioning

    It is easier to locate the pupil center than the spot. Since the gray value of the pupil and the gray value of the surrounding position are quite different in the human eye region, and the image has been preprocessed before the pupil is positioned, so as to select the appropriate threshold for image segmentation.

2.3 Gaze Tracking

The coordinate extraction of the pupil and the center of the spot are the basis for realizing the gaze tracking. The two-dimensional migration vector can be obtained by calculation, that is, the P-CR vector.

$$ x_{e} = x_{p} - x_{c} ,y_{e} = y_{p} - y_{c} $$
(6)

Where \( (x_{p} ,y_{p} ) \) and \( (x_{c} ,y_{c} ) \) is the pupil and spot center coordinates respectively, \( (x_{e} ,y_{e} ) \) is P-CR vector.

In this paper, we use of 6 parameters of the fitting to ensure that the system has good accuracy and real-time.

$$ \begin{aligned} & X_{gaze} = a_{0} + a_{1} x_{e} + a_{2} y_{e} + a_{3} x_{e}^{2} + a_{4} x_{e} y_{e} + a_{5} y_{e}^{2} \\ & Y_{gaze} = b_{0} + b_{1} x_{e} + b_{2} y_{e} + b_{3} x_{e}^{2} + b_{4} x_{e} y_{e} + b_{5} y_{e}^{2} \\ \end{aligned} $$
(7)

During the calibration process, the user needs to follow the calibration point in order to obtain the corresponding P-CR vector. Using the calibration point and P-CR vector to solve the 12 position parameters, this paper uses the least squares method to solve the model parameters.

$$ e^{2} = \sum\limits_{i = 1}^{9} {\left[ {X^{i}_{gaze} - f(x_{e}^{i} ,y_{e}^{i} )} \right]^{2} } $$
(8)

After the parameters are calculated, we calculate the P-CR, the P-CR into the Eq. (7), and to calculate the real scene in the human eye point of view, to achieve the purpose of gaze tracking.

3 The Application of Eye Movement in Mathematics Education

Human eye movement can reflect the human thinking process [9]. The eye movement researcher records the eye movement trajectory based on the eye tracking technique. In the process of mathematics learning, students need to think about the visual mathematical language. In the process of understanding the problem, each different symbol in mathematics represents the different information, the brain for different information symbols, watching, jumping and other different eye movement mode, the response of these eye movement mode of attention, time and pupil diameter and other indicators. In practice, the eyes will be moved to another location once they are gazed at a place after a certain period of time. So if there is a cognitive difficulty in a given area, the fixation time will increase. And, after the students understand the previous content, they can predict the following content. In some areas it will appear the tendency of eye jump. If the students think that part of the content is difficult to understand, there will be short jump distance, so their watching time is longer. At present, the experimenter is the most commonly used eye tracking method with the video image of eye position, that is, to record and analyze the direction of pupil movement by recording video. In this paper, through the real-time eye tracking system designed by ourselves, we can feedback the eye movements in real time. In the aspect of teaching, the type of eye tracker which is often used by teachers is non-contact eye movement. In the experiment, it is necessary to have a host and a camera device. The instrument is used to analyze the subject’s cognitive ability by recording the moving characteristics of the eye when reading the equation and the mathematical symbol of the host. The teachers can improve the teaching quality by improving the teaching design through the student’s eye movement. In terms of student learning mathematics, we judge the students’ acceptance of the teaching content through the eye movement of the students, who can understand the difficult part and the interest part of their own learning. We mainly learn the psychological response from the eye movement indicators of the students (Table 1).

Table 1. Simulated and measured results of the proposed antenna.

Through the two aspects of teaching and learning in mathematics education. Researchers record and analysis the eye movement data of the students, Through the analysis of eye movement data, teachers understand the content of the difficulties of teaching, so as to modify the entire teaching design. On the other hand, students can understand their own learning interests and optimize their own learning strategies. We use their own in the windows platform to develop real-time eye tracking system, the students of mathematics learning eye tracking experiment, because the system is a beta version, only the number of fixations is provided in the interface, and other indicators will be used in the next version (Fig. 3).

Fig. 3.
figure 3

Eye tracking software interface.

In this software, we developed a multidisciplinary education framework, but in the beta version, we only tested the mathematics disciplines. So this paper take mathematics as an example for application of eye tracking (Fig. 4).

Fig. 4.
figure 4

Mathematics test

By students’ watching four different video or photos, we can analyze the students’ eye tracking data. In order to better analyze the fixation point, we add the expression recognition system in the software, while observing the change of the human eye while observing the change of the expression of the student. We can better analyze the learning characteristics of the students. We can make a simple analysis of the students’ learning process from the perspective of the students’ attention to different videos and the changes of the expression in the whole process.

4 Conclusions

In this paper, we apply the eye tracking technology to the teaching of mathematics. Teachers can improve the teaching program according to the eye movement data of the students in the course of learning. On the other hand, students can adjust themselves according to the eye movement situation by the human-computer interaction experiences. Due to the abstraction and complexity of mathematics, this paper provides a feasible solution for mathematics teaching. We will expand the software functions and improve eye tracking accuracy in the future study.