Abstract
Introduction
Human–machine interface (HMI) is a field covers multiple technologies including computer science, artificial intelligence, sociology, etc. The definition of HMI is that when person and machine communication by some kind of interface. Person can control the machine and the machine can report status to the user through the interface. In the other hand, the correct communication of messages and control between humans and computers is the main definition of the HMI. For example, Jira et al. 1 developed a gesture-based human–machine interface for vehicle driving using finger pointers that provide user for controlling an in-vehicle system.
As mentioned above, how to make the user control the machine is an important issue. There are many ways to let the user control the machine. The most traditional is to control by using program instructions, but it is also the most difficult way. Therefore, there are many easier ways to implement control the machine. The easiest and least laborious way is to use visual control.
Vision is crucial to human cognitive behavior. In humans, eyes are one of the primary sensory organs and can distinguish between features of light, distance, and color. The central fossa of the eyeball is densely populated with photoreceptor cells, making it the most sensitive part of the visual perception zone. Eye movements enable the gaze to focus on the central fossa and thereby obtain clear visual information. Therefore, several studies raised the issues of using visual information for HMI and interactive machine.
As mentioned, eye movements are closely related to the target under observation. In addition, the trajectory of gaze probably represents cognitive vision and processing history. In daily life, people look at the things that they have a natural desire to interact with. It is even possible to analyze the state of mental activity through the trajectories of eye movements.
An eye tracker, also known as an eye movement tracking instrument, is an instrument that is used to record eye movements and convert them into gaze trajectories. In recent years, practical products with integrated eye trackers have been commercially produced. The results of gaze tracking can be used for related research in neuroscience, psychology, education, marketing, and advertising analysis.
In addition to being used for research, the eye tracker is also a communication aid for physically disabled people that can be used to express self-awareness through behaviors such as eye movements and blinking. Therefore, for patients with problems such as motor neuron disease, muscular dystrophy, cerebral palsy, spinal injury, multiple sclerosis, or stroke or for those with difficulty speaking, the eye tracker can be used as an aid for communication with the outside world. 2
Since 1990, various eye trackers have been developed for specific applications; however, because such products are still expensive, they are usually used for research purposes. Therefore, this study proposed a low-cost wearable device that could be developed into a low-cost eye tracker.
In recent years, wearable devices have gradually emerged. Some head-mounted devices allow users to have immersive experiences in virtual reality. Such devices can enable the user to switch between scenes regardless of time and space and are widely used in games and education. Therefore, the gaze point and its movement trajectory can also provide a quick and intuitive secondary indicator for the human–machine interface and provide optimized information on the drawing operation to improve the user experience. 3
The most detection algorithms for eye trackers have been classified as model-based methods 4 and feature-based methods. 5 Model-based methods usually match the vision with preset shapes and model formulas and then determine the optimal solution among the limited candidates through voting or classification. With the correct model, the model-based method can consistently locate the iris. For a typical model-based method, the difficult part is the stable and accurate determination of the iris position on the basis of the reflection points created by the ambient light sources shining from different directions. Feature-based methods usually use the features of the eyes. A typical feature-based method fits the shape of the eye in a circular or elliptical shape to locate the center of the iris. Compared with model-based methods, feature-based methods may require a much smaller code that can perform in real time; however, a typical feature-based method is unfavorable compared with a typical model-based method in terms of stability and accuracy.
This study used a highly stable and accurate feature-based method to track gaze and adjusted a pupil elliptical extraction method to make it suitable for head-mounted applications to achieve the purpose of the HMI and develop an interactive machine. For the gaze-point estimation, the polynomial equations of the mapping model are extended to enable the three-dimensional fixation vector to correspond to the two-dimensional visual plane. The correction procedure of the eye model, which simplifies the calibration process of the overall system, minimizes the correction time, and improves ease of use, was integrated into the sampling period established by the mapping model.
Therefore, this study proposes a set of eyesight devices that can capture eyeball images through the Logitech Quick Cam Pro 5000 webcam to achieve the low-cost wearable devices.
Related literature
In the 20th century, Huey 6 created the first invasive eye tracker, as displayed in Figure 1; the device resembled a contact lens with a hole positioned near the middle of the pupil. The remainder of the contact lens was connected to an aluminum pointer; eyeball and pointer movements were synchronized to capture the movement trajectory of the eyeball. However, this was uncomfortable for the subject. In addition, the subject was required to describe the movement of the eyeball and compare it with the captured data. Thus, the results were easily influenced by subjective feelings.

Invasive eye tracker.
As depicted in Figure 2, Babcock et al. 7 created the first noninvasive eye tracker in 1935. This eye tracker sent a beam of light onto the subject's eye and then used photographic film to record the reflection from the eye. The continuous light spot recorded on the film was used as the moving trajectory of the eyeball, and the position of the gaze point was analyzed accordingly. The results of the study were more objective than were those of the invasive eye tracker. However, this noninvasive eye tracker was not popular because of its high cost.

Noninvasive eye tracker.
The detection algorithms for eye trackers have been classified into model-based methods and feature-based methods. Model-based methods usually consider entire images in the context of preset shapes and model formulas. For example, Kao et al. 4 proposed a model-based eye tracker. Kao’s model-based method first defined an eye model and then calculated the angle of rotation of the model. The angle of rotation can match the position of the iris on the image with the position of the iris defined in the model. It can be seen that the center position of the iris on the image is the position of the iris in the model after being rotated through a certain angle. Świrski and Dodgson 5 proposed that it is also a shape-based eye tracker, which uses a pupil image to extract an ellipse and back-projection ellipse into a stereo space. After multiple sampling operations, an eye model can be generated. The pupil ellipse that is subsequently captured can fit the most likely fixation angle from the model. It is computationally intensive and does not rely on reflected spots for tracking.
The feature-based approach uses the common characteristics of the human eye to identify the eye. These include the corneal edge, pupil, or the reflection of light on the cornea. The main purpose of each feature is to obtain information concerning the eyes and face. Therefore, it is not highly sensitive to changes in illumination. For example, Yang used a nonlinear filter to identify suitable candidate points for formalizing information of the iris. However, Yang et al. used method that required high-quality images for accurate detection. 8 Yang et al. 9 used the edge color information of the pupil and iris to obtain the threshold value from the color information around the eye. Sirohey et al. used iterative methods to determine the most suitable threshold value for distinguishing skin color using the model that establishes the eye frame. However, in environments with different brightness levels, the detection tended to fail due to the similarity of eyebrows to shadows. 9 Sigut and Sidha proposed a method for supporting head movement to combine the bright spots reflected on the iris based on shapes and additional light sources. The distance from the center point of the iris to the point of reflection can be used to determine the distance and direction of head movement to correct the effect of head movement. 10 Lee et al. proposed a method for a portable device that involved installing multiple sets of light-emitting diode lights on the device. These light sources reflected the spots on the iris; Lee et al. calculated the dependent positions of each of these points. The vector correction of head rotation was calculated to correct the fixation deviation caused by head movement. 11 Jen et al. 12 proposed a new wearable eye-gaze tracking system with a single webcam mounted on glasses; this system involves skin detection and eyelid removal to extract the region of interest. Dobeš et al. 13 developed a successful method of eye and eyelid localization on a computer using a modified Hough transform. The efficiency of this method was tested on two publicly available face image databases and one private face image database.
The study in Kao et al. 4 proposed a model-based detection method. This paper is a feature-based detection method. The comparative literature 4 is also a feature-based detection method. This paper uses the Hough circle, the least-squares method, and the particle-filtering method to obtain the eye-tracking results. 4 However, in the preprocessing part, the gray-scale morphology and Sigmoid function are used to increase the light and shadow. Changed resistance, and does not use the region of interest to limit the positions of iris detection. Under low lighting and strong lighting conditions, the range of the regions of interest will be cut into the eye, and the iris outside the region will be detected.
In Kao et al.,
4
a high-end and high-speed camera is adopted for eye detection and tracking. This study can accurately be tracked using a low-cost consumer webcam. Comparably, the study in Kao et al.
4
needed a chin to fix the human head position. The input images can be taken by a wearable device in a mobile environment. The advantages of this study are as follows:
The average error of the detection rate is stably within 9 pixels. Under the 50-cm distance from the screen, the error of the fixation mapping is within 3°. Feasibility under different light sources. Accurate detection and tracking of iris center using a low-cost and low-speed camera.
This study combined a low-cost head-mounted device with a feature-based method. An image preprocessing method helped to enhance the characteristics of the iris edge and fit a circle with the characteristics of the iris edge when the system had located the center of the iris. The center of the circle represented the center of the iris and was tracked using a filter. This study developed a low-cost head-mounted eye tracker to provide a communication mechanism for people with disabilities.
Device
The device flowchart is depicted in Figure 3. When the system is operating, a Logitech Quick Cam Pro 5000 captures eyeball images at 30 fps with a resolution of 320 pixels × 240 pixels. The outer casing is removed, and part of a photographic lens is fixed on a pair of glasses. Our built-in visible eye tracker processes gaze points, the coordinates of which are subsequently output and mapped onto the screen to present the eye movement trajectory.

The proposed devices.
When performing fixation and screen gaze-point mapping corrections, the user must remain at a fixed distance from the screen. As shown in Figure 4, a chin support is adopted to assist users in maintaining a fixed distance between the user’s head and the screen.

Chin support.
The photograph presented in Figure 5 indicates that errors are caused in the test environment as a result of distance and screen size; the optimal distance for testing is approximately 50 cm.

Testing process.
Method
The system architecture diagram is presented in Figure 6. This study had three steps. First, image preprocessing was used when the camera captured the image. After thorough image preprocessing, eye detection was conducted. Third, eye tracking and mapping were conducted.

System architecture diagram.
Preprocessing
This study used gray morphology as the preprocessing to increase the amount of iris information obtained. After obtaining the iris information. This process first extracted the part of the iris in the image, then this study uses automatic multilevel thresholding technique to achieve contrast enhancement. As a result, this study will observe the first 10 images. If the iris features are too rare, they will use to create the iris mask to achieve feature enhancement. Its details are as follows:
This study used grayscale images for preprocessing to increase the amount of iris information obtained.
After iris information had been added, an automatic multilevel thresholding technique14–16 obtained the regions of interest for the iris in the image. Mathematical morphology operations of dilation and erosion were then conducted on the thresholded regions. This study used circular structural elements to conduct morphological operations with circular structural elements.
Formulas (1) and (2) describe morphological erosion and dilation, respectively. Formula (1) describes the erosion operation, where
As illustrated in Figure 7, the morphological erosion operation magnifies the darker parts of the grayscale image. The dilation process enlarges the brighter part of the grayscale image. Grayscale images must be treated first, but after erosion, the iris is deformed in the image. When the system executes the dilation process, it attempts to return the iris to as close to its normal size as possible.

Grayscale images with morphological operations. (a) Grayscale Images. (b) Erosion results. (c) Dilation results.
After the system has performed grayscale morphology, it subjects the images to automatic multilevel thresholding.14–16 However, the general automatic multilevel thresholding method has considerable difficulty with thresholds. Therefore, this study used an automatic multilevel thresholding algorithm to automatically find the threshold value.14–16 An automatic multilevel image thresholding system14–16 can homogenize a grayscale image as a histogram. Automatic multilevel thresholding14–16 can then be conducted to identify the most suitable thresholds for different images. To achieve optimal image characteristics, this study performed two runs of automatic multilevel thresholding on images14–16 to automatically determine the threshold. In this study, the first run of automatic multilevel thresholding with images14–16 tended to over amplify eyelid features; this obscured the characteristics of the iris. The second run of automatic multilevel thresholding with images14–16 dramatically increased the resolution of iris characteristics and reduced other noise to increase the iris detection rate. Histogram equalization 17 served to redistribute and accumulate pixel values of grayscale images. As depicted in Figure 8, the quantity corresponding to the pixel value corresponded with the cumulative amount. Histogram homogenization was ineffective for redistribution. Therefore, this study used the sigmoid function to perform the redistribution.

Locating the iris center
After image preprocessing, the iris center must be located. Locating the position of the center of the iris requires two steps: the first uses the Hough circle, 18 and the second uses the least-square circle-based method. 19 Finally, the detected iris center must be tracked through the particle filter.
Figure 9 presents a diagram of the Hough circle detection process. The image subjected to the image feature enhancement process is used as the input for Hough circle detection. The system can obtain a circle after performing the Hough circle conversion. However, the reliability of detection using the Hough circle is insufficient; therefore, the least-square circle-based method must be used for correction.

Hough circle detection result.
The system cannot clearly define the pupil position using Hough circle detection in the image. Therefore, this study used the difference between the iris and the sclera, which is called the limbus, as the basis for judgment. If the Hough circle detection returns the wrong position for the center of the iris, it must be corrected using the least-square circle-based method. The least-square circle-based method
19
scatters the center of the correct result to obtain the feature points of the heterochromatic edge. As shown in Figure 10, two steps must be executed to identify the limbus features: (a) The radius of the largest circle detected by the Hough circle is used as the scattering radius. The system extends

Feature of limbus.
Formula (3) is used for the least-square circle-based method;
When the parameters (
As indicated in formulas (5), (6), and (7), the three unknowns
As demonstrated in equations (8) and (9), after obtaining values for
Figure 11 depicts the result of least-square circle-based detection, which provides the center of the circle, the radius of the center, the standard deviation of the radius of the feature points of the different color edges, and the ratio of the black and white pixels in the detection result circle. These items can be used for follow-up tracking in target feature alteration.

Least-square circle-based detection result.
Tracking the iris center
The method of locating the iris center in motion pictures is difficult to use for detection. Therefore, this study added tracking methods to compensate for the lack of detection. For example, the most common tracking method is the use of the Kalman filter.20–22 The Kalman filter is used to make predictions, and then the results are corrected using the Gaussian distribution and a linear model. However, the Kalman filter employs linear tracking, whereas the movement of the eyeball is nonlinear. Therefore, this study used particle filters for tracking rather than the Kalman filter. Particle filters utilize nonlinear tracking and can design feature and weight values.
As illustrated in Figure 12, the original algorithm for the particle filter is divided into three stages: (1) random sampling; (2) importance sampling; and (3) resampling. First, the particle filter must determine whether tracking is being performed for the first time. If so, then the system must distribute the particles according to the results for the center of the iris. If tracking is not being performed for the first time, then the system must distribute the particles according to the results of the previous trace. Importance sampling constitutes the second stage and has enabled the accurate tracking of particle filters in many studies. Importance sampling serves to determine whether each particle is the same as the target being tracked and assign each particle a weight value. Calculations based on this feature produce the weight value. Finally, the particle with the highest weight value is used as the tracking result. The third phase, resampling, mainly serves to distribute the position of each particle in advance. The distribution is based on the weight value of each particle. Particles with high weight values can generate numerous new particles around them. Particles with low weight values can only exist and cannot produce new particles around them. This stage does not produce an image of the tracking results. Resampling redistributes the position of the new particle based on the weight. After the particles with high weight values have been redistributed, new particles can be generated around them. Only particles with low weight values produce few or no new particles. The object is tracked back and forth until it has disappeared. Figure 13 displays the flow chart of tracking in this study.


Particle filter flow chart of this study.
Formula (10) describes random sampling. The system first confirms whether tracking is being performed for the first time when particles begin entering the particle filter. The first tracking randomly distributes the particles to locate the center of the iris, and subsequent tracking is employed to randomly distribute the particles on the basis of the previous tracking result.
Formula (11) indicates that the input of the research system is a flat image. Therefore, the target position can be divided into a horizontal X-direction and vertical Y-direction. When using noise, the position of each moment should differ slightly.
Formula (12) presents the Gaussian distribution.
This step replaces the resampled portion. Therefore, this study uses the tracking result coordinates at the previous moment and the result coordinates for the center of the iris at this moment as a reference to adjust the range in which the particle group can be dispersed.
Importance sampling must identify the features of each particle, compare the features of the tracking target, and then identify the weight value of each particle. A high weight value indicates that the target is large. A lower weight value indicates that the particle is less similar to the target. As indicated in formula (14),
In formula (18),
Gaze-point calculation and correction
After calculating the iris center point coordinates in the image, it is possible to evaluate the correlation between the center point of the iris and the coordinates for the gaze point on the screen. Figure 14 illustrates the result of the subject gazing at the nine points on the screen for gaze-point correction.

Correction point diagram.
The location of the green dot on the screen varies depending on the subject’s gaze direction. The center coordinates of the iris in the image also differ. Numerous methods exist for correction mapping. For example, the method of calculating the gaze-point map by using the first-order polynomial equation is simple, and few resources are required for calculation, but the accuracy is low.
25
A fifth-order polynomial equation is used for calculation when using the gaze-point mapping method, thus improving accuracy and computational complexity;
26
the most commonly used method is gaze-point mapping through second-order polynomial equations.
27
Because eye movements are nonlinear, this study used the second-order polynomial method.
In formula (21),
Experiment results
Experiment device and environment
Figure 15 presents an iris center detection result diagram. This study conducted three experiments, namely for evaluating the performance of eye detection, iris center locating, eye tracking, and fixation error calculation. In these experiments, seven subjects were tested with normal, weak, and strong light. Users looked at nine points on the screen and moved their heads to make the light change. This study collected data on precision, recall, and accuracy. In addition, this study conducted experiments on iris center gap calculation and fixation error calculation. Please refer to Table 1 for detailed specifications and a description of relevant computers.

Diagram of iris center detection result.
Computer details.
Please refer to Table 2 for detailed specifications of the Acer V276HL screen, which had a resolution of 1920 × 1080 pixels.
Screen details.
This study was conducted in one environment for all seven subjects, who were presented with the same scene. The test items were open eye detection rate and eye-tracking accuracy. The detection rate had four calculation parameters: true positive (TP), which indicates that both the actual situation and detection are positive samples; true negative (TN), which indicates that both the actual situation and detection are negative samples; false positive (FP), which indicates that the actual situation is a negative sample but the detection is a positive sample; false negative (FN), which indicates that the actual situation is a positive sample but is detected as a negative sample. If a positive sample actually occurs, then the event has a value of 1 and the negative sample is zero. If a positive sample is detected, then the detection is 1 and the negative sample is zero.
In formula (22), the accuracy is the ratio of the correct detection of all samples. In formula (23), the precision is the ratio of positive samples of events in all positive samples detected. In formula (24), the recall is the ratio of the positive samples detected in the positive samples of all events.
In the experiment, this study uses the webcam at 30 fps with a resolution of 320 pixels × 240 pixels. In the course of the experiment, the detection speed has reached 0.015 s and it is real time. If that can use a better camera, you can get better images.
Kao et al., 4 Jen et al., 12 and Dobeš et al. 13 are evaluated under similar scenarios with this study. So this study is compared with the Kao et al., 4 Jen et al., 12 and Dobeš et al. 13 Dobeš et al. 13 only provided the performance evaluation results on eye detection, Kao et al. 4 only provided performance evaluation results on locating the iris centers; while Jen et al. 12 provided the performance evaluation results on both eye detection and tracking. Therefore, the method of this study was compared with the methods of Jen et al. 12 and Dobeš et al. 13 for performance evaluation on eye detection results; and compared with the methods of Jen et al. 12 and Kao et al. 4 for tracking results.
As compared with the above-mentioned literatures, the average errors of the eye detection of this study are stable within 9 pixels. The average error of the fixation mapping is within 3° within 50 cm of the screen, and a low-cost and low-speed camera can also be adoptable for eye detection under different lighting sources.
Performance evaluation on eye detection
This experiment is comparable to those conducted by Jen et al. 12 and Dobeš et al. 13 Figure 16 presents an experiment conducted in a room with normal lighting. Users looked at nine points on the screen and were able to move their heads suitably.

Test under normal lighting conditions.
Table 3 lists the results of open eye detection rate under normal lighting conditions for seven users, which reveal that the average accuracy, precision, and recall rates can exceed 97% under normal lighting conditions. In addition, other rates were higher than those reported by Jen et al. 12 and Dobeš et al., 13 except for precision, which was slightly lower than that reported by Dobeš et al. 13
Eye detection rates under normal lighting conditions.
TP: true positive; TN: true negative; FP: false positive; FN: false negative.
Figure 17 demonstrates the setting of experiments performed under weak lighting conditions. The environment and subjects were the same in this experiment, but all indoor lights were turned off and the only light source was a window. Users looked at nine points on the screen and moved their heads to make the light change.

Test under weak lighting conditions.
Table 4 lists the open eye detection rates under weak lighting conditions. Although the average result was unfavorable compared with that reported by Jen et al. 12 and precision was lower than that reported by Dobeš et al., 13 the average accuracy, precision, and recall rates of eye detection results under weak lighting conditions reached 98.55%, 99.17%, and 97.75%, respectively; thus, the experiment demonstrated satisfactory open eye detection rates.
Open eye detection rate under weak lighting conditions.
TP: true positive; TN: true negative; FP: false positive; FN: false negative.
Figure 18 demonstrates the setting of experiments performed under strong lighting conditions. Users looked at nine points on the screen and moved their heads to make the light change.

Test under strong lighting conditions.
As listed in Table 5, the open eye detection rate with strong light demonstrated average results superior to those reported by Jen et al. 12 and Dobeš et al. 13 The average accuracy, precision, and recall rates exceeded 98%.
Open eye detection rates under strong lighting conditions.
TP: true positive; TN: true negative; FP: false positive; FN: false negative.
Performance evaluation of locating the iris center
Figure 19 presents the diagram of the experiment performed for locating the iris center, which is comparable to those performed by Kao et al. 4 and Jen et al. 12 The figure has three points representing different studies. The green circle represents the true center of the iris, the yellow rectangle represents the iris center detected by Kao et al., 4 the pink cross represents the iris center detected by Jen et al., 12 and the blue triangle represents the iris center detected in the present study.

Diagram of locating the iris center.
As shown in Table 6, normal, weak, and strong lighting conditions produced different iris detection results. The experimental results revealed that the performance of iris center locating was worse under normal and strong lighting conditions than under weak lighting conditions. Normal lighting demonstrated an average distance of 6.64 pixels, and strong lighting demonstrated an average distance of 6.95 pixels. Weak light produces less reflective light on the iris and more accurate detection of the iris center. In addition, the experimental results indicate that the iris center was more accurately located in the present study than in those conducted by Kao et al. 4 and Jen et al. 12
Differences in iris center locating performance.
Performance evaluation on eye tracking
Figure 20 presents the comparative eye-tracking results. This experiment compares to those performed by Kao et al. 4 and Jen et al. 12 The experimental also conducted under normal, weak, and strong lighting conditions.

Diagram of eye tracking.
As shown in Table 7, normal, weak, and strong lighting conditions produced different eye tracking error results. The experimental results revealed that the eye-tracking errors were the best under weak lighting conditions among all other lighting conditions. Normal lighting demonstrated an average eye-tracking error distance of 7.15 pixels, while the strong lighting demonstrated an average eye-tracking error distance of 5.16 pixels. Weak light produces the best performance on the eye-tracking results. In addition, the experimental results also indicate that the proposed eye-tracking method can achieve better performance than those conducted by Kao et al. 4 and Jen et al. 12
Differences in eye-tracking error.
Performance evaluation of fixation error
Figure 21 presents the fixation mapping diagram. The subject is given 6 s to stare at each point; the data were divided into three sets. Each data set has central coordinates of the iris for a subject that gazed at nine points. The system must randomly select a datum to calculate the gaze-point coordinate conversion parameters. The remaining two data sets are used with calculated parameters to convert the iris center coordinates into gaze-point coordinates. The system must then select one of the remaining two materials to calculate the conversion parameters after the statistics have been obtained before using this calculated parameter to convert the iris center coordinates of the remaining two data sets. This is the loop that must be iterated until an answer is obtained for each data set. As listed in Table 8, each subject produced an average gaze-point mapping error. The error angles of the sixth, seventh, and eighth points were greater than 3°, but the angles of the remaining points were within 3°. Therefore, the overall accuracy of the mapping was higher than that of mapping performed in other studies.

Fixation mapping diagram.
Average fixation mapping error.
Summary of experiment results
In eye detection rates under normal lighting conditions. The average accuracy, precision, and recall rates can exceed 97% under normal lighting conditions. In addition, other rates were higher than those reported by Jen et al. 12 and Dobeš et al., 13 except for precision, which was slightly lower than that reported by Dobeš et al. 13 In the open eye detection rate under weak lighting conditions. Although the average result was unfavorable compared with that reported by Jen et al. 12 and precision was lower than that those reported by Dobeš et al. 13 In the open eye detection rate under strong lighting conditions. The open eye detection rate with strong lights demonstrated the average results of this study are superior to those reported by Jen et al. 12 and Dobeš et al. 13 As for the differences in iris center locating performance, the performance of iris center locating was worse under strong lighting conditions than under weak lighting conditions. The iris centers were more accurately located in the present study than in those conducted by Kao et al. 4 and Jen et al. 12 In part of differences in eye-tracking error, the experimental results revealed that the eye-tracking errors were the best under weak lighting conditions among all other lighting conditions that the proposed eye-tracking method can achieve better performance than those conducted by Kao et al. 4 and Jen et al. 12 In part of the average fixation mapping error, the overall accuracy of the mapping results of this study was higher than that of mapping performed in other studies.
As a result, the proposed method can achieve an eye tracker in a low cost and simple way. So that the user can conduct interactive control signals by moving the eye movements. Simple instructions are given to achieve HMI results.
Conclusion
This study develops an eye-tracker system for HMI and interactive machine. The eye-tracker system proposed by the research institute is used with natural light. The system can be safely used for a long period of time without placing extra burden on the eyes. It is cheaper than most eye trackers on the market because it uses a general webcam. The experimental results prove that the overall open eye detection precision and accuracy rate exceeded 95% and that the error in detecting the iris center error is also small. The fixation error calculation revealed that the angles of the remaining points were within 3° of each other. This study compared with the methods of Jen et al. 12 and Dobeš et al., 13 which are performance evaluation on eye detection. The method of this study compared with the methods of Jen et al. 12 and Kao et al., 4 which are performance evaluation of locating the iris center. Therefore, the overall mapping accuracy is high. This study designed an eye tracker that only requires a webcam and computer to track the line of fixation and map the gaze point of the user. In addition, all of the hardware components used were inexpensive and easy to obtain. The overall system cost is lower than that of other products on the market.
