Abstract
1. Introduction
Wireless sensor networks (WSN) [1], which have the essential capabilities of sensing, computing, and communicating, have attracted a wide range of attention in the past decade. WSNs are well suited to many applications such as surveillance and monitoring applications [2, 3]. In such applications sensing coverage is one of the fundamental measurement indexes of the QoS (Quality of Service) [4]. A scalar sensor usually has an omnidirectional sensing range (a circular sensing coverage), while a directional sensor has a limited sensing direction and noncircular sensing range [5]. The advances in the technologies of image sensor and embedded system have promoted rapid development of camera/visual sensor networks (VSN) [6, 7]. VSNs belong to a directional sensor network and can provide visual image or video data for the applications of surveillance and monitoring. Numerous related works of object detection, localization, and tracking for WSNs have been proposed [8]. However, most of these approaches are not applicable to the VSNs due to the different characteristics of a visual sensor [9]. In other words, a visual sensor has a limited effective sensing range characterized by its directionality and size-specific sensing angle; moreover, the type of sensed data by a visual sensor is image-based, which are different from the wireless scalar sensors. Different approaches are necessary for the object detection, localization, and tracking in VSN applications. In this paper, the geometric structure of Voronoi diagram [10] is utilized in the proposed design of a visual sensor handover protocol for target tracking in VSNs where a large number of visual sensors are considered to be randomly deployed in a wild field to perform the sensing and tracking tasks. Moreover, the target localization is designed to be completed by single visual sensor without any prior knowledge of target positions, while other related works did the localization by multiple visual sensors. The use of the concept of Voronoi diagram also facilitates the determination of a nearest visual sensor to a detected target object. To the best of our knowledge, this is the first paper that utilizes Voronoi diagram for both target localization and target tracking in a distributed VSN. The major contribution of this paper includes
The remainder of the paper is organized as follows. Section 2 briefs the related works of target localization and tracking for WSNs and VSNs. Section 3 describes the assumptions and preconditions for this study. In Section 4, the proposed scheme of Voronoi-based target tracking and sensor handover are described. Section 5 evaluates the performance by simulations. Finally, concluding remarks are made in Section 6.
2. Related Works
For many applications of WSNs, utilization of a localization technology is a requirement to perform certain functions. Localization techniques utilized in WSNs provide useful location information for subjects such as deployment, coverage, routing, tracking, and rescue. It can be categorized into sensor localization and target localization [11]. Both the categories of localization in WSNs have drawn considerable attention of researchers and the solution schemes can be classified as coarse grained or fine grained, GPS free or GPS based, anchor free or anchor based, centralized or distributed, range free or range based, and stationary or mobile sensors [12, 13]. Target localization is specifically for the applications of surveillance and monitoring because this kind of applications is usually interested in not only the existence but also the location of a target object. Most of the target localization schemes used in scalar WSNs are not applicable to VSNs due to the characteristic of directional FoV of a visual/camera sensor. A sensor selection-based target localization algorithm was proposed in [14]. A set of camera sensors that can detect the target with a high probability is selected as candidates, and then certain of the candidates will be selected for the estimation of target position. The algorithm needs numerous camera sensors deployed with higher density in the surveillance region and the target needs to be observed/detected synchronously by a number of sensors. In [15], the authors also proposed an algorithm of sensor selection to select a set of camera sensors for the improvement of target localization accuracy. The algorithm also needs the target to occur in the overlapped area of FoVs of several camera sensors. In [16], coarse-grained and fine-grained localization procedures are performed, respectively. Ordinary wireless sensors are used for rough estimation of target position and camera sensors are used to determine accurate target location. However, it is assumed that targets are equipped with a wireless transceiver to calculate Received Signal Strength Indication (RSSI) and transmit information to sever. The calculation of target position is completed in the server and it is a centralized scheme. In our study, target localization can be completed by only one sensor and there is no need to equip with an additional localization device or signal transmitter on the targets.
Regarding the target tracking, numerous related works aimed for tracking target in ordinary WSNs [17–20]. Data delivery delay time and network lifetime were considered for target tracking in [17]. A heterogeneous wireless sensor network and a mobile wireless sensor network were used in the studies [18, 19], respectively. In [20], boundary nodes of a monitoring region can be found and used to know the entry and exit of a target through the monitoring region. For VSNs, the related work [16] used a hybrid sensor network composed of ordinary and camera sensors for target tracking. The sensors are regularly deployed and arranged in an array, which is different from a random deployment. The work [21] focuses on the solution of cooperative multiobject tracking among multiple camera sensors in a VSN established with a highly correlated sensing mode. In the approach, global information is needed for each camera sensor. In [22], a centralized 2-camera system was proposed for detecting and tracking humans in a realistic indoor home environment. The work [23] utilizes Graphics Processing Unit (GPU) coprocessors to compare image frames of a camera sensor, so as to acquire human position and velocity for tracking and handoff. In [24], multiple camera sensors which detected the same target will cooperate to track the target. An optimal sensor will be selected from these sensors according to a confidence measurement and this optimal sensor is mainly responsible for the tracking task. When the target is detected by a new camera sensor or there is a target movement, a new optimal camera sensor will be selected. The tracking handoff scheme needs the communication of cooperation data of these sensors. In our Voronoi-based handover protocol, there is no need of cooperation data of target detection and tracking of multiple sensors. In [25], an observation correlation coefficient (OCC) is defined as the ratio of the overlap area of two cameras to the FoV of one camera. It is used for the determination of activating a set of cameras to improve the accuracy of target tracking. The scheme works under the cooperation of multiple sensors which observed the target. More camera sensors need to be deployed in the monitoring field and bring coverage overlaps for performing the algorithm. Our proposed protocol can be performed without high overlapped coverage among sensors. On the contrary, our scheme make camera sensors reduce FoV overlaps to have a high overall field coverage and can track the target by a single camera sensor. Table 1 shows a summary of comparing the proposed scheme with the related works.
Summary of comparison with related works.
3. Preliminaries
3.1. Voronoi Diagram
In this study, the properties of a Voronoi diagram is utilized to propose the sensor handover protocol for target tracking in a VSN. A Voronoi diagram, as shown in Figure 1, has the following properties.
It divides an area into convex polygons, called Voronoi cells, according to a given set of points. Each of the given points lies in exactly one Voronoi cell. The common/shared edge of two adjacent Voronoi cells is the bisector line between the two points in the two cells. For any position For an area with a given set of points, the Voronoi diagram is unique.
A well-known algorithm to construct Voronoi diagram is Fortune's algorithm [26], which is a centralized algorithm. It can generate the Voronoi diagram with the global information of positions of a set of points in a plane. In a VSN, deployed visual sensors can be treated as the set of points for the construction of Voronoi diagram of the surveillance region. For any object which appears in a Voronoi cell, the sensor in the same cell can obtain the optimal sensing quality (the clarity of the picture) of the object and has the strongest reaction capacity to the object. Therefore, the concept of Voronoi diagram is used in the proposed protocol. However, in this study, Voronoi diagram is constructed, distributed, and localized by the deployed sensors without global information. It is not constructed by using a centralized algorithm which needs the global information. Each sensor constructs its own Voronoi cell with local information. Those local Voronoi cells that are constructed by all the deployed sensors will form a complete Voronoi diagram. The detail of the construction of Voronoi cell in this study is described in the Section 4.1.

A Voronoi diagram.
3.2. Assumptions
The basic assumptions in this study are described as follows.
The visual sensors are homogeneous and randomly deployed in the surveillance region at initial phase. They are stationary and direction rotatable. Each visual sensor can obtain its coordinates by a localization technology and has enough communication range or can use a multihop transmission method to transmit information to its neighbor sensors. The objects to be tracked have no positioning component. Their coordinates are originally unknown. The type of target object in the tracking system is determinate or it can be simply recognized by the shape of detected object. Accordingly, the height of the target object can be given approximately.
3.3. Visual Sensing Model
The visual sensors used in this study are directional. And as mentioned in previous Section 3.2 of assumptions, the sensors are direction rotatable. The effective sensing area (Field of View) of the visual sensor is in sector shape. The sensing model for the visual sensors in the proposed protocol is shown in Figure 2 and the related notations are listed in Table 2.
Notations for the sensing model.

Sensing model for visual sensors.
4. Voronoi-Based Sensor Handover Protocol (VSHP)
4.1. Local Voronoi Cell Construction and Sensing Coverage Enhancement
In this study, the concept of Voronoi diagram is utilized for dividing the surveillance region into numerous convex polygons, namely, Voronoi cells. The sensors deployed in the surveillance region can be treated as the points in a Voronoi diagram, and each sensor belongs to one and only one cell in the diagram. As mentioned in Section 3.1, a Voronoi diagram can be generated by the well-known Fortune's algorithm, which is a centralized algorithm, from a given set of points in a plane. However, this study aims at proposing a distributed protocol; the sensors only distributedly construct the local Voronoi cell of their own without the global information of the sensor network.
After the random deployment of sensors, each sensor will broadcast its coordinates to and receive coordinates from neighbors. Figure 3 illustrates the construction of a local Voronoi cell with only the local information of neighbor positions. The sensor

Construction of a local Voronoi cell.
Because the visual sensors are randomly deployed, their positions and sensing directions are also random at initial phase. A coverage enhancement is needed to reduce the overlaps of sensing coverage of these visual sensors. Then, the overall coverage ratio of the surveillance region can be improved for the task of target tracking. The algorithm in our previous work [27] is utilized for coverage enhancement in this study.
4.2. Object Detection
In a tracking system, object detection is necessary prior to the object tracking. For tracking in a VSN, the detection can be done by object segmentation from the camera scenes. Background subtraction [28–30] is an important approach for object segmentation in visual surveillance system, which compares two images to acquire the difference. A simple illustration of background subtraction is shown in Figure 4. A foreground intruder object can be extracted by comparing current image frame,

Dynamic background subtraction for object and movement detection.
4.3. Target Localization
Once an incoming object is detected by a visual sensor, the sensor will estimate the actual position of the object on the ground plane. To estimate the object position, firstly the direction of the object will be calculated and secondly the distance of the object. Regarding the object direction, as shown in Figure 5, the view from the visual sensor is limited by its angle of view and the captured scene is shown on the image in a rectangle shape. Things in a line of sight will appear in a vertical line on the image. As the illustration, the direction of the detected target object (a human) can be calculated by the following equations with the notations listed in Table 3:
Notations for calculation of target object direction.

Calculation of object (target) direction.
Regarding the distance of the object, as shown in Figure 6, the image of real object is projected reversely on the internal image sensor (usually a CCD or CMOS sensor) through the camera lens, and then electronic signals generated by the internal image sensor are processed with an image processor to form the digital image. Since the target object was assumed to be a height of given value, the distance between the target object and visual sensor can be calculated by the following equations with the notations listed in Table 4:
Notations for calculation of target object distance.

Calculation of object (target) distance.
If the height of the target object is not exactly equal to the given assumed value and there is a difference of
Let
As shown in Figure 7, (

Calculation of object (target) coordinates.
4.4. Sensor Handover
After the visual sensor obtains the coordinates of the target object, it adjusts the sensing direction toward the target and then detects object movement by dynamic background subtraction mentioned in Section 4.2. Once a movement of the target object is detected, the new position of the object will be calculated by the localization scheme described in Section 4.3 again. This is a routine procedure for local target tracking by the visual sensor until the tracking task is handed over.
To determine that whether the visual sensor will handover the task of tracking of the target object, firstly the sensor utilizes the structure of the constructed local Voronoi cell and divides the cell into several subregions. As shown in Figure 8, the Voronoi cell associated with the sensor

Subregions divided from a local Voronoi cell according to the vertices.
A target object The object distance is less than or equal to the sensing radius of the sensor:
The distance between the object and the sensor is less than or equal to the one between the object and the neighbor sensor on the opposite side of the bisector line (edge) facing the included angle of the subregion:
The object is located inside the two edges of the included angle of the subregion:
Equation (11) represents that the target object is not located outside the left edge
Once a sensor detected an object and ascertained that the object is located in one of the subregions of the local Voronoi cell, there will be four conditions, while the sensor keeps tracking the moving target object.
As shown in Figure 9(a), the target object can still be detected and it is still located in the same triangular subregion; that is to say, (9), (10), (11), and (12) are still satisfied under the same pair of vertices As shown in Figure 9(b), the target object can still be detected but it has just moved from original subregion to another subregion. For this condition, the sensor will calculate and obtain a new pair of vertices, As shown in Figure 9(c), the target object can still be detected, but the target object has just moved beyond the bisector line As shown in Figure 9(d), the target object is still in the triangular subregion, but it moved out of the sensing radius of the sensor

Different target tracking conditions.
The probability of that a target can be detected by a visual sensor is equal to the overall sensing coverage ratio of the VSN. In addition, once a target has been detected and under tracking, the probability of target missing is equal to
As described in Section 4.1, after visual sensors are deployed initially, an algorithm of sensing coverage enhancement is performed. Whenever a sensor keeps tracking a moving target object or several sensors are notified of adjusting their sensing directions to wait a possible appearance of the lost target, their sensing directions will be changed and the overlapped coverage can be increased; thus overall field coverage can be reduced. This will cause a negative effect on further object detection and tracking. Therefore, in the proposed protocol, those sensors that have changed the sensing direction will return to their original directions if the sensors have lost the target or waited the appearance of target for a long time,
4.5. Procedures of VSHP
Figure 10 summarizes the procedures of the proposed Voronoi-based sensor handover protocol for moving target tracking.

Procedures of the proposed VSHP.
With regard to the issue of energy consumption, the sensor operations about computation, communication, direction rotation, and photographing will consume energy of the sensor. This could cause sensors energy exhaustion and malfunction in the applications of wireless VSNs. A recent research topic and current trend about energy issue of sensor networks are the energy harvesting sensors [31–33]. Energy harvesting is one of the promising solutions to the problem of limited energy capacity in wireless sensor networks. It is easy to foresee that a future sensor network can consist of sensor devices with the integration of the technology of energy harvesting. This paper aims at the target localization and tracking in VSNs; therefore energy issue is not concerned. However, our proposed algorithm can still perform well, while some of the sensors are malfunctioning. This is due to the fact that the proposed Voronoi-based algorithm makes visual sensors be able to easily reconstruct the local Voronoi cells and can keep the VSN tracking operations well performing until most (or all) of the sensors are failed. It has the characteristic of fault tolerance and graceful degradation.
On the other hand, the tracking algorithm described above focuses on the case of tracking single target with one visual sensor. It is applicable to the case of which multiple targets occur in the surveillance region and can be, respectively, tracked by one sensor. Once the multiple targets occur in one Voronoi cell at the same time, they only can be all tracked if
5. Simulation Results
This study evaluates the proposed Voronoi-based approach with simulations. There are four QoS criteria to be used for the evaluation:
Parameters setting of the simulations.
Figure 11 shows the simulation results of both coverage and target-detected latency. OVSN-C and VSHP-C represent the overall field coverage ratios after deployment of ordinary visual sensor network (OVSN) and after enhancement with proposed VSHP, respectively. OVSN-L and VSHP-L represent the ratios of target-detected latency under OVSN and proposed VSHP approach, respectively. The ratio of target-detected latency (

Coverage and target-detected latency (fixed
The total travel time of a target in the surveillance region consists of target-tracked and target-untracked durations. Keeping a higher target-tracked time ratio is an important criterion for target tracking services. Figure 12 shows that VSHP provides a higher target-tracked ratio than OVSN. This is due to the efficacy of the Voronoi-based sensor handover mechanism.

Target-tracked ratio (fixed
The other criterion for measuring the QoS of target tracking in a VSN is the average distance between the sensor and the target, while the target is tracked (monitored) by the sensor. For a visual (camera) sensor, a shorter target distance will bring a clearer (higher) quality of image or video. Figure 13 shows the simulation result of average target distance of which the target is tracked. The proposed VSHP has a shorter average target distance than OVSN does. This is due to the fact that the proposed handover protocol utilizes the characteristic of Voronoi cell to select an appropriate sensor for the tracking task. Moreover, the average target distance is reduced from 40 m to 35 m, while the number of sensor is increased from 50 to 100. This indicates that the VSHP can select more appropriate sensors from the larger number of deployed sensors for the target tracking.

Average target distance (fixed
The results given above are with various values of

Coverage and target-detected latency (fixed

Target-tracked ratio (fixed

Average target distance (fixed
6. Conclusion and Future Works
This paper utilizes the structure and characteristic of Voronoi cells and proposes a new solution for target tracking in visual/camera sensor networks. The solution contains mechanisms of coverage enhancement, object detection, target localization, and sensor handover. Simulations were used for the evaluation of effectiveness of the proposed approach. Four QoS criteria for target tracking were evaluated. The results show that the approach performs well and has an improvement in comparison with target tracking in ordinary visual sensor networks.
Our future work is to implement a practical system with real camera sensors. An experimental evaluation of effectiveness and performance for the practical system will be made. Moreover, the utilizations of mobile visual sensors (e.g., smartphones) and the integration with cloud-based monitoring service will be developed and realized for an extension of this study.
