Abstract
Introduction
Location is highly critical to many services provided by wireless sensor networks (WSNs), 1 such as geographic routing, wildlife monitoring, and health caring. Localization is a fundamental issue of WSNs that has been extensively studied in the literature. In daily life, the global positioning system (GPS) 2 is widely applied to achieve self-localization. However, there are still some situations where it does not work well (e.g. indoors or under the ground). Moreover, having each target GPS-equipped is extremely infeasible and expensive for WSNs.
The limitations of GPS have promoted a large body of localization approaches, which can be classified into two categories: range-based and range-free approaches. Range-based approaches are simple but susceptible to fading, noise, and non–line of sight. Range-free approaches perform localization by exploiting the connectivity between targets and sensors. They do not need extra hardware support but are usually imprecise and easily sensitive to the density of sensors. More seriously, both range-based and range-free approaches are computationally inefficient as they require the exchange of a large number of data to measure distances or determine connectivity. However, in most applications, sensors are resource-constrained (e.g. low power, low memory, and low operational ability), leading these solutions impractical. Therefore, it is quite necessary to develop a localization approach with less data collection and processing.
Compressive sensing (CS)3,4 offers a novel solution to the localization problem in WSNs. As a novel signal processing technique, CS theory asserts that a small number of measurements will suffice for recovering the original sparse or compressible signals. Since the number of targets in the region of interest is limited, localization benefits from CS, sampling number can be greatly reduced. As a consequence, CS attracts considerable attention in the localization field, and a lot of localization approaches based on CS are proposed that will be discussed later. In these approaches, the continuous physical space is discretized into a fixed grid which corresponds to a dictionary. By assuming that all targets fall exactly on some grid points, the measurement vector can be sparsely represented as a sparse linear combination of the dictionary atoms. Then, localization is accomplished by sparse signal recovery followed by support detection.
Unfortunately, as a matter of fact, the assumption usually does not establish that all targets fall exactly on a fixed grid. It is noteworthy that there always exist targets that deviate from the fixed grid no matter how fine the space is sampled. In such a case, there exists mismatch between the assumed and actual sparsifying dictionaries. This is the so-called “dictionary mismatch” problem. Existing research indicates that the existence of dictionary mismatch will deteriorate the performance of CS dramatically.5,6
Since CS has been focused on the signals that can be sparsely or compressibly represented under a finite dictionary, discretization of the physical space is inevitable. It is intuitively reasonable that both dictionary mismatch and recovery error can be reduced with a dense grid. Therefore, one naturally wonders whether a denser grid leads to more accurate sparse signal recovery. In fact, according to CS theory, the sampled grid should not be too dense. This is because the dictionaries, corresponding to densely-sampled grids, have high inter-column correlation. The high correlation of dictionary atoms violates the restricted isometry property (RIP) condition of CS. 19 In fact, the performance of CS may degrade considerably in the presence of dictionary mismatch, even when the physical space is fine-grained discretized.
In this article, we study the dictionary mismatch problem and develop an efficient dictionary refinement algorithm. The key idea is to dynamically adjust the grid to alleviate or even eliminate the mismatch between the assumed and the actual sparsifying dictionaries. First of all, we regard the sparsifying dictionary as a parameterized dictionary, with the sampled grid as the underlying parameters. Consequently, the original counting and localization problem is formulated as a joint sparse signal recovery and grid parameter estimation problem. Then, an iterative two-step algorithm is developed that alternately optimizes over the sparse signal and grid parameters in a manner of optimizing over one while keep another fixed. Therefore, the original joint optimization problem is transformed into two sub-problems that can be effectively solved by existing optimization tools. Finally, we demonstrate via simulation that the proposed dictionary refinement algorithm performs significantly better than the state-of-the-art CS reconstruction algorithms.
The main contributions of this article can be summarized as follows:
We study the dictionary mismatch problem in CS-based localization and develop an efficient dictionary refinement algorithm. Based on the algorithm, the grid can be dynamically adjusted to alleviate or even eliminate the mismatch between the assumed and actual sparsifying dictionaries.
To achieve dictionary refinement, we view the sparsifying dictionary as a parameterized dictionary, with the sampled grid as the underlying parameters. Therefore, the original counting and localization problem is formulated as a joint sparse signal recovery and grid parameter estimation problem.
To solve the joint optimization problem, we decompose it into two sub-problems and develop an iterative two-step algorithm to solve the sub-problems. In each iteration, the sparse signal and grid parameters are updated in turn.
We conduct extensive simulations to evaluate the performance of the proposed dictionary refinement algorithm. The superiority of our algorithm compared with other sparse reconstruction algorithms is confirmed by the simulation results.
The rest of this article is organized as follows. Section “Related work” gives a review on related work. A brief background information about CS is presented in section “CS.” Section “Network model and problem statement” provides the network model and problem statement. In section “Problem formulation,” we mathematically formulate the counting and localization problem. Section “Localization via dictionary refinement” provides detailed descriptions on the proposed dictionary refinement algorithm. The performance of our algorithm is demonstrated in section “Numerical evaluation.” Finally, conclusion is summarized in section “Conclusion.”
Notations used in this article are as follows.
Related work
In this section, we review the related work in the field of CS-based localization. Cevher et al. 7 propose to apply CS to target localization for the first time. The localization problem is transformed into a sparse recovery problem by sampling the physical space into a grid. In addition, a Bayesian framework 8 is proposed and the sparse approximation to its optimal solution is also provided. However, the drawback that a dictionary needs to be maintained at each sensor limits its applications. Meanwhile, the demand of communication among sensors also leads to poor performance in sparse WSNs. Feng et al. 9 model the locations of multiple targets as a sparse matrix and propose a CS-based indoor localization approach. Although it declares to be able to localize multiple targets, the approach can localize only one target once in that the data compression is not sufficient enough. In Feng et al.,10,11 a clustering idea is further introduced to reduce system cost. The localization performance relies heavily on the clustered results and cluster matching algorithms. Nevertheless, in general, it is difficult to choose a proper clustering scheme and an effective cluster matching algorithm. Zhang et al. 12 consider multiple target localization without the prior knowledge of target number. Different from Feng et al., 9 the authors view the locations of multiple targets as a sparse vector whose sparsity corresponds to the number of targets. Then, a greedy matching pursuit (GMP) algorithm is proposed to recover the sparse signal with a high probability. Au et al. 13 apply CS to develop a positioning system which consists of a coarse stage and a fine stage. The coarse stage is executed to obtain the approximate positions of a target using a proximity constraint and then more accurate position is obtained by CS in the fine stage. Guyen et al. 14 propose an indoor positioning system where the radio map is decomposed into a dictionary and a sparse representation matrix using the K-SVD dictionary learning algorithm. Then, the real-time received signal strength (RSS) vector is matched with the columns of the sparse representation matrix. The reference point with the least matching error is determined to be closest to the target. In fingerprint-based localization approaches, the construction of fingerprint database is usually time-costing. By exploiting the spatial and temporal relativity, Zhang et al. 15 develop a new fingerprint database construction method with only a few fingerprint collection. Different from the above-mentioned range-based methods, Liu et al. 16 develop a range-free CS-based localization method. Instead of collecting RSS measurements directly, the information whether targets are in the range of sensors are utilized to achieve localization at the cost of unreliable accuracy. Nguyen and Shin 17 develop a CS-based localization method which collects RSS measurements according to a deterministic sensing matrix rather than random sensing matrix. By exploiting the sparsity of spatial signals and the compressibility of temporal signals, Sun et al. 18 develop a two-dimensional (2D) localization framework using CS.
In the above-mentioned solutions, a common assumption is made that all targets fall exactly on a fixed grid. However, it should be noted that there always exist targets that deviate from the grid no matter how fine the space is sampled. In such a case, the traditional CS-based localization methods perform poor. Additionally, it should be noted that we assume each RSS measurement taken by a sensor is the sum of the strengths of the received signals that come from all targets while the sensor cannot distinguish the signals from different targets. Therefore, it is impossible to directly derive distance estimates or approximation information between sensors and targets. As a result, the traditional range-based or range-free localization methods cannot be applied in our context.
CS
Consider a discrete signal given by the vector
However, it is worthy noting that few signals in practice are truly sparse or compressible themselves; rather they can be sparsely or compressibly represented as
where
Given
Network model and problem statement
Network model
Without loss of generality, we consider the problem of localization for
where
where

(a) The scenario of multiple target localization and (b) an example of grid numbering.
Problem statement
Since the number of targets is limited, localization benefits from CS, the measurement number can be greatly reduced. In order to implement CS, we divide the area of interest into a grid. The grid lines are denoted by vectors
as the atom formed by the
Conventional localization approaches based on CS pre-suppose that all targets fall exactly on the discretized grid, namely,
where
However, in general, the targets may not fall exactly on the discretized grid no matter how fine we discretize the physical space. The existence of off-grid targets leads to the mismatch between the assumed dictionary and the actual dictionary. As a matter of fact, in practice, a small dictionary mismatch may deteriorate the signal recovery performance dramatically.
Problem formulation
To address such a dictionary mismatch problem, an effective solution is to iteratively optimize over the grid parameter
where the regularization parameter
Initially, the continuous space is discretized as a uniform grid

An illustration of the upper and lower bounds of the x-axis grid lines.
Then, we can express the upper and lower bounds of the x-axis grid lines as
where
Considering the bounds, problem (8) can be rewritten as
Furthermore, it is worth noting that the representation coefficient
where
Localization via dictionary refinement
In this section, we first discuss how to solve problem (14) and then present how to achieve accurate localization using the solution.
Dictionary refinement algorithm
The process of the proposed dictionary refinement algorithm is summarized in Algorithm 1. The algorithm starts with an initial equi-spaced grid sampling
We develop an iterative two-step algorithm to solve problem (14). The key idea of the algorithm is to alternately optimize over sparse signal
where superscripts denote algorithm iteration. This is a standard regularized LS optimization problem, and there exist many tools to solve this problem. In our implementation, we use the convex optimization software package. 24
In practice, the solution
where
The second step optimizes over the grid parameters
We solve the problem using a trust-region subspace method
25
where the gradient of objective function must be provided. For the sake of convenience, we transform the grid parameters
Let
Denote by
where
After each iteration, the current system cost is calculated as
Remark 1
Algorithm 1 is a two-layer iteration procedure. In each outer iteration, sparse signal
Remark 2
The algorithm stops when the convergence condition is reached, for example, when the iteration has been executed the maximum outer iteration number
Localization
The output results
where
Numerical evaluation
In this section, we conduct simulations to evaluate the effectiveness and robustness of the proposed dictionary refinement algorithm.
Simulation setup
All simulations are conducted on MATLAB R2015b. We consider a
In order to assign an estimated location to a target, we compute all pairs of the distances between
Definition 1
The average localization error, denoted by
where
We have observed from simulations that the decrement of objective function usually becomes quite small within very few iterations for trust-region subspace method. Based on the observation, to reduce time cost, we set the maximum inner iteration number
Parameter values for simulations.
Simulation without off-grid targets
We first consider an ideal situation where all targets are located exactly on the initial grid. The true locations of three targets are set to (4, 3), (8, 6), and (5, 8). Figure 3 shows the counting and localization results of different algorithms. It can be seen that all algorithms accurately estimate the number of these targets. In terms of localization, it can be seen that BP, SBL, and DicRef are able to accurately estimate the locations of multiple targets. In contrast, OMP fails to localize the first target.

Counting and localization results of different algorithms when all targets fall exactly on the initial grid.
Simulation with off-grid targets
Then, we consider a more general situation where there exist targets deviating from the initial grid. The true positions of three targets are set to (3.85, 3.47), (8.42, 5.84), and (5.35, 8.26), respectively. In such case, all targets are located off the initial grid. Figure 4 shows the counting and localization results of different algorithms.

Counting and localization results of different algorithms when there exist targets deviating from the initial grid.
As can be shown in Figure 4, when there is no noise, BP, SBL, and DicRef correctly estimate the number of targets, while OMP suffers from false targets. As for localization, the traditional CS reconstruction algorithms localize three targets to the nearest grid points. Unsurprisingly, these behaviors are the natural results incurred by dictionary mismatch. In contrast, the positions estimated by DicRef are so accurate that they are visually indistinguishable from their original positions. This example clearly indicates that, compared with the traditional CS reconstruction algorithms, the proposed dictionary refinement algorithm is able to provide better performance when there are targets deviating from the initial grid.
Effect of measurement noise
It is inevitable for measurements to be corrupted with environmental noise. In order to check the robustness of the proposed dictionary refinement algorithm against measurement noise, we intentionally add Gaussian white noise
Figure 5 reports the average estimated target numbers, probabilities of correct counting, and average localization errors of different algorithms under different measurement noise levels. As can be observed, with the increasing of SNR, the average estimated target numbers of all algorithms gradually approximate to the true target number, which is indicated by the horizontal baseline in Figure 5(a). When SNR is low, the proposed dictionary refinement algorithm performs much better than the other algorithms. As SNR is high, BP, SBL, and DicRef can estimate the number of targets accurately while OMP suffers from false targets. As for the probability of correct counting, the behaviors of different algorithms are similar with the average estimated target numbers. The probabilities of correct counting with respect to all algorithms increase with the increasing of SNR. When noise is serious, DicRef performs better than the other algorithms. As noise is slight, the probabilities of correct counting with respect to BP, SBL, and DicRef approximate to 1. However, the probability of correct counting of OMP keeps below 0.21 in the entire SNR range. Additionally, the average localization errors for all algorithms decrease as the SNR increases. As a matter of fact, this is reasonable as the signal reconstruction accuracy is proportional to SNR. Another important observation is that the proposed dictionary refinement algorithm performs better than the other algorithms no matter how high the noise level is. The reason lies in the fact that the proposed algorithm dynamically adjusts the grid to refine the dictionary while the other algorithms do not. Furthermore, we can see that the proposed dictionary refinement algorithm can tolerate a certain level of measurement noise. For example, when SNR is higher than 30 dB, the average localization error with respect to DicRef is no more than 0.1 m.

Counting and localization performance versus SNR when measurement number
Then, to further examine the performance of the proposed dictionary refinement algorithm, we compare the cumulative distribution functions of the average localization errors of different algorithms when SNR

Cumulative distribution functions (CDFs) of average location errors.
Effect of measurement number
Now, we check the performance of the proposed dictionary refinement algorithm with different measurement numbers. We set target number

Counting and localization performance versus measurement number when
Results on different parameters
Then, to further test the performance of the proposed algorithm, we sample the same area into a denser grid with
Complexity analysis
Eventually, we analyze the complexities of different algorithms in terms of computational complexity and CPU running times. The results are shown in Table 2. For BP, OMP, and SBL, the computational complexity are
Complexity analysis of different algorithms.
BP: basis pursuit; OMP: orthogonal matching pursuit; SBL: sparse Bayesian learning.
Conclusion
In this article, we investigated the dictionary mismatch problem in CS-based localization and developed an efficient dictionary refinement algorithm. Different from the other CS reconstruction algorithms using a fixed grid, the proposed dictionary refinement algorithm dynamically adjusts the grid to alleviate or even eliminate dictionary mismatch. To achieve this, we view the sparsifying dictionary as a parameterized dictionary, with the sampled grid as adjustable parameters; then the localization problem is transformed into a joint sparse signal recovery and parameter estimation problem; at last, an iterative two-step algorithm is developed to solve the joint optimization problem. Simulation results show that, compared to the most existing CS reconstruction algorithms, the proposed dictionary refinement algorithm achieves better performance with high time cost. In the future, we will seek for alternative solutions to reduce the cost of the proposed algorithm at the expense of allowed performance loss.
