We study the performance of diffusion LMS (Least-Mean-Square) algorithm for distributed parameter estimation problem over sensor networks with quantized data and random topology, where the data are quantized before transmission and the links are interrupted at random times. To achieve unbiased estimation of the unknown parameter, we add dither (small noise) to the sensor states before quantization. We first propose a diffusion LMS algorithm with quantized data and random link failures. We further analyze the stability and convergence of the proposed algorithm and derive the closed-form expressions of the MSD (Mean-Square Deviation) and EMSE (Excess Mean-Square Errors), which characterize the steady-state performance of the proposed algorithm. We show that the convergence of the proposed algorithm is independent of quantized data and random topology. Moreover, the analytical results reveal which factors influence the network performance, and we show that the effect of quantization is the main factor in performance degradation of the proposed algorithm. We finally provide computer simulation results that illustrate the performance of the proposed algorithm and verify the results of the theoretical analysis.
1. Introduction
Wireless sensor networks have recently been studied among researchers from the fields of communications, computer technology, signal processing, controls, and many others [1–3], because they have many applications [4–7]. In wireless sensor networks, distributed parameter or state estimation is a very important research aspect. Meanwhile, distributed adaptive estimation is extremely attractive and challenging for theoretical analysis and applications in recent years. Adaptive networks [8, 9] are a prevailing solution to distributed adaptive estimation problem. The adaptive network consists of a collection of nodes observing temporal data arising from different sources with possibly different statistical profiles. These nodes are interconnected to each other and required to estimate and infer some parameters of interest from noisy measurements in a collaborative manner. The existing strategies that enable learning and adaptation over such network can be roughly classified into incremental strategy [10–18], consensus strategy [19, 20], and diffusion strategy [8, 9, 21–28]. In incremental strategy, a cyclic path is first determined over a network, and then information is passed from one node to the next node over the cyclic path, repeating the process until all nodes are visited. However, it is difficult to determine the cyclic path in network. Consensus and diffusion strategy do not need to determine the cyclic path. Compared with consensus strategy, diffusion strategy outperforms consensus strategy for distributed parameter estimation. Moreover, when the step sizes are constant to enable continuous learning, diffusion strategy can enhance adaptation performance and widen stability ranges [29]. For these reasons, we focus on diffusion strategy in this paper.
In wireless sensor networks, communication with unquantized values is impractical due to the power and bandwidth constraints, which prevents the exchange of high-precision (analog) data among sensors. Therefore, quantization is usually required before exchanging data through internetworks communications [30–32]. What is more, node that receives data from the neighbors loses certain information because the quantization procedure induces some noise to the original data. This makes the problem more challenging from the previous works [21]. In distributed parameter estimation problem, the quantization issue is considered in consensus strategy [32–34], but it has not been considered in diffusion strategy. Meanwhile, because the wireless environment is very complicated, nodes and links may be subject to failure. Therefore, the randomness of network topology is an important topological characteristic in wireless networks, and then it is considered in diffusion algorithm [35]. In conclusion, the data must be quantized and the network topology is random in wireless sensor networks.
However, the existing works in diffusion algorithm do not consider the effect of quantization (see [21–28] and references therein). Nevertheless, due to the power and bandwidth of each sensor node constrain in sensor network, the quantization is necessarily considered, and random topology is also considered due to wireless environment. These issues are our motivation for this paper. To the best of our knowledge, this paper presents the first performance analysis of diffusion LMS algorithm, which considers the quantization together with random topology. Our objective in this paper is to analyze the convergence and steady-state performances of the proposed algorithm and derive some closed-form expressions of MSD and EMSE that characterize the steady-state performance of the proposed algorithm.
The main contributions are as follows:
Some theoretical results of diffusion LMS algorithm with quantized data and random topology are derived. Some models for the transient and steady-state behavior of the diffusion LMS algorithm are obtained. Explicit expressions are derived for variance relation which contain moments that represent the effects of quantization and random topology.
We study the performance of diffusion LMS with quantized data and random topology by deriving closed-form expressions for Mean-Square Deviations (MSD) and Excess Mean-Square Errors (EMSE) for Gaussian data and sufficiently small step sizes. Meanwhile, we obtain the closed-form expressions for MSD and EMSE to explain the steady-state performance at each individual node.
The convergence of the dithered quantized diffusion LMS algorithm in networks with random links is proved. What is more, the effect of quantized error is the main factor in performance degradation of the diffusion LMS algorithm.
The remainder of this paper is organized as follows. Section 2 briefly describes the diffusion LMS algorithm without quantized data and with fixed topology. In Section 3, we propose a diffusion LMS algorithm with quantized data and random topology. In Section 4, we analyze the mean-square convergence of the modeled diffusion algorithm. The steady-state performances are obtained in Section 5. Simulation results and analysis are given in Section 6. Finally, conclusions are described in Section 7.
Notation. In order to distinguish between deterministic variables and random variables, we use boldface letters to represent random quantities and normal font represents their deterministic (nonrandom) quantities. In this paper, all vectors are column vectors, with the exception of the regression vector. We use to denote a column vector with entries a and b stacked on top of each other and use to represent a (block) diagonal matrix with entries a and b. The notation denotes complex conjugation for scales and Hermitian transpose for matrices, and represents transposition for matrices. The notation denotes the identity matrix of sizes . The notations and denote the sets of real numbers and integer numbers, respectively.
2. Diffusion Algorithm without Quantized Data: Fixed Topology
We consider a network consisting of N nodes distributed over a spatial domain. In a connected network, two nodes are said to be neighbors if the nodes may be connected directly by an edge; that is, the nodes can share information. Consider a network with a fixed topology; we define the neighborhood of node k, which consists of neighbors of node k and node k itself. The neighborhood of node k is denoted by . At every time index i, each node k is able to obtain a realization , , where is a scalar measurement and is a regression row vector. The data arise from a jointly wide-sense stationary random processes , where is a random variable with zero mean and covariance matrix of the form . In this paper, to estimate some unknown column vector , we assume that the measurements satisfy a linear regression model of the following form [36]:
where is the measurement noise. We further assume that is independent and identically distributed over time and space and with zero mean and variance , and is independent of for all l.
The objective of the network is to estimate the column vector by minimizing the cost function (2) below in a fully distributed manner and in real-time. The cost function is defined as follows:
where denotes the expectation operator with respect to the probability space of interests. By minimizing the cost function , the optimal solution satisfies the following equation:
where and . For convenience of presentation, we assume that is positive-definite; that is, . Hence, is invertible; then, the optimal solution satisfies the normal equation [36]:
However, the optimal solution cannot be determined from the solution of the normal equation, since are not known beforehand in practical. Therefore, in order to solve , the CTA (Combine-then-Adaptive) diffusion LMS algorithm is proposed in [21]:
where is the local step size and is the local estimate for the parameter that is evaluated by node k at time index i. are nonnegative coefficients, which can be treated as free weighting parameters are chosen by the designer, and satisfy
In this section, we briefly describe the diffusion LMS algorithm without quantized data and with fixed topology. However, the sources (e.g., power and bandwidth) are limited and the topology is random in wireless sensor networks. Therefore, we consider diffusion LMS algorithm with quantized data and random topology in Section 3.
3. Diffusion LMS Algorithm with Quantized Data: Random Topology
In this section, we consider the effects of quantization and random topology in diffusion LMS algorithm. Therefore, we first establish random topology model and quantization model and then derive diffusion LMS algorithm with quantized data and random topology.
3.1. Random Topology Model
In this paper, we consider an undirected graph for simplicity; that is, , where are coefficients at time index i. To model the topological randomness, we assume that the nodes and links are random entities. We assume that, at any given time index i, the coefficient (or link weight) , which connects node k to node l, will either assume a nominal value with probability (since we assume , therefore, we can let ) or be zero with probability ; that is,
where .
Since the link failures exist in the network, therefore, we assume a normal topology , including a fixed number of nodes N and edges (or links). Thus, the edges give rise to different subnetworks , with each subnetwork with probability , composed of faultless and faulty links.
Since the topology matrix is random, thus we define the mean topology matrix and , as [35]
where , , and . The symbols ⊗ and ⊙ denote the Kronecker product and block Kronecker product, respectively.
3.2. Dithered Quantization Model
In this paper, we assume that each internode communication channel uses a uniform quantizer with quantization step Δ. To model the communication channel, we introduce the quantizing function, , ; that is,
where is the channel input. Therefore, can be written as follows:
where is the quantization error, which satisfies
Since the error terms are not stochastic, which fail to lead to a reasonable solution, therefore, we introduce dither to randomize the node states prior to quantizing the perturbed stochastic state. As [33], the quantization error sequence, , is defined as follows, when dither is added before quantization:
where and are random sequences. Further, the sequence is independent of the sequence . If the dither sequence, , satisfies the Schuchman condition [37], then in (13) is independent and identically distributed random variable and uniformly distributed on . The quantization error sequence is independent of the input sequence [38, 39].
3.3. Dither Quantized Diffusion with Random Topology
We now return to the problem formulation of diffusion LMS algorithm with random topology and quantized data. In the diffusion LMS algorithm, we consider both quantized data and random topology in the network. Therefore, we introduce the sequence vector , where each entity in the vector is independent and identically distributed random variable, uniformly distributed on . Then, using (13), the state update equations in (5) and (6) change to
where is the neighborhood of node k at time i and and are defined as follows:
According to the properties of the dither sequence and the quantization error sequence , and are zero mean random vector and have i.i.d. (independent and identically distributed) components, independent of , with zero mean and with covariance matrices:
where and .
In order to facilitate the analysis, we introduce the quantities as follows:
Using (1) and definitions (20), (21), (23), and (26), we have
With this relation, (14) and (15) can be written as follows:
Expression (29) can be rewritten in a more compact state-space form as follows:
In this section, we derive the diffusion LMS algorithm with quantized data and random topology. In Sections 4 and 5, we analyze the performances of the proposed algorithm.
4. Mean-Square Convergence Analysis
It is well known that studying the performance of single stand-alone adaptive filters is a challenging work. We now face a network of adaptive network, where the nodes influence each other's behavior. Therefore, its performance analysis is more complicated than the single adaptive filter. In order to proceed with the analysis, we introduce the weight error vector:
Noting that , by subtracting from both sides of (30) and using the expression (27), we get
We assume that the regression data are temporally and spatially independent random variables; taking expectation of both sides of (32) leads to
where is a block diagonal matrix and . Therefore, from (33), we can see that the mean of the weight error vector depends on the mean topology matrix and on the block diagonal matrix .
Theorem 1.
Assume the fact that data model (27) and the regression data are temporally and spatially independent holds. Then, all estimators converge in the mean to the optimal solution if the positive parameters satisfy the following condition:
where denotes the maximum eigenvalue of Hermitian matrix .
Proof.
It follows from (33) that the weight error vector converges in the mean to zero if and only if the matrix is a stable matrix. However, if the matrix is stable, then the matrix is stable. Therefore, the matrix is stable if and only if the positive step sizes parameters are selected to satisfy
where is the spectral radius of matrix . Therefore, we have
or
Thus, the matrix is stable if and only if the parameters satisfy
It follows that
Namely,
5. Steady-State Performance Analysis
5.1. Variance Relations
In order to study the mean-square performance of the diffusion LMS algorithm with quantized data and random topology, therefore, we introduce the variance relation of weight error vectors . From expression (32), we get
where denotes the real part of its matrix argument, Σ is some arbitrary Hermitian positive-definite matrix, for any column vector x, and
However, the closed form of the variance relation is difficult to derive. So we consider that the regressor data arise from circular Gaussian sources with zero mean in the next subsection.
5.2. Gaussian Data
In order to evaluate the network mean-square performance of each individual node by using the above variance relation, thus, we need to calculate the data moments. In particular, since the last terms in (42) and (43) are difficult to calculate in closed form for arbitrary distribution of the data, thus, we assume that the regressors arise from circular Gaussian sources with zero mean [36]. We define the transformed quantities as follows:
where we introduce the eigendecomposition , T is unitary, and , is a diagonal matrix, and follows from (22). Therefore, expressions (41), (42), and (43) change to
For Gaussian data sources, expression (45) is rewritten by the following recursion:
where
For the derivation of (48), see the Appendix. In the above, we are using vector σ to replace its matrix representation Σ. In the next subsection, we use (48) and (49) to capture the essence of stability behavior of the diffusion LMS algorithm with quantized data and random topology.
5.3. Steady-State Performance
We now analyze the steady-state performance of the diffusion LMS algorithm with quantized data and random topology. We consider the steady-state quantities MSD and EMSE, which are defined as
Compared with in [21], the terms , , and are added, since we consider that the sensor data is quantized. Thus, according to a similar way in [21], we obtain the global MSD and EMSE as follows:
where and .
In order to analyze the performance of individual node, we introduce the following quantities:
where is a block of zeros and is the diagonal matrix with the eigenvalues of node k. Local MSD and EMSE can measure the steady-state of the individual node. Therefore, following a similar way in [21], the local MSD and EMSE are obtained as follows:
In this section, we derive the expressions of the global MSD and EMSE and local MSD and EMSE. We will verify theoretical results by simulations in Section 6.
6. Simulation and Analysis
In this section, we compare the theoretical expressions with simulation results by computer simulation. In order to achieve this aim, we consider a sensor network with and the network topology as shown in Figure 1. In simulation, we set , , and for all . Meanwhile, we assume the regression data generated via Gaussian 1-Markov sources with covariance matrices, where maximum eigenvalue of each covariance matrix is set to 5 and minimum eigenvalue of each covariance matrix is set to 1. According to the generation of the regression data, the mean of regression vector is zero. The statistical profiles for simulation are shown in Figure 2.
Network topology.
Statistical profiles for simulation.
The global MSD and EMSE curves of diffusion LMS algorithm with static topology and unquantized data and diffusion LMS algorithm with random topology and quantized data are shown in Figures 3 and 4, which are obtained by averaging across all nodes and over several experiments. As shown in Figures 3 and 4, the performances of the diffusion LMS algorithm are decreased because of quantization and random topology. What is more, the effect of quantization is the main factor in performance degradation of the diffusion LMS algorithm.
Global MSD curve for diffusion LMS algorithm.
Global EMSE curve for diffusion LMS algorithm.
The local MSD and EMSE curves of diffusion LMS algorithm are shown in Figures 5 and 6, which show the network performances of the individual node in steady-state. As shown in Figures 5 and 6, a very good match between closed-form expressions and simulations is presented.
Local MSD curve for diffusion LMS algorithm.
Local EMSE curve for diffusion LMS algorithm.
7. Conclusion
In this paper, we studied the distributed estimation problem based on diffusion LMS (Least-Mean-Square) algorithm in wireless sensor networks with quantized data and random topology. First, we established weighted spatial-temporal energy conservation relation. The mean stability of diffusion LMS algorithm with quantized data and random topology is analyzed, and the analysis result shows that the mean stability is independent on quantized data and random topology. We derived a variance relation simultaneously. The closed-form expressions that describe the steady-state performance in terms of the MSD (Mean-Square Deviation) and EMSE (Excess Mean-Square Errors) quantities are derived. The results show that the effect of quantization is the main factor in performance degradation of the diffusion LMS algorithm with quantized data and random topology. Meanwhile, the simulation results show the good match between the closed-form expressions and simulations.
Footnotes
Competing Interests
The authors declare that they have no competing interests.
Acknowledgments
This work was partially supported by the National Basic Research Program of China (973 Program) under Grant no. 2013CB329102,by the National Science and Technology Major Project under Grant no. 2015ZX03003002-002,by the National Natural Science Foundation of China (NSFC) under Grant nos. 61372112,61232017,61522103,61370221,and U1404611,and in part by the Program for Science & Technology Innovation Talents in the University of Henan Province under Grant no. 16HASTIT035.
References
1.
EstrinD.GovindanR.HeidemannJ.KumarS.Next century challenges: scalable coordination in sensor network'sProceedings of the 5th ACM/IEEE International Conference on Mobile Computing and Networking1999263270
2.
CassandrasC. G.LiW.Sensor networks and cooperative controlEuropean Journal of Control2005114-543646310.3166/ejc.11.436-463MR2201570ZBL1293.930692-s2.0-33645395957
3.
GharaviH.KumarS. P.Special issue on sensor networks and applicationsProceedings of the IEEE20039181151115310.1109/jproc.2003.8149252-s2.0-3042745888
4.
LiD.WongK. D.HuY. H.SayeedA. M.Detection, classification, and tracking of targetsIEEE Signal Processing Magazine2002192172910.1109/79.9856742-s2.0-0036503203
5.
XuC.LiuT.GuanJ.ZhangH.MunteanG.-M.CMT-QA: quality-aware adaptive concurrent multipath data transfer in heterogeneous wireless networksIEEE Transactions on Mobile Computing201312112193220510.1109/tmc.2012.1892-s2.0-84884689328
6.
XuC.ZhaoF.GuanJ.ZhangH.MunteanG.-M.QoE-driven user-centric VoD services in urban multi-homed P2P-based vehicular networksIEEE Transactions on Vehicular Technology20136252273228910.1109/tvt.2012.22286822-s2.0-84876577335
7.
XuC.ZhaoJ.MunteanG.-M.Congestion control design for multipath transport protocols: a surveyIEEE Communications Surveys & Tutorials201610.1109/comst.2016.2558818
8.
SayedA. H.TuS.-Y.ChenJ.ZhaoX.TowficZ.Diffusion strategies for adaptation and learning over networks: an examination of distributed strategies and network behaviorIEEE Signal Processing Magazine201330315517110.1109/msp.2012.22319912-s2.0-84876115108
9.
SayedA. H.Adaptive networksProceedings of the IEEE2014102446049710.1109/jproc.2014.2306253
10.
TsitsiklisJ. N.AthansM.Convergence and asymptotic agreement in distributed decision problemsIEEE Transactions on Automatic Control1984291425010.1109/tac.1984.1103385MR734244
11.
TsitsiklisJ. N.BertsekasD. P.AthansM.Distributed asynchronous deterministic and stochastic gradient optimization algorithmsIEEE Transactions on Automatic Control198631980381210.1109/tac.1986.1104412MR8534312-s2.0-0022783899
12.
BertsekasD. P.A new class of incremental gradient methods for least squares problemsSIAM Journal on Optimization19977491392610.1137/S1052623495287022MR1479606ZBL0887.490252-s2.0-0031285678
13.
NedićA.BertsekasD. P.Incremental subgradient methods for nondifferentiable optimizationSIAM Journal on Optimization200112110913810.1137/s1052623499362111MR18705882-s2.0-0036342213
14.
RabbatM. G.NowakR. D.Quantized incremental algorithms for distributed optimizationIEEE Journal on Selected Areas in Communications200523479880810.1109/JSAC.2005.8435462-s2.0-17144419228
15.
LopesC. G.SayedA. H.Incremental adaptive strategies over distributed networksIEEE Transactions on Signal Processing20075584064407710.1109/tsp.2007.896034MR24644162-s2.0-34547870144
16.
RastegarniaA.TinatiM. A.KhaliliA.Performance analysis of quantized incremental LMS algorithm for distributed adaptive estimationSignal Processing20109082621262710.1016/j.sigpro.2010.02.019ZBL1194.941312-s2.0-77951208272
17.
LopesC. G.SayedA. H.Randomized incremental protocols over adaptive networksProceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP '10)March 2010Dallas, Tex, USA3514351710.1109/icassp.2010.54959512-s2.0-78049374572
18.
CattivelliF. S.SayedA. H.Analysis of spatial and incremental LMS processing for distributed estimationIEEE Transactions on Signal Processing20115941465148010.1109/TSP.2010.21003862-s2.0-79952634926
19.
KarS.MouraJ. M.RamananK.Distributed parameter estimation in sensor networks: nonlinear observation models and imperfect communicationIEEE Transactions on Information Theory20125863575360510.1109/tit.2012.2191450MR29243862-s2.0-84861357379
20.
NedićA.OlshevskyA.Distributed optimization over time-varying directed graphsIEEE Transactions on Automatic Control201560360161510.1109/tac.2014.2364096MR33183902-s2.0-84923608183
21.
LopesC. G.SayedA. H.Diffusion least-mean squares over adaptive networks: formulation and performance analysisIEEE Transactions on Signal Processing2008567, part 23122313610.1109/TSP.2008.917383MR25167082-s2.0-46649108330
22.
CattivelliF. S.LopesC. G.SayedA. H.Diffusion recursive least-squares for distributed estimation over adaptive networksIEEE Transactions on Signal Processing20085651865187710.1109/tsp.2007.913164MR24504182-s2.0-46649115239
23.
CattivelliF. S.SayedA. H.Diffusion LMS strategies for distributed estimationIEEE Transactions on Signal Processing20105831035104810.1109/tsp.2009.2033729MR27514192-s2.0-77955401339
24.
ChenJ.SayedA. H.Diffusion adaptation strategies for distributed optimization and learning over networksIEEE Transactions on Signal Processing20126084289430510.1109/TSP.2012.2198470MR29604962-s2.0-84862605412
25.
ChenJ.SayedA. H.On the learning behavior of adaptive networks—part I: transient analysisIEEE Transactions on Information Theory20156163487351710.1109/tit.2015.2427360MR33525112-s2.0-84929195637
26.
ChenJ.RichardC.SayedA. H.Diffusion LMS over multitask networksIEEE Transactions on Signal Processing201563112733274810.1109/tsp.2015.2412918MR33455832-s2.0-84928712188
27.
ZhaoX.TuS.-Y.SayedA. H.Diffusion adaptation over networks under imperfect information exchange and non-stationary dataIEEE Transactions on Signal Processing20126073460347510.1109/TSP.2012.2192928MR29432502-s2.0-84862613480
28.
AbdoleeR.ChampagneB.Diffusion LMS strategies in sensor networks with noisy input dataIEEE/ACM Transactions on Networking201624131410.1109/tnet.2014.23500132-s2.0-84907705171
29.
TuS.-Y.SayedA. H.Diffusion strategies outperform consensus strategies for distributed estimation over adaptive networksIEEE Transactions on Signal Processing201260126217623410.1109/TSP.2012.2217338MR30064142-s2.0-84866491634
30.
MegalooikonomouV.YeshaY.Quantizer design for distributed estimation with communication constraints and unknown observation statisticsIEEE Transactions on Communications200048218118410.1109/26.8235482-s2.0-0033888581
31.
LiJ.AlRegibG.Distributed estimation in energy-constrained wireless sensor networksIEEE Transactions on Signal Processing200957103746375810.1109/TSP.2009.2022874MR26831312-s2.0-70349640575
32.
AysalT. C.CoatesM. J.RabbatM. G.Distributed average consensus with dithered quantizationIEEE Transactions on Signal Processing200856104905491810.1109/tsp.2008.927071MR25172212-s2.0-53149154509
33.
KarS.MouraJ. M.Distributed consensus algorithms in sensor networks: quantized data and random link failuresIEEE Transactions on Signal Processing2010583, part 11383140010.1109/tsp.2009.2036046MR27302112-s2.0-80052335250
34.
LiD.KarS.AlsaadiF. E.DobaieA. M.CuiS.Distributed Kalman filtering with quantized sensing stateIEEE Transactions on Signal Processing201563195180519310.1109/TSP.2015.2450200MR33928892-s2.0-84941120723
35.
LopesC. G.SayedA. H.Diffusion adaptive networks with changing topologiesProceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP '08)April 2008Las Vegas, Nev, USA3285328810.1109/icassp.2008.45183522-s2.0-51449108885
36.
SayedA. H.Fundamentals of Adaptive Filtering2003New York, NY, USAWiley
37.
ScotchmanL.Dither signals and their effect on quantization noiseIEEE Transactions on Communication Technology196412416216510.1109/tcom.1964.10889732-s2.0-78650273796
38.
GrayR. M.StockhamT. G.Jr.Dithered quantizersIEEE Transactions on Information Theory199339380581210.1109/18.256489ZBL0784.940102-s2.0-0027601946
39.
SripadA. B.SnyderD. L.A necessary and sufficient condition for quantization errors to be uniform and whiteIEEE Transactions on Acoustics, Speech, and Signal Processing197725544244810.1109/tassp.1977.11629772-s2.0-0017542211
40.
SayedA. H.Adaptive Filters2008New York, NY, USAJohn Wiley & Sons