Professional Documents
Culture Documents
Research Article
Block Least Mean Squares Algorithm over Distributed Wireless
Sensor Network
T. Panigrahi,1 P. M. Pradhan,2 G. Panda,2 and B. Mulgrew3
1 Department
1. Introduction
A wireless sensor network (WSN) consists of a group
of sensors nodes which perform distributed sensing by
coordinating themselves through wireless links. Since the
nodes operate in a WSN function with limited battery power,
it is important to design the networks with a minimum of
communication among the nodes to estimate the required
parameter vector [1, 2]. In the literature, a number of
research papers have appeared which address the energy
issues of sensor networks. According to the energy estimation
scheme based on the 4th power loss model with Rayleigh
fading [3], the transmission of 1 kb of data over a distance
of 100 m, operating at 1 GHz using BPSK modulation
with 106 bit-error rate, requires 3 J of energy. The same
energy can be used for executing 300 M instructions in a
100 MIPS/watt general purpose processor. Therefore, it is of
great importance to minimize the communication among
nodes by maximizing local estimation in each sensor node.
Each node in a WSN collects noisy observations related
to certain desired parameters. In the centralized solution,
2
ratio (SNR) over the network, recently an ecient adaptive
combination strategy has been proposed [8]. Also a fully
distributed and adaptive implementation to make individual
decisions by each node in the network is dealt with in [9].
Since in block filtering technique [10], the filter coecients are adjusted once for each new block of data in contrast
to once for each new input sample in the least mean square
(LMS) algorithm, the block adaptive filter permits faster
implementation while maintaining equivalent performance
as that of widely used LMS adaptive filter. Therefore, the
block LMS algorithms could be used at each node in order
to reduce the amount of communications.
With this in mind, we present a block formulation of
the existing cooperative algorithm [4, 11] based on the distributed protocols. Distinctively, in this paper, the adaptive
mechanism is proposed in which the nodes of the same
neighborhood communicate with each other after processing
a block of data, instead of communicating the estimates
to the neighbors after every sample of input data. As a
result, the average bandwidth for communication among
the neighboring nodes decreases by a factor equal to the
block size of the algorithm. In real-time scenarios, the
nodes in the sensor network follow a particular protocol
for communication [1214], where the communication time
is much more than the processing time. The proposed
block distributed algorithm provides an excellent balance
between the message transmission delay and processing
delay, by increasing the interval between two messages
and by increasing the computational load on each node
in the interval between two successive transmissions. The
main motivation here is to propose communication-ecient
block distributed LMS algorithms (both incremental and
diusion type). We analyze the performance of the proposed
algorithms and compare them with existing distributed LMS
algorithms.
The reminder of the paper is organized as follows. In
Section 2, we present the BDLMS algorithm and its network
global model. The performance analysis of BDLMS and its
learning characteristics obtained from a simulation study
are presented in Section 3. Performance analysis of the
BILMS and its simulation results are presented in Section 4.
The performance of the proposed algorithms in terms of
communication cost and latency is compared with the
conventional distributed adaptive algorithms in Section 5.
Finally, Section 6 discusses the conclusions of the paper.
(1)
(2)
i = 0, 1, 2, . . . , L 1, j = 1, 2, . . .,
(3)
T
Xk = uk jL , uk jL + 1 , . . . , uk jL + L 1
(4)
T
dk = dk jL , dk jL + 1 , . . . , dk jL + L 1
(5)
k,
ek = dk j Xk w
(6)
Xbg = col X1 , X2 , . . . , XN ,
j
(7)
dbg = col d1 , d2 , . . . , dN .
The objective is to estimate the M 1 vector w from the above
quantities, those collected the data across N nodes. By using
this global data, the block error vector for the whole network
is
j
ebg = dbg Xbg w.
(8)
.
min E dbg Xbg w
(9)
w
1
T
T
w
E dbg dbg E dTbg Xbg w
T E Xbg
dbg
L
T
.
Xbg w
w
T E Xbg
(10)
Let the input regression data u be Gaussian and defined
by the correlation function r(l) = 2 |l| in the covariance
matrix, where is the correlation index and 2 is the variance
of the input regression data, then the relation between
correlation and cross-correlation quantities among blocked
and unblocked data can be denoted as [10]
bg
RX = LRU ,
bg
bg
RdX = LRdu ,
bg
Rd = LRd ,
bg
(11)
bg
T
T
Xbg ], RdX = E[Xbg
dbg], and Rd = E[dTbg dbg ],
where RX = E[Xbg
which are the autocorrelation and cross-correlation matrices
for global data in blocked form. Similarly, the correlation
g
matrices for unblocked data are defined as RU = E[UTg Ug ],
g
g
Rdu = E[UTg dg ], and Rd = E[dTg dg ] where the global
distribution of data across the network is represented as Ug =
[u1 , u2 , . . . , uN ]T and dg = [d1 , d2 , . . . , dN ]T . These relations
are also valid for node data in individual nodes.
Now, the block mean square error (BMSE) in (10) is
reduced to
BMSE =
=
1
g
g
g
g
LRd LRdu w
Lw
T Rdu LwT Ru w
L
g
Rd
g
Rdu w
w
g
Rdu
g
Ru w
w
w
0 = w
j 1 ,
j
w
k = w
k1 +
1
k L
uk jL + q
L q=0
j
dk jL + q uTk jL + q w
k1
j
=w
k1 +
k j T
j
j j
k1 ,
X k dk Xk w
L
for k = 1, 2, . . . N,
w
j =w
N,
(14)
(12)
= MSE.
(13)
j 1
lNk, j 1
j
j 1
w
k = k
ckl w
k ,
k (1) = 0,
1
k L
uk jL + q
L q=0
j 1
dk jL + q uTk jL + q k
.
(15)
j 1
w
k = k
k j T
j
j j 1
Xk dk Xk k .
L
j 2
= E w
k1
k = E ea,k
(MSD),
(17)
(EMSE),
2
k = E ek ( j) = k + ,k
j
(18)
k1 .
ea,k = uk w
The algorithm described in (15) is looking like the interconnection of block adaptive filters instead of conventional LMS
adaptive algorithm among all the nodes across the network.
As shown in (12) that the block LMS algorithm has similar
properties to those of the conventional LMS algorithm, the
convergence analysis of the proposed block diusion LMS
algorithm can be carried out similar to the diusion LMS
algorithm described in [18, 21].
The estimated weight vector for jth block across the
network is defined as
1; . . . ; w
w
j = w
N .
(19)
(20)
X j = diag X1 , . . . , XN .
(22)
dbg = Xbg wg + v j ,
(23)
v j = v1 ; , . . . , ; vN (LN 1)
(24)
(21)
(25)
where the step sizes for all the nodes are embedded in a
matrix S,
and the local error signals such as weight error vector and a
priori error at kth node for jth block are given as
j
(MSE),
w
k1 = w w
k1 ,
dk = Xk w + vk ,
(16)
j
k
(26)
(27)
1
j = INM SRX GE w
j 1 ,
E w
L
(28)
(29)
(30)
1
jT j
T
E v X SSX j v j ,
2
L
T
1
GT E X j X j SG
L
(32)
T
1 T
jT j
G E X X SSE X j X j G.
2
L
jT
T
= GT G GT SE U U j G GT E U j
jT
jT
2
1
6
8
20
20
jT
40
(31)
T
1
= GT G GT SE X j X j G
L
/2 unit
j = E w
j 1 +
E w
60
U j SG
40
7
60
60
40
20
+ GT E U U j SSE U U j G,
(33)
which is similar to (45) in [21]. Using the properties of
expectation and trace [18], the second term of (31) is solved
as
T
1
j
1 jT j
jT j
jT
Ev
X
SSX
v
=
E
tr
X
SSX
E vj vj
L2
L2
= E n j U j SSU j n j ,
(34)
where the noise variance vector n j is not in block form, and
it is assumed that the noise is stationary Gaussian. Equations
(31) and (32) may therefore be written as
j = E w
j 1 +
E w
(35)
T
= GT G GT SE U j U j G GT E U j U j SG
20
40
60
block LMS and are compared with that of DLMS. The row
regressors with shift invariance input [18] are used with each
regressor having data as
uk (i) = [uk (i), uk (i 1), . . . , uk (i M + 1)]T .
(37)
uk (1)
Xk (1) =
uk (2) uk (1)
0
,
1
T
T
En j U j SSU j n j ,
2
L
0
/2 unit
(38)
Xk (2) =
uk (5) uk (4) uk (3).
+ GT E U j U j SSE U j U j G.
(36)
It may be noted that variance estimate (36) for BDLMS
algorithm is exactly the same as that of DLMS [21]. In the
block LMS algorithm, the local step size is chosen to be L
times that of the local step size of diusion LMS in order
to have the same level of performance. As the proposed
algorithm and the diusion LMS algorithm have similar
properties, the evolution of their variances is also similar.
Therefore, the recursion equation of the global variances for
BDLMS will be similar to (73) and (74) in [21]. Similarly,
the local node performances will be similar to (89) and (91)
of [21].
3.3. Learning Behavior of BDLMS Algorithm. The learning
behavior of BDLMS algorithm is examined using simulations. The characteristic or variance curves are plotted for
[1, 1, . . . , 1]T / M.
The input sequence {uk (i)} is assumed to be spatially
correlated and is generated as
uk (i) = ak uk (i 1) + bk nk (i),
i > .
(39)
0.9
0.9
0.8
0.8
0.7
0.7
2
Power profile-u,k
Correlation index-k
0.6
0.5
0.4
0.3
0.6
0.5
0.4
0.3
0.2
0.2
0.1
0.1
Node k
Node k
(a)
(b)
10
10
Figure 2: Network statistics used for the simulation of BDLMS. (a) Network corelation index per node. (b) Regressor power profile.
15
20
25
30
35
15
20
25
30
35
40
40
45
45
50
1000
2000
3000
4000
5000
50
1000
Block number i
Diffusion LMS
BDLMS
2000
3000
Block number i
4000
5000
Diffusion LMS
BDLMS
7
in detail in literature. It is observed that the theoretical
performance of block incremental LMS and conventional
incremental LMS algorithms are similar because both have
the same variance expressions. Simulation results provide the
validation of this analysis.
5
10
15
20
25
30
35
1000
2000
3000
4000
5000
Block number i
Diffusion LMS
BDLMS
j
Xk
j 2
e p,k
2 .
j
Xk
(40)
2
k 2
j T j
j
j 2
E V X X j T V j
k = E w
k1 +
E w
k
k
k
k
L
k
T j
T j
=
X j k Xk + X j k Xk
L
2
k j T j
X X X j T X j .
k k
L k k
+
(41)
4.1. Simulation Results of BILMS Algorithm. For the simulation study of IBLMS, we have used the regressors with
shift-invariance as with the same desired data used in the
case of BDLMS algorithm. The time-correlated sequences are
generated at every node according to the network statistics.
The same network has been chosen here for simulation
study as defined for block diusion network in Section 3.3.
In incremental way of cooperation, each node receives
information from its previous node, updates it by using own
data, and sends the updated estimate to the next node. The
ring topology used here is shown in Figure 8. We assume
that the background noise to be temporarily and spatially
uncorrelated additive white Gaussian noise with variance
103 . The learning curves are obtained by averaging the
performance of 100 independent experiments, generated by
5,000 samples in the network. It can be observed from figures
that the steady-state performances at dierent nodes of the
network achieved by BILMS matche very closely with that of
ILMS algorithm. The EMSE plots which are more sensitive to
local statistics are depicted in Figures 9(a) and 9(b). A good
match between BILMS and ILMS is observed from these
plots. In [19], the authors have already proved the theoretical
matching of steady-state nodal performance with simulation
results. As the MSE roughly reflects the noise power and the
plot indicates the good performance of the adaptive network,
it may be inferred that the adaptive node performs well in the
steady state.
The global MSD curve shown in Figure 10 is obtained
( j 1) 2
k
across all the nodes and over 50
by averaging E
experiments. Similarly, the global EMSE and MSE plots
are displayed in Figures 11 and 12, respectively. These are
( j 1)
k
obtained by averaging Eea,k ( j)2 , where ea,k ( j) = xk, j
across all the nodes over 50 experiments.
If the weights are updated after L data points and then
communicated for local diusion, the number of communications between neighbors is reduced by L times that of ILMS
where the weights are updated after processing each sample
of data. Therefore, similar to BDLMS, the communication
overhead in BILMS also gets reduced by L times that of ILMS
algorithm.
The performance comparison between two proposed
algorithms BDLMS and BILMS for the same network is
shown in Figures 1315. One can observe from Figure 13
that the MSE for BILMS algorithm is converging faster
than BDLMS. Since the same noise model is used for both
the algorithms, therefore after convergence, the steady-state
performances are the same for both of them. But in case of
MSD and EMSE performances in Figures 14 and 15, little
dierence is observed. It is due the dierent cooperation
scheme used for dierent algorithms. However, the diusion
cooperation scheme is more adaptive to the environmental
change compared to the incremental cooperation. But
5
10
15
10
20
25
30
35
20
25
30
35
40
40
45
50
15
1000
2000
3000
4000
5000
45
1000
2000
3000
Block number i
Block number i
Diffusion LMS
BDLMS
4000
5000
Diffusion LMS
BDLMS
(a)
(b)
Figure 6: Local mean-square deviation (MSD) comparison between block diusion LMS and diusion LMS. (a) MSD curve at node 1. (b)
MSD curve at node 7.
60
10
4
40
10
20
20
1
/2 unit
30
40
20
6
8
50
1000
2000
3000
4000
5000
40
Block number i
Diffusion LMS
BDLMS
7
60
60
40
20
0
/2 unit
20
40
60
5. Performance Comparison
In this section, we present an analysis of communication
cost and latency to have a theoretical comparison of the
performances of distributed LMS with block distributed
LMS.
29.3
38
29.4
40
29.5
29.6
29.7
29.8
29.9
30
42
44
46
48
50
52
54
Node, k
Node, k
ILMS
BILMS
ILMS
BILMS
(a)
(b)
20
22
24
26
28
30
32
34
36
38
40
Node, k
ILMS
BILMS
(c)
Figure 9: Network nodal performance. (a) Mean-square error (MSE) versus node. (b) excess mean-square error (EMSE) versus node. (c)
Mean-square deviation (MSD) versus node.
CILMS = (N 1)K.
(42)
CBILMS =
(N 1)K
.
L
(43)
10
10
15
20
25
30
15
20
25
30
35
40
10
35
1000
2000
3000
Iteration i
4000
5000
1000
2000
3000
Iteration i
4000
5000
ILMS
BILMS
ILMS
BILMS
Figure 10: Global mean-square deviation (MSD) curve for incremental LMS and block incremental LMS.
Figure 12: Global mean-square deviation (MSE) curve for incremental LMS and block incremental LMS.
0
5
10
15
20
25
30
35
10
15
20
25
30
40
45
1000
2000
3000
Iteration i
4000
5000
ILMS
BILMS
35
1000
2000
3000
Block number i
4000
5000
BDLMS
BILMS
ni ,
(44)
i=1
(45)
Figure 13: Global mean-square error curve for BILMS and BDLMS.
cK
.
L
(46)
11
The communication delay is mostly due to the implementation of protocols for transmission and reception, which
remains almost the same for dierent nodes. The location
of nodes will not have any major contribution to the delay
unless the destination node is far apart, and a relay node is
required to make the message reach the destination. In this
backdrop, we can assume that the same average delay h is
required to transfer each message for all receiver-transmitter
pairs in the network.
0
5
10
15
20
25
30
35
5.2.1. Estimation of Delays for the ILMS and BILMS Algorithms. In case of ILMS, the duration of each updating cycle
by all the nodes is
40
45
50
1000
2000
3000
Block number i
4000
5000
BDLMS
BILMS
ND + (N 1)h,
(48)
(49)
0
5
10
15
20
25
30
LBILMS =
35
40
(50)
45
50
[NMD + (N 1)h]K
.
L
1000
2000
3000
Block number i
4000
5000
LBILMS = ND +
BDLMS
BILMS
Figure 15: Global excess mean-square error curve for BILMS and
BDLMS.
(N 1)h
K.
L
(51)
Those are processing delay to perform the necessary computations in a node to obtain the estimates to be updated
and the communication delay involved in transferring the
message to the receiver node(s). The processing delay will
very much depend on the hardware architecture of the nodes
to perform the computation which could be widely varying.
But, without losing much of the generality of analysis, we can
assume that each node has M parallel multipliers and one
full adder to implement the LMS algorithm. Let TM and TA
be the time required for executing a multiplication and an
addition, respectively. Therefore, the processing delay needed
for single update in LMS is
5.2.2. Estimation of Delays for the DLMS and BDLMS Algorithms. Similar to ILMS, it is also assumed here that the
updates of a node reaches all the connected nodes after the
same average delay h. Therefore, the communication delay
remains the same as that of ILMS, but in this case, it needs
more processing delay to process the unbiased estimates
received from the connected neighboring nodes. The total
communication delay in a cycle in this case can be given
by cTA + NTM , where c is the total number of messages
transferred in a cycle given by (44). Now, the total duration of
a cycle in diusion LMS with the same hardware constraints
is given by
D = 2TM + (M + 1)TA .
(47)
(52)
12
ILMS
Parameter
Communication
(N 1)K
Cost
Duration of
[ND + (N 1)h]K
convergence
BILMS
DLMS
BDLMS
(N 1)K/L
cK
cK/L
[NMD + (N 1)h]K/L
ILMS
BILMS
DLMS
BDLMS
950
95
4500
450
9.5 s
0.95 s
9.5 s
0.95 s
(53)
6. Conclusion
We have proposed the block implementation of the distributed LMS algorithms for WSN. The theoretical analysis
and the corresponding simulation results demonstrate that
the performance of the block-distributed LMS algorithms
is similar to that of the sequential-distributed LMS. The
remarkable achievement of the proposed algorithms is that
a node requires L (block size) times of less communications
compared to the conventional sequential-distributed LMS
algorithms. This would be of great advantage in reducing
the communication bandwidth and power consumption
involved in the transmission and reception of messages
across the resource-constrained nodes in a WSN. In the
coming years, with continuing advances in microelectronics,
we can accommodate enough computing resources in the
nodes to reduce the processing delays in the nodes, but
the communication bandwidth and communication delay
could be the major operational bottlenecks in the WSNs. The
proposed block formulation therefore would have further
References
[1] D. Estrin, L. Girod, G. Pottie, and M. Srivastava, Instrumenting the world with wireless sensor networks, in Proceedings
of the IEEE Interntional Conference on Acoustics, Speech, and
Signal Processing (ICASP01), pp. 20332036, May 2001.
[2] D. Estin, R. Govindan, J. Heidemann, and S. Kumar, Next
century chanlanges: scalable cordination in sensor network,
in Proceedings of the International Conference on Mobile
Computing and Networks (MobiCOM 99), pp. 263270, 1999.
[3] G. J. Pottie and W. J. Kaiser, Wireless integrated network
sensors, Communications of the ACM, vol. 43, no. 5, pp. 51
53, 2000.
[4] M. G. Rabbat and R. D. Nowak, Decentralized source
localization and tracking [wireless sensor networks], in
Proceedings of the IEEE International Conference on Acoustics,
Speech and Signal Processing, (ICASSP04), vol. 3, pp. III921
III924, 2004.
[5] A. Nedic and D. P. Bertsekas, Incremental subgradient
methods for nondierentiable optimization, SIAM Journal on
Optimization, vol. 12, no. 1, pp. 109138, 2001.
[6] F. S. Cattivelli and A. H. Sayed, Diusion LMS strategies for
distributed estimation, IEEE Transactions on Signal Processing, vol. 58, no. 3, pp. 10351048, 2010.
[7] C. G. Lopes and A. H. Sayed, Diusion adaptive networks
with changing topologies, in Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing,
(ICASSP08), pp. 32853288, April 2008.
[8] N. Takahashi, I. Yamada, and A. H. Sayed, Diusion adaptive networks with changing topologies, in Proceedings of
the International Conference on Acoustics, Speech and Signal
Processing, (ICASSP 06), pp. 28452849, 2009.
[9] F. S. Cattivelli and A. H. Sayed, Distributed detection
over adaptive networks using diusion adaptation, IEEE
Transactions on Signal Processing, vol. 59, no. 5, pp. 19171932,
2011.
[10] G. A. Clark, S. K. Mitra, and S. R. Parker, Block implementation of adaptive digital filters, IEEE Transactions on Signal
Processing, vol. 29, no. 3, pp. 744752, 1981.
[11] F. S. Cattivelli and A. H. Sayed, Analysis of spatial and
incremental LMS processing for distributed estimation, IEEE
Transactions on Signal Processing, vol. 59, no. 4, pp. 14651480,
2011.
[12] X. Liu, Q. Huang, and Y. Zhang, Combs, needles, haystacks:
balancing push and pull for discovery in large-scale sensor networks, in Proceedings of the Second International Conference
on Embedded Networked Sensor Systems, (SenSys04), pp. 122
133, November 2004.
13