Novel Light Weight Compressed Data Aggregation Using Sparse Measurements for IoT Networks

Optimal data aggregation aimed at maximizing IoT network lifetime by minimizing constrained on-board resource utilization continues to be a challenging task. The existing data aggregation methods have proven that compressed sensing is promising for d…

Authors: Amarlingam M, Pradeep Kumar Mishra, P Rajalakshmi

Novel Light Weight Compressed Data Aggregation Using Sparse Measurements   for IoT Networks
1 No v el Light W eight Compressed Data Aggre gation Using Sparse Measurements for IoT Networks Amarlingam M, Pradeep Kumar Mishra, P Rajalakshmi, Sumohana S. Channappayya, and C. S. Sastry Abstract —Optimal data aggregation aimed at maximizing IoT network lifetime by minimizing constrained on-board resource utilization continues to be a challenging task. The existing data aggregation methods have prov en that compressed sensing is promising f or data aggregation. Howe ver , they compromise either on ener gy efficiency or recov ery fidelity and require complex on-node computations. In this paper , we propose a novel Light W eight Compressed Data Aggregation (L WCD A) algorithm that randomly divides the entire network into non-overlapping clus- ters f or data aggregation. The random non-overlapping clustering offers two important advantages: 1) energy efficiency , as each node has to send its measurement only to its cluster head, 2) highly sparse measurement matrix, which leads to a practically implementable framework with low complexity . W e analyze the properties of our measurement matrix using restricted isometry property , the associated coher ence and phase transition. Thr ough extensive simulations on practical data, we show that the mea- surement matrix can reconstruct data with high fidelity . Further , we demonstrate that the L WCD A algorithm reduces transmission cost significantly against baseline approaches, implying thereby the enhancement of the network lifetime. Index T erms —Compressed sensing, data aggregation, Internet of Things, network lifetime. I . I N T R O D U C T I O N The sensor nodes used in Internet Of Things (IoT) appli- cation deployments such as remote sensing and monitoring are typically inexpensi ve, untethered and are powered through batteries [1]. Howe ver , relaying on battery power limits the lifetime of the nodes. Further , regular recharging or replace- ment of batteries leads to additional cost and is a laborious task [2]. Thus, the network lifetime is a critical concern for data aggregation in IoT networks. W ireless transmission consumes significant amount of energy during the data aggregation [3]. Indeed, reducing the number of packet transmissions and minimizing routing path for data aggregation in the network can improve the network lifetime. Sev eral approaches hav e been proposed to address this problem [4]. Compressed Sensing (CS) [5] is a signal processing tech- nique that has prov en to be very promising for data aggregation [6]. CS provides a new perspecti ve for data aggreg ation in IoT networks enabling the compression and route mini- mization jointly for energy efficienc y over the network [7]- [10]. Most of the CS aided data aggregation techniques use Amarlingam M, Pradeep Kumar Mishra, P Rajalakshmi and Sumo- hana S. Channappayya are with the Department of Electrical Engi- neering and C. S. Sastry is with the Department of Mathematics, Indian Institute of T echnology Hyderabad, Hyderabad, India (e-mail: ee13p1003@iith.ac.in; ee16mtech11039@iith.ac.in; raji@iith.ac.in; sumo- hana@iith.ac.in; csastry@iith.ac.in). either dense [7]-[10] or sparse random measurements [11]- [14]. These methods hav e proposed the encoding by utilizing the structural properties of the measurement matrix. In dense random measurements based data aggregation techniques, it is assumed that the individual columns of the measurement matrix are generated at the respectiv e nodes and compute the corresponding measurement [7]-[10]. The sparse random measurements based data aggregation techniques computes the measurements by collecting the data from the interested nodes for each measurement, while assuming that the sparse measurement matrix is stored at each node [11]-[14]. These approaches [7]-[14] aggregate the measurements from all the nodes by minimizing the routing path to reduce the energy consumption in data aggregation. Most of the existing CS aided data aggregation approaches do not consider the feasibility of hardware implementation [7]- [14]. The bottleneck for hardware implementation of the CS aided data aggregation techniques is in the encoding process at IoT nodes that are sev erely resource constrained. The size of the measurement matrix depends on sparsity of the sensing data and the number of nodes deployed in the network [6]. As IoT nodes are resource constrained devices, for sparse random measurements based data aggregation techniques, storage is- sues can crop up in large-scale network applications. In case of dense random measurements, the dependency of column size on sensing data sparsity poses multiple constraints in real-time implementation for the applications where data to be sensed has low sparsity [15]. In contrast, the measurement matrix content can be combined enroute to the sink instead of generating individual columns or storing the matrix while aggregating the data from the nodes using CS. This class of methods called as routing measurements based data aggre- gation approaches. Some existing methods in the literature [16]-[18] ha ve inv estigated data aggregation using routing measurements. Howe ver , these methods compromise either on recovery fidelity (due to low coherence) [17] or energy efficienc y (due to higher number of transmissions) [16] [18]. Designing a low complexity CS based data aggregation technique that minimizes total energy consumption as well as guarantees the reconstruction is still a challenging problem. T o address this problem, in this article, we propose a data ag- gregation method called “Light W eight Compressed Data Ag- gregation (L WCD A)", which is light-weight (low complexity), energy efficient and provides good recovery fidelity . In contrast to some existing approaches [16]-[18], we utilize clustering for data aggregation which is proven to be promising for energy efficient routing [14] [19]. In addition, the aggregated data from cluster heads is collected using a minimum spanning tree 2 to minimize energy consumption. In the proposed algorithm, each node measures a data sample follo wed by generating a random value from a Bernoulli distribution for computing the measurement. The cluster heads recei ve the measurements from their descendants, process them to compute the final measurement before transporting it to the sink. W e find that the measurement matrix constructed from our algorithm is highly sparse and possesses properties to guarantee the recovery of data such as high incoherence, good recovery region and satisfy the Restricted Isometry Property (RIP) when combined with some popular bases. The contributions of this article are summarized as follows: 1) Low complexity CS aided data aggregation technique that constructs a sparse measurement matrix from the network. 2) Performance ev aluation of the measurement matrix with respect to RIP , coherence and phase transition. 3) Comparativ e analysis of the algorithm in terms of re- construction error and transmission cost using real data sets. 4) A practical implementation using IITH Motes [20] to demonstrate hardware feasibility of the proposed L WCDA algorithm. The paper is organized as follows: Section II explains the basics of compressed sensing in IoT networks. Section III describes the proposed L WCD A data aggregation method. Section IV ev aluates the RIP and coherence of the proposed measurement matrix and presents the phase transition analysis. Simulation results of L WCD A method are described in Section V and Section VI describes the hardware implementation performed. Section VII concludes the paper . I I . C O M P R E S S E D S E N S I N G F O R I O T N E T W O R K S A. Compr essed Sensing: A r eview For a giv en N dimensional signal (hereafter data and signal are used interchangeably) that can be sparsely represented using a basis, CS promises to deliv er a full recovery of the signal with high probability from far fewer samples [21]. Let X = [ x 1 , x 2 , x 3 , . . . , x N ] T ∈ R N be sparsely repre- sented in a basis (e.g., Discrete Cosine Transform (DCT), Discrete Fourier Transform (DFT), Discrete W av elets T rans- form (DWT), etc.) Ψ = [ ψ 1 , ψ 2 , . . . , ψ N ] ∈ R N × N with k large coefficients ( k -sparse), where k  N , i.e., X = Ψ θ , θ = [ ψ T 1 X, ψ T 2 X, . . . , ψ T N X ] ∈ R N and k θ k 0 ≤ k . The CS theory computes the compressed M -dimensional vector , Y = Φ X , (1) where Y ∈ R M is the measurement vector and M is the number of measurements and M < N which influences re- construction of the signal. It has been shown that the number of random measurements required for successful reconstruction of a k -sparse signal is M = O ( k l og N ) [22]. The matrix Φ = [ ϕ T 1 , ϕ T 2 , . . . , ϕ T M ] T ∈ R M × N is called the measurement matrix. The problem here is to reconstruct X from Y , which is under-determined and can hav e infinitely many solutions. CS theory shows that the problem of recovering X from its linear measurements can be posed as a l 0 -minimization problem as shown in (2) and it is computationally intractable. A family of greedy algorithms hav e been proposed in [23] and [24] to solve the l 0 -minimization problem. min θ k θ k 0 subject to ΦΨ θ = Y . (2) The most prev alent decoding technique to solve the problem in (2) is l 1 -minimization, which is a conv ex optimization problem [25] and hence, computationally tractable [26], min θ k θ k 1 subject to ΦΨ θ = Y . (3) From the solution θ obtained using l 0 or l 1 -minimization, X can be reconstructed as, b X = Ψ θ . (4) The CS matrix A = ΦΨ plays a crucial role in the recovery of the N dimensional original signal X . In [27], it is shown that the CS matrix A should satisfy the property known as RIP for successful recovery of X using l 1 minimization. A matrix A ∈ R M × N is said to satisfy the RIP of order k with constant δ k ∈ (0 , 1) if (1 − δ k ) k u k 2 2 ≤ k Au k 2 2 ≤ (1 + δ k ) k u k 2 2 , ∀ u ∈ Σ k , (5) where u is a k -sparse vector and Σ k is set of all k -sparse vectors. On the other hand, if X can be sparsely represented in Ψ domain, then to achiev e successful recov ery , the theory of CS requires low mutual coherence between the columns of the CS matrix A = ΦΨ . The mutual coherence of the CS matrix can be defined as µ ( A ) = max 1 ≤ p 6 = q ≤ N |h a p , a q i| , (6) where a p and a q are normalized columns of A . B. Related W orks In this section we discuss the contributions of the rele vant literature. Most of the CS aided data aggregation techniques can be classified into three classes, dense random measure- ments [7]-[10], sparse random measurements [11]-[14] and routing measurements [16]-[18] based data aggregation meth- ods. Dense random measurements based methods [7]-[10] achiev e CS aided data aggregation by considering individual column generation of the measurement matrix at node lev el using pseudo-random sequences. These methods aggregate the measurements from all the nodes by minimizing routing path to achieve energy efficiency . The size of the measurement matrix depends on the number of nodes and sparsity of the data. IoT nodes are constrained devices possessing minimal on-board resources (in terms of physical memory , processing capability , internal memory , energy). Therefore, generating individual columns of the measurement matrix at a node in case of a large-scale network application where sensing data sparsity is low is computationally intensiv e and poses multiple constraints in real-time implementation. W ang et al. [11] showed that sparse random measurements (projections) reduce communication cost per sensor node for 3 data aggregation. In [12]-[14], data aggregation techniques hav e been proposed to achie ve energy ef ficiency for IoT networks by using the sparse random measurements [11]. These algorithms find the optimal route to collect data from the interested nodes for each measurement, while assuming that the sparse measurement matrix is stored at each node. Since the measurement matrix depends on the network size, storage issues can crop-up for large-scale networks. In other words, commercially av ailable nodes that hav e minimal on- board resources may not be able to support the storage large measurement matrices. In contrast to dense and sparse random measurements, the routing measurements based data aggregation methods aggregate the measured data from the nodes by computing measurements on the fly enroute to the sink [16]-[18]. In [16], the routing paths are iteratively built through a greedy choice to minimize the coherence of the CS matrix and energy required for data aggregation. Howe ver , building of routing paths in an iterative manner is computationally intensive and requires more transmissions rendering the process highly en- ergy inef ficient. In [17], the algorithm picks up a portion of the nodes randomly from the network to generate measurements by utilizing shortest path routing. Howe ver , such an approach does not achieve good performance with respect to coherence. In [18], the authors showed that data aggregation from fixed length random walks starting at randomly located nodes can reconstruct the data using CS. Howe ver , recovery performance of the method depends on the length of the random walks. An increase in the length of the walk increases the number of transmissions which in turns data aggregation to be energy inefficient. C. Pr oblem Statement As discussed above, CS based data aggregation algorithms proposed in the literature do not address the aspects of low complexity and energy efficienc y jointly . The approaches proposed based on dense random measurements [7]-[10] as well as sparse random measurements [11]-[14] are energy efficient but not real-time implementable. On the other hand, the approaches proposed in [16]-[17] are light weight, howe ver they are either energy inef ficient or do not achiev e good performance in terms of coherence and recov ery . These lim- itations provide the motiv ation for this work. Specifically , the problem is to design a low complexity (real-time) CS aided data aggregation method that is energy efficient and can guarantee a successful recovery of the data for IoT networks. I I I . P R O P O S E D D A TA A G G R E G AT I O N P R OT O C O L In this section we first present the network model that will be used in our analysis and next describe the proposed data ag- gregation protocol which forms the light weight measurement matrix. A. Network Model Consider an IoT network with N nodes deployed in a rect- angular area (an example network with grid-wise deployment of N = 100 nodes is shown in Fig. 1). The network can be represented by a graph G ( V , E ) , where V is the set of vertices or nodes and E represents the set of edges or links between the nodes. The sink node S is the node that collects data from all the other nodes in the network. W e assume that all the nodes are loosely time synchronized and hav e homogeneous transmission coverage. Unit disc coverage model is considered for all the nodes. W e consider the communication range of the nodes to be D = q 5 N ∗ a [17]. Here, a is the length of the maximum side of the considered area and N is the number of nodes. Data aggregation proceeds in c ycles (rounds) and each node generates one sample per cycle. For example, the i th node acquires data sample x i in each cycle and N samples X = [ x 1 , x 2 , x 3 , . . . , x N ] T ∈ R N will be acquired from all the nodes per cycle. W e also assume that there is no packet loss in data aggregation. W e consider both grid [29] [30] and random deployment [31] scenarios for analysis in further sections as these network deployments have their own significance in different application scenarios. B. Pr oposed Data Aggre gation Protocol As described above, X ∈ R N is a signal of length N that contains measurements from N nodes in the network. T o ag- gregate data from all the nodes, M nodes are randomly picked such that each node is a Cluster Head (CH) with a probability P C H = M N . The remaining ( N − M ) leaf nodes connect to their respective nearest CH through the shortest path. Accordingly , the whole network gets di vided into M non- ov erlapping clusters to aggregate sensors data. The M clus- ters { c 1 , c 2 , . . . , c M } can contain distinct { n 1 , n 2 , . . . , n M } number of nodes. Every node in the cluster measures its data sample x i (e.g., temperature, humidity , light intensity , etc.) and multiplies it with a random value α i generated from a Bernoulli distribution with a success probability of 0 . 5 . In other words, the i th node performs α i x i , where α i is randomly drawn from the set {− 1 , 1 } with a Bernoulli distribution and i ∈ [1 , N ] . Each leaf node sends the measurement α i x i to its CH. The CH adds the received measurements from the leaf nodes including its o wn measurement. The final measurement at j th CH, y j = P i ∈ c j α i x i is the linear combination of α i and x i , where the nodes belonging to the cluster take non- zero values i.e., { α i 6 = 0 , x i 6 = 0 } ∈ c j and the nodes that do not belong to the cluster can be assumed to be zeros i.e., { α i = 0 , x i = 0 } / ∈ c j . The CHs deliv er the computed mea- surements to the sink node through the Minimum Spanning T ree (MST). Dijkstra’ s and Kruskal algorithms can be used to create MST of CHs along with the sink node. The CHs follow the pack and forward method [8] that provides the feasibility to encapsulate the current measurement of a CH with the relaying packet from descendant CHs along the MST towards the sink. From the CS formalism in Section II, each cluster can be considered as a ro w of the measurement matrix Φ and each node in the network corresponds to a column of Φ . In other words, M randomly formed clusters and the nodes in each cluster correspond to rows and respectiv e columns of Φ . The j th cluster c j forms the j th row of Φ , i.e., ϕ j . The support vector of ϕ j is ∆ j = { i : i ∈ [1 , N ] , i ∈ c j } , ϕ j ∆ j = { α i : 4 i ∈ ∆ j } and ϕ j ∆ c j = 0 . In other words, the j th row of Φ at respectiv e columns of nodes that are connected as a cluster i ∈ c j will be assigned values from the set {− 1 , +1 } with a Bernoulli distribution. The remaining entries in the row will be zeros. More concretely , Φ ∈ R M × N , Φ = [ ϕ T 1 , ϕ T 2 , . . . , ϕ T M ] T contains elements in each row ϕ j i = ( − 1 or + 1 if i ∈ c j 0 otherwise . Packets received at the sink node from the MST contain elements of the measurement vector Y = [ y 1 , y 2 , . . . , y M ] T ∈ R M which are linear combinations of the measured data and the random values of nodes, i.e., Y =      y 1 y 2 . . . y M      =      ϕ 1 ϕ 2 . . . ϕ M       X  = Φ X, (7) where X = [ x 1 , x 2 , . . . , x N ] T , X ∈ R N , ϕ m ∈ R N , y m ∈ R where m ∈ [1 , M ] . T o gain insight into the described L WCD A, we consider a network of grid-wise deployed 100 nodes with a sink node (S = 101), which is placed at the center of the network as shown in Fig. 1. Fig. 1 shows the measurement matrix formation from the network and the sink node. Consider the example node of 98 from Fig. 1, which is a CH and has two descendant nodes 88 , 97 . The measurement matrix Φ contains a ro w which represents the cluster with the nodes 98 , 88 , 97 and contains non-zero values from the set {− 1 , +1 } that are drawn from a Bernoulli distribution with a success probability of 0 . 5 at respectiv e columns, as shown in Fig. 1. T o recover the original signal X ∈ R N from the mea- surement vector Y ∈ R M , the sink node needs to have the knowledge of Φ . The information of Φ can be shared with the sink by maintaining synchronized seeds and pseudo- random number generators between the nodes and the sink [7]. Practically , to share pseudo-random number seeds, each node has to send its seed to the sink [18] and this requires a large number of transmissions for large-scale networks. Another approach is to transmit information of Φ to the sink along with the measured data if the message overhead is negligible [17]. In our data aggregation algorithm, each node needs to send or share the information of α = − 1 or +1 with the sink, which can take a maximum of one octet of packet payload. W e consider the case that ev ery j th CH sends individual α values of the nodes that belong to that cluster c j and their indices ∆ j along with the measurement P i ∈ c j α i x i to the sink in the first cycle of data collection. The system of linear equations in (7) ( M < N ) is under-determined and will giv e infinitely many solutions while recovering X from Y . The sink node reconstructs full dimension b X ∈ R N from the receiv ed measurement vector Y ∈ R M by solving either of the optimization problems discussed in Section II. 1) Node-level Complexity for Encoding: The node-level complexity of measuring the data is computed in terms of generating or storing the number of random values. The proposed data aggregation algorithm constructs Φ on the fly while data is being aggregated from the nodes. Note that each node is required to generate only a random v alue − 1 or +1 from a Bernoulli distribution as discussed above. The node lev el complexity of our method in terms of generating or storing number of random v alues is Θ(1) which is independent of sensing data sparsity and network size. The Θ( ) refers the formal notation for stating the exact bound on growth of resource needs (computation and storage) of an algorithm. Baseline data aggregation approaches [7]-[10] which utilize dense random measurements require the generation of the respectiv e columns at each node which is the size of Θ( M ) units. In case of sparse random measurements based data aggregation methods [12]-[14], ev ery node in the network stores the complete Φ . The size of the required storage is Θ( M N ) . Some of the methods which use sparse random measurements such as [11] generate the respecti ve row of Φ at ev ery node and the row size is Θ( N ) units. The values of M and N are proportionally related and depend on sensing data sparsity and the network size. This dependency poses multiple constraints on the real-time implementation of the large-scale network applications where the data to be sensed has low sparsity . The proposed approach is lightweight as it completely eliminates the burden of generating a specific column or storing the entire Φ at the node to perform data aggregation in the network. Consequently , the proposed method can be implemented in commercially av ailable low end IoT nodes. The measurement matrix Φ should satisfy certain properties as discussed in Section II for it to allow data recovery . In the following section we ev aluate the properties of the Φ and demonstrate how it can guarantee the reconstruction. I V . M E A S U R E M E N T M A T R I X A NA LY S I S T o analyze the proposed measurement matrix Φ , we rely on RIP , coherence and Phase T ransition (PT) [28] analyses. W e considered both grid and random deployments scenarios as both deployments have their own significance for different application scenario [29]-[31]. W e considered DCT , DFT , D WT , Laplacian and Diffusion W av elet (DiWT) bases ( Ψ ) for the analyses. The DCT , DFT and DWT bases ( Ψ ) can sparsify data from regular (grid-wise) IoT deployments [7], [18]. In case of randomly deployed networks, the Laplacian [18] and Diffusion wav elet (DiWT) [32] can accommodate irregularity and provide a sparse representation of the data. A. Numerical Experiments: RIP Analysis As discussed in Section II, RIP is a standard tool to analyze near-orthonormal performance of a CS matrix while operating with sparse input vectors. This property measures the performance of a compressed sensing matrix in terms of the Restricted Isometry Constant (RIC) δ k . As a result, δ k can be used to e valuate the ability to recov er a sparse signal from the measurement vector . From the definition of RIP of a matrix 5 Fig. 1: The procedure of measurement matrix designing from a network with N = 100 nodes and M = 40 . The sink S = 101 is represented by a star, square boxes represent the CHs and remaining nodes are leaf nodes. A ∈ R M × N , for k -sparse vectors with a constant δ k , (5) can be rewritten as, δ k = max T ⊂ [ N ] , | T |≤ k k A ∗ T A T − I d k 2 → 2 , (8) where I d ∈ R | T |×| T | is an identity matrix and T is the support set of k -sparse vector [21]. For any matrix A that satisfies RIP with a RIC of δ k , the following condition holds: (1 − δ k ) ≤ λ min ( A ∗ T A T ) ≤ λ max ( A ∗ T A T ) ≤ (1 + δ k ) , (9) where λ min and λ max are the minimum and maximum eigen values of the symmetric matrix A ∗ T A T respectiv ely . 1) Numerical verification of δ k , λ min and λ max : The DFT , DCT , D WT , Laplacian and DiWT matrices are separately considered as bases Ψ N × N for the empirical ev aluation of δ k . T o verify with DWT , we ev aluated the performance of the CS matrix A across several popular wa velets such as Daubechies, Symlets, Coiflets and chose the Daubechies-4 wa velet for all our analysis as it gives a better range for k compared to other wavelets. The compressed sensing matrix A M × N = Φ M × N Ψ N × N with N = 1024 , at different com- pression rates with M = 103 , 308 , 717 , 922 are considered for ev aluation. For a compressed sensing matrix A ∈ R M × N , the compression rate Γ% can be written as, Γ% =  1 − M N  × 100 . The procedure follo wed for empirical e valuation of δ k , λ min and λ max is described below: 1) Generate the measurement matrix Φ and the basis Ψ for fixed N = 1024 and for each M . 2) For a combination ( N , M ) , k is varied across [1 : M ] . 3) Consider a k -sparse vector u ∈ R N . The vector u contains non-zero values at k randomly chosen locations and the values themselves are chosen from a normal distribution. 4) Find the support set for u , i.e., T . 5) Repeat steps 2 and 3 for 10000 iterations for each combination ( N , M , k ) and calculate δ k from (8). 6) Compute λ min ( A ∗ T A T ) , λ max ( A ∗ T A T ) , where T is the support set corresponding to δ k from step 5. T ABLE I: Sparsity value k where RIC δ k ∈ (0 , 1) ∀ u ∈ Σ k for different Γ . Sparsity value k Regular deployment Random deployment Compression rate Γ DFT DCT DWT Laplacian DiWT 90% 7 6 1 2 1 70% 15 14 3 4 3 30% 36 35 9 13 9 10% 67 66 11 19 13 The calculated RIC δ k values, λ max , λ min with respect to sparsity value k at different compression rates Γ , are plotted in Fig. 2. In Fig. 2, δ kf , δ kc , δ kw , δ kl , δ kd refer to RICs of CS matrix A where Ψ is DFT , DCT , DWT , Laplacian and DiWT respectively . λ min , λ max refer to the minimum and the maximum eigen values of CS matrix A respectiv ely when Ψ is DFT . Similar behavior of eigen values is also observed with DCT , DWT , Laplacian and DiWT bases. Sparsity values k obtained while δ k ∈ (0 , 1) for the proposed CS matrix A with dif ferent bases are tabulated in T able I (the same can be observed from Fig. 2 as well). The interesting observ ation made from T able I is that the CS matrix A gives better range for k with DFT compared to that of DCT and DWT bases. In the random deployment case, CS matrix A gives slightly better range for k with Laplacian then DiWT basis. B. Coher ence Analysis As discussed in Section II, if X can be sparsely represented in an arbitrary basis Ψ , then for successful recovery , CS theory requires low mutual coherence between columns of the matrix A = ΦΨ . The mutual coherence µ of the matrix A with different bases at v arious compression rates Γ is calculated using (6), i.e., the CS matrix A M × N = Φ M × N Ψ N × N where N = 1000 and M is chosen to vary from 100 to 900 in steps of 100 ( M = 100 : 100 : 900 ) for calculating µ . The resultant mutual coherence with different bases are shown in Fig. 3. The CS matrix A provides better incoherence for the DCT and DFT bases compared to the D WT basis where Φ is constructed from grid deployment. In case of random 6 10 20 30 40 50 60 70 0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 RIC and eigenvalues for M=921 N=1024 1 - kf 1 + kf 1- kc 1+ kc max min 1+ kw 1- kw 1+ kl 1- kl 1+ kd 1- kd (a) Γ = 10% . 5 10 15 20 25 30 35 40 0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 RIC and eigenvalues for M=717 N=1024 1 - kf 1 + kf 1- kc 1+ kc max min 1+ kw 1- kw 1+ kl 1- kl 1+ kd 1- kd (b) Γ = 30% . 2 4 6 8 10 12 14 16 18 0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 RIC and eigenvalues for M=307 N=1024 1 - kf 1 + kf 1- kc 1+ kc max min 1+ kw 1- kw 1+ kl 1- kl 1+ kd 1- kd (c) Γ = 70% . 1 2 3 4 5 6 7 8 9 0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 RIC and eigenvalues for M=103 N=1024 1 - kf 1 + kf 1- kc 1+ kc max min 1+ kw 1- kw 1+ kl 1- kl 1+ kd 1- kd (d) Γ = 90% . Fig. 2: RIC δ k and eigen value bounds ( λ min , λ max ) for the proposed CS matrix A = ΦΨ . Here, Φ is the measurement matrix constructed using L WCDA algorithm and analyzed for different Ψ (DCT , DFT , DWT , Laplacian and DiWT) matrices. (a) - (d) show values of δ k , λ max and λ min for matrix A at different compression rates Γ . CS matrix A gives a better range of k with DFT and DCT compared to DWT basis in grid deployment. In the random deployment case, CS matrix A gives slightly better range of k with Laplacian compared DiWT . 10 20 30 40 50 60 70 80 90 Compression rate Γ % 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Coherence µ DiWT DCT DFT DWT Laplacian Fig. 3: Comparison of mutual coherence µ of the CS matrix A = ΦΨ with different bases where matrix Φ is constructed from L WCDA. Columns of the matrix A are highly incoherent with DFT basis among all. deployment, the coherence of the matrix A with Laplacian is fairly better compared to DiWT basis across all compression rates. It is observed from Fig. 3 that among all the bases, DFT provides high incoherence for all compression rates. C. Phase T ransition Analysis For a given CS matrix, the phase diagram can be generated as a numerical representation of successful recovery probabil- ity P s ov er the space ( k / M , 1 − ( M / N )) ∈ [0 , 1] 2 , as in [28]. This space is discretized and we performed multiple compression and decompression experiments at each grid point. The phase diagram is finally approximated by using successful recovery probability P s = P r { e ≤ e T H } , where the reconstruction error e =    X − ˆ X    2 / k X k 2 , with an appropriately selected threshold e T H . W e considered error threshold e T H = 10 − 8 in our analysis. For PT analysis, A M × N = Φ M × N Ψ N × N is considered with N = 1000 and ev aluated for different compression rates ( Γ ) with M = 100 : 100 : 900 . Fig. 4 sho ws the phase diagram of CS matrix A = ΦΨ , where Φ is the measurement matrix and Ψ is the basis. Fig. 4 also illustrates that the proposed measurement ma- trix Φ with DCT and DFT bases provides promising recov ery region compared to DWT where Φ is constructed from grid deployment. In case of random deployment, Laplacian basis provides slightly better recov ery region compared to DiWT basis. This ev aluation has shown that the proposed measurement 7 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 (a) DCT . 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 (b) DFT . 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 (c) DWT . 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 (d) Laplacian. 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 (e) DiWT . Fig. 4: Phase transition analysis of CS matrix A = ΦΨ for different bases, where Φ is the proposed measurement matrix and Ψ is the basis matrix. The color bar indicates successful recovery probability P s . CS matrix A with DCT and DFT basis yields promising recovery region. matrix Φ giv es better performance with DCT and DFT bases compared to the DWT basis in terms of RIC, coherence and PT analysis where Φ is designed from grid-wise deployed network. Further, in random deployment scenario, Laplacian and DiWT bases giv e comparable performance. The proposed matrix Φ with DCT and DFT bases ( Ψ ) has the ability to recov er the signals successfully even though they hav e fairly low sparsity . Whereas in case of DWT , Laplacian and DiWT , the matrix Φ can recover the signals on the condition that they are highly sparse. T o extend the proposed L WCDA method to fairly low sparse data cases especially in random deployment scenario, we propose a technique called spatial logical node mapping, which is described in the following subsection. D. Spatial Logical Node Mapping Before in voking the L WCD A algorithm, we first model the network as a logical chain based on the Euclidean distance between the nodes. The algorithm starts from any random node and gives sequential node IDs along the chain. The method used to form the logical chain is similar to that in [33]. W e consider that in the initial phase, each node sends the distance information of the nodes that are in its cov erage area to the sink. The sink maps the new node IDs from old node IDs and sends it back to the nodes to change. Fig. 5 provides more insight into the Spatial Logical Node Mapping (SLNM) with an example network of N = 30 nodes. This preprocessing will introduce spatial correlation in the data since adjacent nodes in the chain tend to be the nodes which are geographically close to each other [33] [34]. The spatial correlation among the samples generated from the nodes which are geographically close to each other can make the signal sparse in the regular DFT and DCT bases. SLNM adds the adv antage to L WCD A to guarantee the recovery of the measured data from the random deployment as it introduces sparsity for the data in DFT and DCT bases. In the following section, we ev aluate the performance of the proposed data aggregation method in an application scenario. V . R E S U LT S A N D E V A L U A T I O N In this section, the performance of the proposed data aggre- gation method is analyzed using the following metrics: 1. Reconstruction error . 2. Transmission cost. 0 20 40 60 80 100 0 10 20 30 40 50 60 70 80 90 100 Network of N=30 nodes 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 Sensor node (a) Randomly deployed nodes. 0 20 40 60 80 100 0 10 20 30 40 50 60 70 80 90 100 Network of N=30 nodes 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 Sensor node (b) Logical mapped nodes. Fig. 5: Logical node mapping. 21 22 23 24 25 26 Fig. 6: T emperature sensing field. 0 100 200 300 400 500 600 700 800 900 1000 Node index( N ) 20 22 24 26 28 Measured v alue 0 100 200 300 400 500 600 700 800 900 1000 Node index ( N ) 20 22 24 26 28 Measured v alue Fig. 7: Measured data from random (top) and grid-wise (bottom) deployed 1024-node network. A. Reconstruction Err or Analysis W e extend the application of the proposed algorithm for compressible signals by using a real data set for ev aluation. Real temperature data which is obtained by capturing ther- mal images from the top view of an area 100 m × 100 m is considered for analysis. Fig. 6 visualizes the considered temperature data for recovery performance of the proposed 8 method. Fig. 7 shows the measured data from random and grid-wise deployed 1024 sensor nodes on the field. W e used MA TLAB R2015b software for performing all our simulations. Ideally the sparsity value k of X in a basis Ψ is measured using the l 0 norm, k = k θ k 0 , where θ = Ψ X . For real-time data which is approximately sparse, only few large coef ficients contribute a large proportion of the total energy . W e use numerical sparsity [32] as the measure of sparsity which represents the number of ef fectiv e large coef ficients. If a vector X can be represented using a sparsifying basis Ψ as X = Ψ θ , then the numerical sparsity of X can be calculated as s = k θ k 1 2 k θ k 2 2 . (10) Numerical sparsity of the considered temperature data with different bases ( Ψ ) (DFT , DCT , DiWT and Laplacian) are tabulated in T able II. 10 20 30 40 50 60 70 80 90 Compression rate Γ % 0 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.1 Reconstruction error e DFT DCT DWT (a) 10 20 30 40 50 60 70 80 90 Compression rate Γ % 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 Reconstruction error e DiWT Laplacia DFT DCT (b) Fig. 8: A verage reconstruction error against different compression rates. Here, 1024 temperature data points are considered. Figures (a) and (b) depicts recovery of grid-wise and Randomly deployed nodes measured data respectiv ely . DFT basis provides lo w recov ery error compared to all bases in both grid and random deployment scenario. T able II shows considered data is more sparse in DCT and DFT bases compared to the others in both grid-wise and random deployment scenarios. T o ev aluate recovery perfor- mance of the measured data from grid-wise deployed nodes the basic L WCDA is used to construct the measurement matrix Φ . In case of random deployment, Φ is constructed from L WCDA and SLNM. The OMP algorithm [24] is used for the recovery of the compressed data. W e ev aluated the performance of the proposed data aggregation method in terms of the reconstruction error e against the compression rate Γ . Fig. 8 compares av erage reconstruction error of our method with dif ferent bases. In our analysis, e is a veraged o ver 100 iterations for each Γ . From Fig. 8a, it can be observed that DFT and DCT can recov er the data which is measured from grid-wise deployed nodes with a low error compared to D WT for all compression rates. Fig. 8b illustrates the data recov ery performance of the proposed method where the data is measured from randomly deployed sensor nodes. From Fig. 8b, it can be observed that DCT and DFT can recover the data with a low error across all compression rates compared to other bases. Howe ver , DiWT and Laplacian result in high recov ery error as they require the signal to be highly sparse. This ev aluation has shown that the proposed L WCD A method provides high recovery fidelity using the DFT basis for the T ABLE II: Sparsity measure of the temperature data Numerical sparsity value s Regular deployment Random deployment SLNM DFT DCT DWT Laplacian DiWT DFT DCT 2.2205 2.5251 7.6707 15.925 53.0402 2.6219 2.7569 0 20 40 60 80 100 0 1 2 3 4 5 6 7 8 9 10 4 Non-CS Hybrid CS CWCDA LWCDA-Random SPRM LWCDA-Grid Fig. 9: Comparison of the transmission cost required for data aggregation from 625 nodes deployed in an area 256 m × 256 m using the proposed L WCD A and SPRM, Hybrid CS, CWCD A and Non-CS methods against the compression rate Γ . Transmission cost of the L WCDA is significantly low compared to all methods almost for all compression rates ( Γ ≤ 80% ). data measured from both the random and grid-wise deployed nodes. In the following section, we perform a comparative analysis of the transmission cost of our algorithm with traditional CS based data gathering methods. T o demonstrate the efficienc y of our algorithm, we compare with SPRM for the grid- wise deployment scenario, CWCD A, Hybrid CS and Non-CS methods for the random deployment scenario. B. T ransmission Cost Analysis T ransmission cost of the network G ( V , E ) is defined as [7], T cst = X ( i,j ) ∈ E t ij c ij , (11) where t ij represents the traffic on the link ( i, j ) ∈ E and c ij is the cost of the link. W e considered one packet as one unit of traffic on the link and cost of the link c ij is considered as the Euclidean distance between the nodes i and j . ZigBee protocol is considered for simulations as the ZigBee stack is one of the most commonly used protocols among commercially available off-the-shelf IoT solutions. The size of PHY layer data field of the packet of ZigBee is 128 bytes, of which 87 bytes can be used for application payload as the remaining octets are reserved for packet header information of higher layers. The number of bits required to represent the data sample and the address field (short address mode) is considered to be 2 octets. For transmission cost comparison, a network deployment of 625 nodes in an area of 256 m × 256 m is considered. The com- parison of the transmission cost for data aggregation using the proposed method (L WCD A), SPRM, Hybrid CS, CWCD A and a Non-CS method with respect to the change in compression 9 0 20 40 60 80 100 0 20 40 60 80 100 120 140 160 180 200 Non-CS Hybrid CS CWCDA SPRM Fig. 10: Comparison of the percentage of disbursed transmission cost of the proposed L WCDA with respect to SPRM, Hybrid CS, CWCDA and Non-CS methods against the compression rate Γ . Here, data aggreg ation is considered from 625 nodes deplo yed in an area 256 m × 256 m . Percentage of disbursed transmission cost of the L WCDA with respect to SPRM, Hybrid CS, CWCDA and Non-CS methods is low for almost all compression rates ( Γ ≤ 80% ). rate ( Γ ) is shown in Fig. 9. In particular , we considered the con ventional shortest path algorithm [35] for data gathering as the Non-CS approach, where each node in the network sends its data to the sink through the shortest path. From Fig. 9, it can be observed that our L WCD A (labeled as L WCDA- Random for random deployment) method requires very low transmission cost for data aggregation compared to Non-CS, Hybrid CS and CWCD A for all compression rates where nodes are randomly deployed. In case of grid deployment, compared to SPRM the proposed L WCDA (labeled as L WCDA-Grid for grid deployment) method outperforms until Γ = 80% for data aggregation. In the proposed L WCD A method, an increase in compression rate Γ decreases the number of required clusters for data aggregation. A decrease in number of clusters increases the required transmission cost for data aggregation as the leaf nodes need to send their measurements to CHs from farther distance. Further, as Γ increases, the required transmission cost to collect measurements from CHs (using MST) also decreases. This results in an increase of total transmission cost T cst at higher compression rates ( Γ > 80% ) as sho wn in Fig. 9. Fig. 10 illustrates the percentage of disbursed transmission cost D T cst of the proposed L WCDA with respect to that of Hybrid CS, L WCD A, SPRM and Non- CS methods. Percentage of disbursed transmission cost D T cst of a giv en method P with respect to the method Q is defined as, D T cst % = T cst of method P T cst of method Q × 100 . (12) The SPRM method at high compression rates ( Γ > 80% ) results in lesser transmission cost as compared to that of the proposed L WCD A method. This in turn results in the per- centage of disbursed transmission cost of L WCD A (L WCDA- Grid) to go beyond 100% as sho wn in Fig. 10 for higher compression rates. This is because, in the SPRM method, very few randomly selected nodes are required to send data through the shortest path to the sink at high compression rates. Although, SPRM offers higher compression rates with lower transmission costs, it does not achiev e good performance with 100 225 400 625 900 0 1 2 3 4 5 6 7 8 9 10 4 Non-cs Hybrid CS CWCDA LWCDA-Random SPRM LWCDA-Grid Fig. 11: T ransmission cost comparison of the proposed L WCDA method against node density with SPRM, Hybrid CS, CWCD A and Non-CS methods at Γ = 50% . Here, an area of 256 m × 256 m is considered for the network deployment and number of nodes deplo yed ( N ) are varied. Transmission cost of the L WCDA is significantly lo w compared to SPRM, Hybrid CS, CWCDA and Non-CS methods for all considered node densities. 100 225 400 625 900 0 10 20 30 40 50 60 70 80 90 100 110 Non-CS Hybrid CS CWCDA SPRM Fig. 12: Comparison of the percentage of saved transmission cost of the proposed L WCDA with respect to SPRM, Hybrid CS, CWCDA and Non-CS methods against node density at Γ = 50% . Here, an area of 256 m × 256 m is considered for the network deployment and number of nodes ( N ) deployed are varied. The proposed L WCD A method offers savings in the transmission cost consistently with respect to the baseline approaches for all the considered node densities. respect to coherence leading to higher reconstruction errors at higher compression rates, thereby not guaranteeing a success- ful reconstruction (as discussed in [17]). For all compression rates in both grid-wise ( Γ ≤ 80% ) and random deployment scenario, the proposed L WCDA method can deliv er the data to the sink with a lower transmission cost as illustrated in Fig. 9 and with a lower percentage of disbursed transmission cost as shown in Fig. 10, thereby enhancing the network lifetime as compared to the considered baseline approaches. T o ev aluate the effect of changing the node density on the required transmission cost for data aggregation of the proposed L WCDA, we performed an experiment where the number of nodes deployed is v aried in the considered area of 256 m × 256 m . The transmission cost of data aggregation with respect to the changing in node density with Γ = 50% compression rate is shown in Fig. 11. From Fig. 11, it is observed that the transmission cost increases with an increase in the node density . The interesting observation made from Fig. 11 is that the transmission cost for L WCD A is signifi- 10 0 50 100 150 200 250 0 5 10 15 10 4 Non-CS Hybrid CS CWCDA LWCDA-Random SPRM LWCDA-Grid Fig. 13: Comparison of the transmission cost required for data aggregation from 625 nodes deployed in an area of 256 m × 256 m using the proposed L WCD A and SPRM, Hybrid CS, CWCDA and Non-CS methods against the sink location. The sink node location ( X , Y ) varies according to the line X = Y where X , Y ∈ [0 , 256] . Transmission cost of the proposed L WCD A method is robust and lo wer compared to all baseline approaches for all considered sink locations. cantly lo w as compared to that of the traditional methods for all considered density lev els in both deployment scenarios. Fig. 12 shows the percentage of savings in the transmission cost with respect to Non-CS, Hybrid CS, CWCD A and SPRM methods. Percentage of sav ed transmission cost S T cst of a giv en method P with respect to the method Q is defined as, S T cst % =  1 − T cst of method P T cst of method Q  × 100 . (13) From Fig. 12, one can observe that the proposed method consistently offers sa vings in transmission cost under the considered varying node densities. W e can infer that for large-scale dense networks, L WCD A algorithm can achiev e significant improvements in the network lifetime compared to traditional approaches. The location of the sink node affects the required transmis- sion cost for data aggregation [36]. T o analyze the dependence of the transmission cost on the sink location for data aggre- gation, we considered a 625 -node network deployed (grid and random deployment) in an area of 256 m × 256 m with varying sink locations. Fig. 13 compares the transmission cost of the proposed L WCDA algorithm with that of SPRM (for grid-wise deployment) and Non-CS, Hybrid CS and CWCDA (for ran- dom deployment) with respect to various sink locations. Note that the variables X , Y ∈ [0 , 256] represent the geographic coordinates of the sink node on the considered area. The sink location ( X , Y ) varies on the line X = Y . The observation that can be made from Fig. 13 is that the transmission cost of baseline approaches except CWCD A strongly depends on the sink location. Transmission cost required for data aggregation with CWCD A is robust to sink location, but it requires more transmission cost compared to the proposed method across all sink locations. The considered traditional approaches (SPRM, Hybrid CS and NoN-CS) yield lower transmission cost when the sink is at the center of the considered area. In fact, if the sink is at the center of the considered area, e very node can connect to the sink with the shortest distance. An interesting inference that can be made from Fig. 13 is that the transmission cost of the proposed L WCD A algorithm for data aggregation in both grid-wise and random deployment scenario is robust to the sink location and is much lower compared to that of the traditional methods for all the considered sink locations. This can be justified by noting that the proposed L WCDA algorithm aggregates data through clustering where required transmission cost is independent of the sink location. In addition, to aggregate measurements from randomly dis- tributed CHs, which are connected through MST along with the sink node, incur almost same transmission cost irrespectiv e of the sink location. The data aggregation algorithms proposed in the literature such as [7]-[18] do not discuss hardware implementation details. The implementation procedure and assumptions con- sidered for software simulations differ when it comes to real- time hardware implementation. For the completeness of the proposed algorithm and analysis, we describe an optimal way of implementation which shows the efficacy of the proposed method in a real-time scenario. V I . R E A L - T I M E I M P L E M E N TA T I O N The in-house IITH Motes [20] are used for implementing the proposed data aggregation algorithm (L WCDA). The IITH Mote is a ZigBee system-on-chip combining a 2.4 GHz IEEE 802.15.4 radio transcei ver with a 8 MHz, 8-bit processor having 128 kB of flash memory and 8 kB of RAM. Tin yOS [37] is used to program the proposed L WCDA algorithm on the nodes. Based on the required compression rate Γ , the threshold value T hr will be decided. From the selected M CHs, the probability of the i th node becoming a CH is P C H = M N as discussed in Section III. Let T u i denote the generated uniform random value at the i th node, i.e., T u i ∈ U ∼ [0 , 1] . If T u i ≤ T hr then the i th node becomes a CH. CH probability can be re written as P C H = P r { T u i ≤ T hr } = T hr where i ∈ [1 , N ] . For example, if the threshold is considered to be T hr = 0 . 3 then on an average 30% of the nodes become CHs ( P C H = 0 . 3 ) and Γ = 70% compression can be achieved. The sink node broadcasts a starting packet with the specified threshold T hr value. Each node in the network broadcasts this packet once so that the threshold value reaches e very other node in the network. The nodes calculate Receiv ed Signal Strength Indication (RSSI) values from the receiv ed packets and stores them in a table. It is important to note that each node will have RSSI v alues of all the other nodes that are in its radio range (communication range). Using the created RSSI table, the nodes, which are selected as leaf nodes, connect to nearer CHs and CHs form MST . As the sink node requires the knowledge of Φ , i.e., { α i } values and respecti ve indices ∆ j , where i ∈ c j and j ∈ [1 , M ] , in the initial phase (i.e., first cycle of data aggregation), CH sends { α i } , ∆ j to the sink along with the final measurement P i ∈ c j α i x i . This is a small overhead as α and the respectiv e node index (node address) together can take a maximum of three octets when short address mode is considered. By the end of the initial phase, all the nodes register their respective destination node addresses. In data sensing phase (i.e., from second data aggregation cycle on-wards), in each cycle, all the 11 Algorithm 1 Pseudo code for the data aggregation algorithm at node level Require: T hr 1: Data collection round r = 0 2: Generate uniform random value T u i ∈ U ∼ [0 , 1] ( i refers node number) 3: if T u i ≤ T hr then 4: T y pe = C H 5: else 6: T y pe = Leaf node 7: end if 8: while r ≥ 0 do 9: if T y pe == C H then 10: r = r + 1 11: if r == 1 then 12: Broadcast CH packet 13: Generate uniform random value R i ∈ U ∼ [0 , 1] 14: if R i ≤ 0 . 5 then 15: α i = − 1 16: else 17: α i = 1 18: end if 19: Discover the next hop destination node C H dest : CH node or the leaf node in MST towards the sink 20: end if 21: Measure data sample x i 22: Compute: α i x i 23: Receiv e data packets from all the leaf nodes and descendant CHs 24: Compute: P i ∈ c j α i x i 25: Send CH data packet to C H dest using pack and forward method 26: else 27: r = r + 1 28: if r == 1 then 29: Find RS S I h = max h { RS S I of CHs which are in the radio range } 30: Leaf dest = C H h 31: if Leaf dest == N U LL then 32: Disco ver the next hop destination node Leaf dest = neighboring leaf node in the shortest path towards nearer CH 33: end if 34: Generate uniform random value R i ∈ U ∼ [0 , 1] 35: if R i ≤ 0 . 5 then 36: α i = − 1 37: else 38: α i = 1 39: end if 40: end if 41: Measure data sample x i 42: Compute: α i x i 43: Send the data packet to Leaf dest 44: end if 45: end while leaf nodes compute their measurements and send them to their respectiv e destined CHs. Further , each CH computes the final measurement and forwards it to the sink. Pseudo code of the node level implemented algorithm is described in Algorithm 1. The proposed L WCD A algorithm is independent of the deployment scenario. As an example to verify the implemen- tation efficacy of the L WCD A, we deployed 50 nodes grid- wise in an area of 321 . 44 f t 2 as shown in Fig. 14. The sink node is connected to a PC that collects measured data from all the nodes in the network. For illustration, we considered a threshold T hr = P C H = 0 . 3 and obtained 14 CHs among the deployed 50 nodes in a particular realization of the e xperiment, while the remaining nodes are connected to their respectiv e CHs. Accordingly , 14 clusters were formed, and thus the rows of the measurement matrix Φ 14 × 50 were generated. The resultant measurement matrix Φ 14 × 50 is sho wn in Fig. 15. T o Fig. 14: Experimental setup with N = 50 nodes deployed in an area of 321 . 44 f t 2 . 10 20 30 40 50 Node index 2 4 6 8 10 12 14 Clus ter num ber -1 -0.8 -0.6 -0.4 -0.2 0 0.2 0.4 0.6 0.8 1 Fig. 15: Measurement matrix Φ 14 × 50 constructed from the real field deploy- ment with N = 50 nodes and Γ = 70% . ev aluate the efficac y of the proposed method, we considered coherence as the metric. W e repeated the above experiment for a range of threshold values T hr = 0 . 1 : 0 . 1 : 0 . 9 , and in each case of T hr , the measurement matrix Φ was constructed. T o compute the coherence of Φ against all the compression rates, we obtained Φ for 10 realizations and for each T hr . Each realization gives one mutual coherence value µ for a pair of Φ and Ψ . W e then av eraged µ ov er 10 realizations for each T hr . T o compare with the real deployment, we simulate a similar scenario in software. A verage coherence values of the matrix A designed from both the experiment (exp.) as well as the simulation (sim.) are plotted in Fig. 16. Fig. 16 illustrates that the coherence values of the matrix A where the proposed measurement matrix is constructed from the experiment as well as the simulation with DFT , DCT and D WT bases are in excellent agreement. These results show efficac y of the proposed method in a real-time implementation. It justifies our claim that the proposed method does not require any extra computational overhead (such as the generation of the individual columns of the matrix Φ , storage of Φ etc.). Hence, the proposed method can be implemented on low end commercial off-the-shelf IoT nodes. 12 10 20 30 40 50 60 70 80 90 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 LWCDA CS DCT exp. LWCDA CS DCT sim. LWCDA CS DFT exp. LWCDA CS DFT sim. LWCDA CS DWT exp. LWCDA CS DWT sim. Fig. 16: Comparison of the mutual coherence µ for the CS matrix A = ΦΨ where Φ is constructed from the real-time deployment and simulations against different compression rates Γ , with N = 50 nodes and for different bases DFT , DCT and DWT . Mutual coherence curves from simulation and real- time deployment are very close and demonstrate the ef ficacy of the proposed method. V I I . C O N C L U S I O N In this paper , we proposed a CS based data aggregation method for IoT networks which is both light weight (possess- ing low complexity) and energy efficient. In the proposed data aggregation algorithm, to minimize the transmission cost, data is aggregated from non-overlapping clusters where each node can contribute to one measurement. Hence, the columns of the measurement matrix constructed from the proposed algorithm are coherent and recovery is not possible for the data which is sparse in the canonical basis (Identity matrix). Ho wev er , we showed that the measurement matrix when combined with the popular bases (DFT , DCT , DWT , Laplacian and DiWT) can guarantee the recovery of data with high fidelity . Unlike con ventional methods, in the proposed data ag- gregation method the node-le vel complexity is independent of the network size and data sparsity . The comparison of the transmission cost concludes that the proposed method is energy efficient and can aid in extending the network lifetime by achie ving minimal transmission cost. Hardware implementation demonstrated the ef ficacy of the proposed algorithm in a real-time implementation. Further, through the analysis of the measurement matrix combined with the popular bases, we found that our data aggregation method using the DFT basis yields a better reconstruction quality than other bases. Howe ver , it is still unknown whether there exists a relation between the measurement matrix and the DFT basis. W e hope to provide a deeper insight in our future in vestigations and present theoretical guarantees. W e observed that there is a slight variation in the performance obtained through DFT and DCT bases. In future, we will pursue a thorough analysis of this discrepancy in the performance variation and study the behavior of energy consumption of the proposed method in the presence of interference. R E F E R E N C E S [1] A. Zanella, N. Bui, A. Castellani, L. V angelista, and M. Zorzi, “Internet of things for smart cities," IEEE Internet of Things J . , v ol. 1, no. 1, pp. 22-32, Feb. 2014. [2] W . Xu, Y . Zhang, Q. Shi, and X. W ang, “Energy management and cross layer optimization for wireless sensor network powered by heteroge- neous energy sources," IEEE T rans. W ir eless Commun., vol. 14, no. 5, pp. 2814-2826, May 2015. [3] R. Jurdak, A. G. Ruzzelli, and G. M. P . O’Hare, “Radio sleep mode optimization in wireless sensor networks," IEEE Tr ans. Mobile Comput., vol. 9, no. 7, pp. 955-968, Jul. 2010. [4] R. Rajagopalan and P . K. V arshney , “Data-aggregation techniques in sensor networks: A survey ," IEEE Commun. Surv . T utorials, v ol. 8, no. 4, pp. 48-63, Fourth Quarter 2006. [5] D. L. Donoho, “Compressed sensing," IEEE T rans. Inf. Theory, vol. 52, no. 4, pp. 1289-1306, Apr . 2006. [6] J. Haupt, W . U. Bajwa, M. Rabbat, and R. Nowak, “Compressed sensing for netw orked data," IEEE Signal Process. Mag., vol. 25, no. 2, pp. 92- 101, Mar . 2008. [7] L. Xiang, J. Luo, and C. Rosenberg,“Compressed data aggregation: Energy-ef ficient and high-fidelity data collection," IEEE/ACM T rans. Netw ., vol. 21, no. 6, pp. 1722-1735, Dec. 2013. [8] C. Caione, D. Brunelli, and L. Benini, “Distributed compressive sam- pling for lifetime optimization in dense wireless sensor networks," IEEE T rans. Ind. Informat., vol. 8, no. 1, pp. 30-40, Feb . 2012. [9] J. Luo, L. Xiang, and C. Rosenberg, “Does compressed sensing improve the throughput of wireless sensor networks?," in Pr oc. IEEE Int. Conf. Commun., 2010. [10] L. Xiang, J. Luo, and A. V asilakos, “Compressed data aggregation for energy efficient wireless sensor networks," in Proc. 8th annu. IEEE commun. Soc. Conf. Sensor , Mesh, Ad hoc Commun., Netw ., 2011. [11] W . W ang, M. Garofalakis, and K. Ramchandran, “Distributed sparse random projections for refinable approximation," in Pr oc. 6th Int. Symp. on Inform. Process. in Sensor Netw ., Cambridge, MA, pp. 331-339, 2007. [12] D. Ebrahimi and C. Assi, “Optimal and efficient algorithms for projection-based compressive data gathering," IEEE Commun. Lett., vol. 17, no. 8, pp. 1572-1575, Aug. 2013. [13] D. Ebrahimi and C. Assi, “ A distributed method for compressive data gathering in wireless sensor networks," IEEE Commun. Lett., vol. 18, no. 4, pp. 624-627, 2014. [14] S. A. Daresari and J. Abouei, “T oward cluster-based weighted compres- siv e data aggregation in wireless sensor networks," J. Ad Hoc Netw ., vol. 36, pp. 368-385, 2016. [15] K. C. Lan and M. Z. W ei, “ A Compressibility-Based Clustering Algo- rithm for Hierarchical Compressive Data Gathering", IEEE Sensors J., vol. 17, no. 8, pp. 2550-2562. [16] S. Lee, S. Pattem, M. Sathiamoorthy , B. Krishnamachari, and A. Ortega, “Compressed sensing and routing in multi-hop networks," USC CENG, T ech. Rep., 2009. [17] G. Quer, R. Masiero, D. Munaretto, M. Rossi, J. Widmer , and M. Zorzi, “On the interplay between routing and signal representation for com- pressiv e sensing in wireless sensor networks," in Proc. Inform. Theory and Applicat., Feb. 2009. [18] H. Zheng, F . Y ang, X. Tian, X. Gan, X. W ang, and S. Xiao, “Data gathering with compressive sensing in wireless sensor networks: a random walk based approach," IEEE Tr ans. P arallel and Distrib. Syst., 2015. [19] S. P . Tirani and A. A vokh, “On the performance of sink placement in WSNs considering energy-balanced compressive sensing-based data aggregation," J. Network and Computer Applicat., vol. 107, pp. 38-55, 2018. [20] P . Rajalakshmi. (2012). IITH Mote-Wir eless Sensor Communication Module. [Online]. A vailable: http://www .iith.ac.in/ ∼ raji/downloads/ IITH-mote-webpage.pdf. [21] S. Foucart and H. Rauhut, “ A Mathematical Introduction to Compressiv e Sensing," Basel, Switzerland: Birkhäuser , 2013. [22] E. J. Candes and T . T ao, “The Dantzig selector: Statistical estimation when p is much larger than n," Ann. Statist., vol. 35, no. 6, pp. 2313- 2351, 2007. [23] D. Needell and J. A. Tropp, “CoSaMP: Iterative signal recovery from incomplete and inaccurate samples," Appl. and Computational Harmonic Anal., vol. 26, no. 3, pp. 301-321, May 2009. [24] J. A. T ropp and A. C. Gilbert, “Signal recovery from random measure- ments via orthogonal matching pursuit," IEEE T rans. Inf. Theory , vol. 53, no. 12, pp. 4655-4666, Dec. 2007. [25] D. Donoho, “Compressed sensing," IEEE T rans. Inf. Theory, vol. 52, pp. 1289-1306, 2006. [26] M. Grant, S. Boyd, and Y . Y e, “CVX: Matlab software for disciplined con vex programming," 2008. 13 [27] E. Candies and T . T ao, “Decoding by linear programming," IEEE T rans. Inf. Theory , vol. 51, no. 12, pp. 4203-4215, Dec. 2005. [28] D. Donoho and J. T anner, “Precise undersampling theorems," Proc. IEEE, vol. 98, no. 6, pp. 913-924, Jun. 2010. [29] K. Derr and M. Manic, “W ireless sensor netw orks-node localization for various industry problems," IEEE T rans. Ind. Informat., vol. 11, no. 3, pp. 752-762, Jun. 2015. [30] W . Poe and J. Schmitt, “Node deployment in large wireless sensor networks: Coverage, energy consumption, and worst-case delay ," in Proc. Asian Internet Eng. Conf., pp. 77-84, 2009. [31] S. Kumar , T . H. Lai, and J. Balogh, “On k-cov erage in a mostly sleeping sensor network," in Pr oc. ACM 10th Annu. Int. Conf. on Mobile Computing and Networking, 2004. [32] C. Lva, Q. W anga, W . Y anb, and Y . Shena, “Diffusion wavelet basis algorithm for sparse representation of sensory data in WSNs," J . Signal Pr ocess., vol. 140, pp. 12-31, 2017. [33] S. Lindsay and C. S. Ragha vendra, “PEGASIS: Power-ef ficient gathering in sensor information system," in Pr oc. Int. Conf. Commun. (ICC), Bejing, China, vol. 3, pp. 1125-1130, 2002. [34] K. C. Lan and M. Z. W ei, “ A compressibility-based clustering algorithm for hierarchical compressiv e data gathering," IEEE Sensors J., vol. 17, no. 8, pp. 2550-2562, Apr . 2017. [35] J. N. Al-Karaki and A. E. Kamal, “Routing techniques in wireless sensor networks: A survey ," IEEE W ireless Commun., vol. 11, no. 6, pp. 6-28, Dec. 2004. [36] H. Safa, W . El-Hajj, and H. Zoubian, “ A robust topology control solution for the sink placement problem in WSNs," J. Network and Computer Applicat., vol. 39, pp. 70-82, 2014. [37] Tin yOs. [online]. A vailable: http://www .tinyos.stanford.edu/tinyos- wiki/index.php/T inyOS_Documentation_Wiki

Original Paper

Loading high-quality paper...

Comments & Academic Discussion

Loading comments...

Leave a Comment