Optimal computation of symmetric Boolean functions in Tree networks
In this paper, we address the scenario where nodes with sensor data are connected in a tree network, and every node wants to compute a given symmetric Boolean function of the sensor data. We first consider the problem of computing a function of two n…
Authors: Hemant Kowshik, P. R. Kumar
Wireless sensor networks are composed of nodes with sensing, wireless communication and computation capabilities. In sensor network applications, one is not interested in the raw sensor measurements, but only in computing some relevant function of the measurements. For example, one might want to compute the mean temperature for environmental monitoring, or the maximum temperature in fire alarm systems. Thus, we need to move away from a data-centric paradigm, and focus on efficient in-network computation and communication strategies which are function-aware.
There are two possible architectures for sensor networks that one might consider. First, one can designate a single collector node which seeks to compute the function of interest. Alternately, one can suppose that every node in the network wants to compute the function. The latter architecture can be viewed as providing implicit feedback to each sensor node, which could be useful in applications like fault monitoring and adaptive sensing. For example, sensor nodes could modify their sampling rate depending on the value of the function. In this paper, we will focus on strategies which achieve function computation with zero error for all sensor nodes.
In this paper, we abstract out the medium access control problem, and view the network as a graph with edges representing essentially noiseless wired links between nodes. Hence, we focus on strategies for combining information at intermediate nodes, and optimal codes for transmissions on each edge. This is a significant departure from the traditional decode and forward paradigm in wireless networks. Moreover, the strategy for computation may benefit from interactive information exchange between nodes.
We consider a graph where each node has a Boolean variable and focus on the problem of symmetric Boolean function computation. We adopt a deterministic formulation of the problem of function computation, allowing zero error. We consider the problem of worst-case function computation, without imposing a probability distribution on the node measurements. Further, instead of restricting a strategy to compute just one instance of the problem, we allow for nodes to accumulate a block of B measurements, and realize greater efficiency by using block codes. The set of admissible strategies includes all interactive strategies, where a node may exchange several messages with other nodes, with node i's transmission being allowed to depend on all previous transmissions heard by node i, and node i's block of measurements.
We begin with the two node problem in Section III, where each node i has an integer variable X i and both nodes want to compute a function f (X 1 , X 2 ) which only depends on X 1 + X 2 . We use a lower bound from the theory of communication complexity, by constructing an appropriate fooling set [1]. For achievability, we devise a single-round strategy so as to minimize the worst-case total number of bits exchanged under the Kraft inequality constraint. For the class of sumthreshold functions, which evaluate to 1 if X 1 + X 2 exceeds a threshold, this single-round strategy is indeed optimal. However, for the class of sum-interval functions, which evaluate to 1 if a ≤ X 1 + X 2 ≤ b, the upper and lower bounds do not match. However, the achievable strategy involving separation, followed by coding, can be used for any general function.
In Section IV, we consider Boolean symmetric function computation on trees. Since every edge is a cut-edge, we can obtain a cut-set lower bound for the number of bits that must be exchanged on an edge, by reducing it to a two node problem. For the class of sum-threshold functions, we are able to match the cut-set bound by constructing an achievable strategy that is reminiscent of message passing algorithms. In Section V, for general graphs, we can still derive a cut-set lower bound by considering all partitions of the vertices. We also propose an achievable scheme that consists of activating a subtree of edges and using the optimal strategy for transmissions on the tree. While the upper and lower bounds do not match even for very simple functions, for complete graphs we show that aggregation along trees provides a 2-OPT solution.
The problem of function computation in sensor networks has received much attention recently. In [2], the problem of worst-case block function computation with zero error was formulated. The authors identify two classes of symmetric functions namely type-sensitive functions exemplified by Mean, Median and Mode, and type-threshold functions, exemplified by Maximum and Minimum. The maximum rates for computation of type-sensitive and type-threshold functions in random planar networks are shown to be Θ( 1 log n ) and Θ( 1 log log n ) respectively, where n is the number of nodes. For the case of a designated collector node interested in computing the function, one can derive a per-link optimal strategy for block function computation in tree networks [3].
In contrast, in this paper, we require that every node must compute the function. The latter approach naturally allows the use of tools from communication complexity [1]. The communication complexity of Boolean functions has been studied in [4]. Further, one can consider the direct-sum problem [5] where several instances of the problem are considered together to obtain savings. This block computation approach is used to compute the exact complexity of the Boolean AND function in [6]. This result was considerably generalized in [7] to derive optimal strategies for computing symmetric Boolean functions in broadcast networks. The average and randomized complexity of Boolean functions are studied in [8].
While we focus on worst-case computation in this paper, we could suppose that the measurements are drawn from some joint probability distribution. The problem of source coding with side information has been studied for the vanishing error case in [9], and for the zero error case in [10]. The problem of source coding for function computation with side information has been studied in [11]. However, a tractable information theoretic formulation of the problem of function computation has proved elusive.
In this paper, we allow for all interactive strategies where each node's transmission is allowed to depend on all previous transmissions and the node's measurements. Thus, all network coding strategies [12] are subsumed in this class. The rate region for multi-round interactive function computation has been characterized for two nodes [13], and for collocated networks [14].
Consider two nodes 1 and 2 with variables X 1 ∈ {0, 1, . . ., m 1 } and X 2 ∈ {0, 1, . . ., m 2 }. Both nodes wish to compute a function f (X 1 , X 2 ) which only depends on the value of X 1 + X 2 . To put this in context, one can suppose there are m 1 Boolean variables collocated at node 1 and m 2 Boolean variables at node 2, and both nodes wish to compute a symmetric Boolean function of the n := m 1 + m 2 variables. We pose the problem in a block computation setting, where each node i has a block of B independent measurements, denoted by X B i . We consider the class of all interactive strategies, where nodes 1 and 2 transmit messages alternately with the value of each subsequent message being allowed to depend on all previous transmissions, and the block of measurements available at the transmitting node. We define a round to include one transmission by each node. A strategy is said to achieve correct block computation if for every choice of input (X B 1 , X B 2 ), each node i can correctly decode the value of the function block f B (X 1 , X 2 ) using the sequence of transmissions b 1 , b 2 , . . . and its own measurement block X B i . This is the direct-sum problem in communication complexity.
Let S B be the set of strategies for block length B, which achieve zero-error block computation, and let C( f , S B , B) be the worst-case total number of bits exchanged under strategy S B ∈ S B . The worst-case per-instance complexity of computing a function f (X 1 , X 2 ) is defined as
In this paper, we are only interested in functions f (X 1 , X 2 ) which only depend on X 1 + X 2 . Let us suppose without loss of generality that m 1 ≤ m 2 . We define an interesting class of {0, 1}-valued functions called sum-threshold functions.
Definition 1 (sum-threshold functions): A sum-threshold function Π θ (X 1 , X 2 ) with threshold θ is defined as follows:
For the special case where m 1 = 1, m 2 = 1 and θ = 2, we recover the Boolean AND function, which was studied in [6]. Throughout this paper, we will use tools introduced in [6]. Theorem 1: Given any strategy S B for block computation of the function
Further, there exist single-round strategies S * B and S * * B , starting with nodes 1 and 2 respectively, which satisfy
Thus, the complexity of computing Π θ (X 1 , X 2 ) is given by C(Π θ (X 1 , X 2 )) = log 2 {min(2θ + 1, 2m 1 + 2, 2(n -θ + 1) + 1)}. Proof of achievability: We consider three cases: (a) Suppose θ ≤ m 1 ≤ m 2 . We specify a strategy S * B in which node 1 transmits first. We begin by observing that inputs X 1 = θ , X 1 = (θ + 1) . . . , X 1 = m 1 need not be separated, since for each of these values of X 1 , Π θ (X 1 , X 2 ) = 1 for all values of X 2 .
Thus node 1 has an effective alphabet of {0, 1, . . . , θ }. Suppose node 1 transmits using a prefix-free codeword of length l(X B 1 ). At the end of this transmission, node 2 only needs to indicate one bit for the instances of the block where X 1 = 0, 1, . . . , (θ -1). Thus the worst-case total number of bits is
where w j (X B 1 ) is the number of instances in the block where X 1 = j. We are interested in finding the codebook which will result in the minimum worst-case number of bits. From the Kraft inequality for prefix-free codes we have
Consider a codebook with l(X B 1 ) = ⌈B log 2 (2θ + 1)⌉w(x B 1 ). This satisfies the Kraft inequality since
Hence there exists a prefix-free code for which the worstcase total number of bits exchanged is ⌈B log
The strategy S * * B starting at node 2 can be similarly derived. Node 2 now has an effective alphabet of {0, 1, . . . , θ }, and we have C(Π θ (X 1 , X 2 ), S * * B , B) ≤ ⌈B log 2 (2θ + 1)⌉. (b) Suppose m 1 ≤ m 2 < θ . Consider a strategy S * B in which node 1 transmits first. The inputs X 1 = 0, X 1 = 1, . . . , X 1 = m 2 -1 need not be separated since for each of these values of X 1 , Π θ (X 1 , X 2 ) = 0 for all values of X 2 . Thus node 1 has an effective alphabet of {θm 2 -1, θm 2 , . . . , m 1 }. Upon hearing node 1's transmission, node 2 only needs to indicate one bit for the instances of the block where
. This satisfies the Kraft inequality and we have
The strategy S * * B starting at node 2 can be analogously derived. (c) Suppose m 1 < θ ≤ m 2 . For the case where node 1 transmits first, we construct a trivial strategy S * B where node 1 uses a codeword of length ⌈B log 2 (m 1 + 1)⌉ bits and node 2 replies with a string of B bits indicating the function block. Thus we have C(Π θ (X 1 , X 2 ), S * B , B) ≤ ⌈B log 2 (2m 1 + 2)⌉. Now consider a strategy S * * B where node 2 transmits first. Observe that the inputs X 2 = 0, X 2 = 1, . . . , X 2 = θm 1 -1 need not be separated since for each of these values of X 2 , Π θ (X 1 , X 2 ) = 0 for all values of X 2 . Further, the inputs X 2 = θ , X 2 = θ + 1, . . . , X 2 = m 2 need not be separated. Thus node 1 has an effective alphabet of {θ -m 1 -1, θ -m 1 , . . . , θ }. Upon hearing node 2's transmission, node 1 only needs to indicate one bit for the instances of the block where
. This satisfies the Kraft inequality and we have L = ⌈B log 2 (2(n -θ + 1) + 1)⌉. Since m 1 < θ ≤ m 2 , we have that
The lower bound is shown by constructing a fooling set [1].
A set E ⊆ X × Y is said to be a fooling set, if for any two distinct elements (x 1 , y 1 ), (x 2 , y 2 ) in E, we have either
Proof of Lower Bound: Define the measurement matrix M to be the matrix obtained by stacking the row X B 1 over the row X B 2 . Let E denote the set of all measurement matrices which are made up only of the column vectors from the set
We claim that E is the appropriate fooling set. Consider two distinct measurement matrices M 1 , M 2 ∈ E. Let f B (M 1 ) and f B (M 2 ) be the block function values obtained from these two matrices. If f B (M 1 ) = f B (M 2 ), we are done. Let us suppose f B (M 1 ) = f B (M 2 ), and note that since M 1 = M 2 , there must exist one column where M 1 and M 2 differ. Suppose M 1 has
, where z 1a + z 2a = z 1b + z 2b .
Assume without loss of generality that z 1a < z 1b and z 2a > z 2b .
• If z 1a + z 2a = z 1b + z 2b = θ -1, then the diagonal element f (z 1b , z 2a ) = 1 since z 1b + z 2a ≥ θ . Thus, if we replace the first row of M 1 with the first row of M 2 , the resulting measurement matrix, say
Thus, if we replace the second row of M 1 with the second row of M 2 , the resulting matrix M * is such that f (M * ) = f (M 1 ). Thus, the set E is a valid fooling set with cardinality |Z| B . For any strategy S B , we have C( f , S B , B) ≥ B log 2 |Z|. The cardinality of Z can be modeled as the sum of the coefficients of Y θ and Y θ -1 in a carefully constructed polynomial:
This is solved using the binomial expansion for 1
This completes the proof of Theorem 1. ✷
Further, there exists a single-round strategy S * B which satisfies
)}⌉. Thus, we have obtained the complexity of computing Π θ (X 1 , X 2 ) to within one bit.
The strategy for achievability used in Theorems 1 and 2 suggests an achievable scheme for any general function f (X 1 , X 2 ) of variables X 1 ∈ X 1 and X 2 ∈ X 2 which depends only on the value of X 1 + X 2 . This is done in two stages. Separation: Two inputs x 1a and x 1b need not be separated if f (x 1a , x 2 ) = f (x 1b , x 2 ) for all values x 2 . By checking this condition for each pair (x 1a , x 1b ), we can arrive at a partition of {0, 1 . . ., m 1 } into equivalence classes, which can be considered a reduced alphabet, say A := {a 1 , . . . , a l }. Coding: Let A 0 denote the subset of the alphabet A for which the function evaluates only to 0, irrespective of the value of X
Consider a tree graph T = (V, E), with node set V = {0, 1, . . . , n} and edge set E. Each node i has a Boolean variable X i ∈ {0, 1}, and every node wants to compute a given symmetric Boolean function f (X 1 , X 2 , . . . , X n ). Again, we allow for block computation and consider all strategies where nodes can transmit in any sequence with possible repititions, subject to:
• On any edge e = (i, j), either node i transmits or node j transmits, or neither, and this is determined from the previous transmissions. • Node i's transmission can depend on the previous transmissions and the measurement block X B i . For sum-threshold functions, we have a computation and communication strategy that is optimal for each link.
Theorem 3: Consider a tree network where we want to compute the function Π θ (X 1 , . . . , X n ). Let us focus on a single edge e ≡ (i, j) whose removal disconnects the graph into components A e and V \ A e , with |A e | ≤ |V \ A e |. For any strategy S B ∈ S B , the number of bits exchanged along edge e ≡ (i, j), denoted by C e (Π θ (X 1 , . . . , X n ), S B , B), is lower bounded by C e (Π θ (X 1 ,... ,X n ),S B ,B) ≥ B log 2 {min(2θ + 1,2|A e | + 2,2(n -θ + 1) + 1)}.
Further, there exists a strategy S * B such that for any edge e, C e (Π θ (X 1 ,... ,X n ),S * B ,B) ≤ ⌈B log 2 {min(2θ +1,2|A e |+2,2(n-θ +1)+1)}⌉.
The complexity of computing Π θ (X 1 , . . . , X n ) is given by C e (Π θ (X 1 , . . . , X n )) = log 2 {min(2θ + 1, 2|A e | + 2, 2(n -θ + 1) + 1)}.
Proof: Given a tree network T , every edge e is a cut edge.
The achievable strategy S * B is derived from the achievable strategy for the two node case in Theorem 1. While the transmissions back and forth along any edge will be exactly the same, we need to orchestrate these transmissions so that conditions of causality are maintained. Pick any node, say r, to be the root. This induces a partial order on the tree network. We start with each leaf in the network transmitting its codeword to the parent. Once a parent node obtains a codeword from each of its children, it has sufficient knowledge to disambiguate the letters of the effective alphabet of the subtree, and subsequently it transmits a codeword to its parent. Thus codewords are transmitted from child nodes to parent nodes until the root is reached. The root can then compute the value of the function and now sends the appropriate replies to its children. The children then compute the function and send appropriate replies, and so on. This sequential strategy depends critically on the fact that, in the two node problem, we derived optimal strategies starting from either node. For any edge e, the worst-case total number of bits exchanged is given by C e (Π θ (X 1 ,... ,X n ),S * B ,B) ≤ ⌈B log 2 {min(2θ +1,2|A e |+2,2(n-θ +1)+1)}⌉.✷
One can similarly derive an approximately optimal strategy for sum-interval functions, which we state here without proof.
Original Paper
Loading high-quality paper...
Comments & Academic Discussion
Loading comments...
Leave a Comment