Presentation is loading. Please wait.

Presentation is loading. Please wait.

ONR MURI: NexGeNetSci Optimizing Information Flow in Networks Third Year Review, October 29, 2010 Babak Hassibi Department of Electrical Engineering California.

Similar presentations


Presentation on theme: "ONR MURI: NexGeNetSci Optimizing Information Flow in Networks Third Year Review, October 29, 2010 Babak Hassibi Department of Electrical Engineering California."— Presentation transcript:

1 ONR MURI: NexGeNetSci Optimizing Information Flow in Networks Third Year Review, October 29, 2010 Babak Hassibi Department of Electrical Engineering California Institute of Technology

2 Theory Data Analysis Numerical Experiments Lab Experiments Field Exercises Real-World Operations First principles Rigorous math Algorithms Proofs Correct statistics Only as good as underlying data Simulation Synthetic, clean data Stylized Controlled Clean, real-world data Semi- Controlled Messy, real-world data Unpredictable After action reports in lieu of data Hassibi Optimizing information flow

3 Overview of Work Done 1.Network information theory wired and wireless networks, entropic vectors, Stam vectors, groups, matroids, Ingleton inequality, Cayley’s hyperdeterminant, entropy power inequality 2.Estimation and control over lossy networks asymptotic analysis of random matrix recursions, universal laws for networks 3.Social network problems searchibility and distance-dependent Kronecker graphs many-to-one matchings over social networks 4. Distributed adaptive consensus

4 Network Information Theory Network information theory deals studies the limits of information flow in networks. Unlike point-to-point problems (solved by Shannon in 1948) almost all network information theory problems are open. ONR MURI: NexGeNetSci relay transmitterreceiver P(y2|x) P(y1|x) s1 s2 x y1 y2 transmitter P(y1|x1,x2) P(y2|x1,x2) y1 y2 x1 x2

5 So How are Current Networks Operated? ONR MURI: NexGeNetSci Almost invariably via a two-step procedure: use coding to make each link error free (up to the Shannon capacity) view information as a regular “flow” and solve a flow problem over the network (routing, etc.) But is information a regular flow?

6 Is Information a Regular Flow? Is Information a Regular Flow? ONR MURI: NexGeNetSci routing delivers 1.5 bits per receiver taking delivers 2 bits per receiver

7 A General Network Problem ONR MURI: NexGeNetSci Network Suppose each source wants to communicate with its corresponding destination at rate The problem with the above formulation is that it is infinite letter, and that for each T it is a highly non-convex optimization problem (both in the input distributions and the “network operations”).

8 Entropy Vectors Consider n discrete random variables of alphabet size N and define the entropy of as This defines a dimensional vector called an entropy vector The space of all entropic vectors is denoted by and can be shown to be a convex cone

9 A Convex Formulation of the Network Problem Associate a random variable with each edge of the network. The sum rate capacity of the network is given by ONR MURI: NexGeNetSci subject to and (for sources) (at each node) (at each edge)

10 Remarks Network information theory is basically the problem of identifying which is open for n>3. The following issues need to be addressed: - given a vector in, is it entropic? - given an entropic vector, realize it. - can these by done in a distributed way? The framework results in an explosion in the number of variables - is this really necessary? ONR MURI: NexGeNetSci

11 Stam Vectors and Wireless Networks Consider n continuous random vectors of dimension N and define the Stam entropy of as This defines a dimensional vector call a Stam vector The space of all Stam vectors is denoted by and can be shown to be a compact convex set Main Result: Network information theory problems for wireless networks can be cast as linear optimization over For n=2, is characterized by the entropy power inequality

12 Some Related Objects Entropy vectors are related to: Quasi-uniform distributions - typical sequences Finite groups - statistical mechanics, symmetric group - random walk over entropy vectors Matroids - representibility of matroids - optimal linear networks codes via linear programming Determinantal inequalities - Cayley’s hyperdeterminant ONR MURI: NexGeNetSci

13 Entropy and Groups Given a finite group and n subgroups the dimensional vector with entries where is entropic Conversely, any entropic vector for some collection of n random variables, corresponds to some finite group and n of its subgroups Abelian groups are not sufficient to characterize all entropic vectors---they satisfy a so-called Ingleton inequality, which entropic vectors can violate –this is why linear network coding is not sufficient---linear codes form an Abelian group

14 Codes from Non-Abelian Groups ONR MURI: NexGeNetSci If a and b are chosen from a non-Abelian group, one may be able to infer them from ab and ba. There is also a larger set of signals that one may transmit.

15 Where is this Coming From? Ans: Stat Mech Suppose we have T particles that can be in one of N states with probability Then the typical micro-states will be those for which The entropy is the log of the number of microstates One can think of the numerator as the size of the symmetric group of T elements and the denominator as the size of a certain subgroup of. ONR MURI: NexGeNetSci

16 Entropy and Partitions ONR MURI: NexGeNetSci 1 2 3 1’ 2’ 3’

17 Staking out the Entropy Region Take a set of size T and for each random variable partition it into N sets The entropies and joint entropies can be computed from the partitions and their intersections By making local changes to the partitions, we can move from one entropy vector to the next As T and N grow, one can stake out the entire entropic region to desired accuracy This idea can be used to perform random walks on entropy vectors and thereby MCMC methods for entropy optimization ONR MURI: NexGeNetSci

18 I<0 is the Ingleton bound. The above is a MCMC maximization of I for T=100 and N=2. The resulting value of.025 is much superior to previous best known violation 0.007.

19 Optimizing the Information Flow in Networks This optimization can be done in networks, provided we respect the network topology. The sum rate can be optimized in a distributed fashion: each output edge randomly changes its partition based on information received by the sinks ONR MURI: NexGeNetSci

20 The Vamos Network ONR MURI: NexGeNetSci Constructed from the Vamos matroid, the smallest nonrepresentable matroid Capacity unknown. Known to be less than 60/11.

21 ONR MURI: NexGeNetSci Using the distributed MCMC method, we can find a binary solution with sum rate 5.

22 The Non-Pappus Network ONR MURI: NexGeNetSci A 9 element, rank 3 nonrepresentable matroid. Capacity of network unknown.

23 N=2, T=100, C=2/3 ONR MURI: NexGeNetSci

24 N=3, T=100, C=0.8228 ONR MURI: NexGeNetSci

25 The Group PGL(2,p) We have performed computer search to find the smallest finite group that violates the Ingleton inequality It is the projective linear group PGL(2,5) with 120 elements The groups PGL(n,p) and GL(n,p) can be used to construct codes stronger than linear network codes ONR MURI: NexGeNetSci

26 Entropy and Matroids A matroid is a set of objects along with a rank function that satisfies submodularity Entropy satisfies submodularity and therefore defines a matroid However, not all matroids are entropic A matroid is called representable if it can be represented by a collection of vectors over some (finite) field. All representable matroids are entropic, but not all entropic matroids are representable When an entropic matroid is representable, the corresponding network problem has an optimal solution which is a linear network code (over the finite field which represents the matroid)

27 The Fano Matroid ONR MURI: NexGeNetSci The Fano matroid has a representation only over GF(2) ab c de f g

28 The Non-Fano Matroid ONR MURI: NexGeNetSci The Non-Fano matroid has a representation over every field except GF(2) ab c de f g

29 A Network with no Linear Solution ONR MURI: NexGeNetSci i wants cwant bwants a df e g wants cwants bwants a abc This network has no linear coding solution with capacity 7 The linear coding capacity can be shown to be 70/11 < 7 A nonlinear network code of capacity 7 can be found hjk

30 Matroid Representations Unfortunately, determining whether a general matroid is representable is a classical open problem in matroid theory However, the question of whether a matroid is binary representable has a relatively simple answer –the matroid must have no U(2,4) minor, i.e., no 4- element minor such that all pairs are independent and all triples dependent---see matrix below Similar results hold for ternary and quaternary representability, but that is about it. ONR MURI: NexGeNetSci

31 Binary Entropic Vectors A vector in is the entropy vector of n linearly-related binary random variables iff: 1.it has integer entries 2. 3.it satisfies submodularity 4.for every and every the 15-dimensional entropy vector corresponding to is not U(2,4). ONR MURI: NexGeNetSci

32 Optimal Linear Binary Network Codes The sum rate capacity of a network over the class of linear binary network codes is given by ONR MURI: NexGeNetSci subject to, the poly-matroidal cone, and (for sources) (at each node) (at each edge) for every and every the 15-dimensional entropy vector corresponding to is in the convex cone of the entropy region of 4 binary rvs.

33 Comments We have reduced the problem of optimal linear binary network coding to linear programming In general, the complexity of the linear program is exponential - variables; submodular inequalities minors to consider However, if we define r = no. of sources, then - we only have variables - we only have minors to consider - we could significantly fewer submodular inequalities ONR MURI: NexGeNetSci

34 Estimation and Control over Lossy Networks There is a great deal of recent interest in estimation and control over lossy networks. While in many cases (especially in estimation) determining the optimal algorithms is straightforward, determining the performance (stability, mean-square- error, etc.) can be quite challenging (see, e.g., Sinopoli et al). The main reason is that the system performance is governed by a random matrix Riccati recursion, which is incredibly difficult to analyze. ONR MURI: NexGeNetSci

35 Large System Analysis When the dynamical system being estimated or controlled has a large state space dimension, we have proposed a method of analysis based on large random matrix theory. The contention is that when the system dimension and the network are large, the performance of the system exhibits universal laws that depend only on the macroscopic properties of the system and network. The main tool for the analysis is the Stieltjes transform of a random matrix A ONR MURI: NexGeNetSci s(z) = E ( trace (zI – A)^{-1} )/n From which the marginal eigendistribution of A can be found via p(λ) = lim Im( s(λ+j) )/2π →0+

36 Example Consider a MIMO linear time-invariant system with random output measurements that are randomly dropped across some lossy network ONR MURI: NexGeNetSci

37 Consensus What is Consensus? –Given a network where nodes have different values, update over time to converge on a single value –In many cases, we would like convergence to the sample average –Simple local averaging often works Motivation –Sensor network application –Synchronizing distributed agents Agree on one value to apply to all agents

38 Distributed Adaptive Consensus How to quickly reach consensus in a network is important in many applications Local weighted averaging often allows consensus across a network –for example, Metropolis weighting (which requires only the knowledge of the degree of one’s node) works However, if global knowledge of the network topology is available, optimal weights (to minimize the consensus time) can be found using semi-definite programming However, the semi-definite program cannot be made distributed, since the sub-gradient of the second largest eigenvalue requires global knowledge We have developed an algorithm that simultaneously updates both the local averages and the weights using only local information –it computes the gradient of a certain quadratic cost It can be shown to reach consensus faster than Metropolis weighting

39 ONR MURI: NexGeNetSci

40 Social Network Problems 1.Kronecker graphs recently introduced by Leskovec while having many nice properties, these graphs are not searchable we have extended these graphs to “distance-dependent” Kronecker graphs which are searchable 2.Many-to-one matching problems over social networks prove the existence of pairwise stable matchings develop greedy algorithms to achieve such matchings characterize price-of-anarchy of such algorithms

41 Overview of Work Done 1.Network information theory wired and wireless networks, entropic vectors, Stam vectors, groups, matroids, Ingleton inequality, Cayley’s hyperdeterminant, entropy power inequality 2.Estimation and control over lossy networks asymptotic analysis of random matrix recursions, universal laws for networks 3.Social network problems searchibility and distance-dependent Kronecker graphs many-to-one matchings over social networks 4. Distributed adaptive consensus


Download ppt "ONR MURI: NexGeNetSci Optimizing Information Flow in Networks Third Year Review, October 29, 2010 Babak Hassibi Department of Electrical Engineering California."

Similar presentations


Ads by Google