Random Matrices, Integrals and Space-time Systems Babak Hassibi California Institute of Technology DIMACS Workshop on Algebraic Coding and Information.

Slides:



Advertisements
Similar presentations
10.4 Complex Vector Spaces.
Advertisements

Capacity of MIMO Channels: Asymptotic Evaluation Under Correlated Fading Presented by: Zhou Yuan University of Houston 10/22/2009.
Ch 7.7: Fundamental Matrices
MIMO Communication Systems
Matrix Theory Background
Eigenvalues and Eigenvectors
Quantum One: Lecture Canonical Commutation Relations 3.
Symmetric Matrices and Quadratic Forms
Chapter 5 Orthogonality
Random Matrices Hieu D. Nguyen Rowan University Rowan Math Seminar
SUMS OF RANDOM VARIABLES Changfei Chen. Sums of Random Variables Let be a sequence of random variables, and let be their sum:
Tch-prob1 Chapter 4. Multiple Random Variables Ex Select a student’s name from an urn. S In some random experiments, a number of different quantities.
Elementary Linear Algebra Anton & Rorres, 9 th Edition Lecture Set – 08 Chapter 8: Linear Transformations.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
Boyce/DiPrima 9th ed, Ch 11.2: Sturm-Liouville Boundary Value Problems Elementary Differential Equations and Boundary Value Problems, 9th edition, by.
Dirac Notation and Spectral decomposition
Capacity of multi-antenna Gaussian Channels, I. E. Telatar By: Imad Jabbour MIT May 11, 2006.
Lecture II-2: Probability Review
1 10. Joint Moments and Joint Characteristic Functions Following section 6, in this section we shall introduce various parameters to compactly represent.
Orthogonal Matrices and Spectral Representation In Section 4.3 we saw that n  n matrix A was similar to a diagonal matrix if and only if it had n linearly.
MIMO Multiple Input Multiple Output Communications © Omar Ahmad
CHAPTER SIX Eigenvalues
Wireless Communication Elec 534 Set IV October 23, 2007
Compiled By Raj G. Tiwari
GROUPS & THEIR REPRESENTATIONS: a card shuffling approach Wayne Lawton Department of Mathematics National University of Singapore S ,
8.1 Vector spaces A set of vector is said to form a linear vector space V Chapter 8 Matrices and vector spaces.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Two Functions of Two Random.
Linear Algebra (Aljabar Linier) Week 10 Universitas Multimedia Nusantara Serpong, Tangerang Dr. Ananda Kusuma
1 7. Two Random Variables In many experiments, the observations are expressible not as a single quantity, but as a family of quantities. For example to.
Network Systems Lab. Korea Advanced Institute of Science and Technology No.1 Appendix A. Mathematical Background EE692 Parallel and Distribution Computation.
Multiple Random Variables Two Discrete Random Variables –Joint pmf –Marginal pmf Two Continuous Random Variables –Joint Distribution (PDF) –Joint Density.
1 Two Functions of Two Random Variables In the spirit of the previous lecture, let us look at an immediate generalization: Suppose X and Y are two random.
Elementary Linear Algebra Anton & Rorres, 9th Edition
Eigenvectors and Linear Transformations Recall the definition of similar matrices: Let A and C be n  n matrices. We say that A is similar to C in case.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Quantization Codes Comprising Multiple Orthonormal Bases Alexei Ashikhmin Bell Labs  MIMO Broadcast Transmission  Quantizers Q(m) for MIMO Broadcast.
Diagonalization and Similar Matrices In Section 4.2 we showed how to compute eigenpairs (,p) of a matrix A by determining the roots of the characteristic.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Two Random Variables.
University of Houston Cullen College of Engineering Electrical & Computer Engineering Capacity Scaling in MIMO Wireless System Under Correlated Fading.
1 8. One Function of Two Random Variables Given two random variables X and Y and a function g(x,y), we form a new random variable Z as Given the joint.
Elementary Linear Algebra Anton & Rorres, 9 th Edition Lecture Set – 07 Chapter 7: Eigenvalues, Eigenvectors.
The Mathematics for Chemists (I) (Fall Term, 2004) (Fall Term, 2005) (Fall Term, 2006) Department of Chemistry National Sun Yat-sen University 化學數學(一)
Principal Component Analysis (PCA)
Mathematical Tools of Quantum Mechanics
Joint Moments and Joint Characteristic Functions.
One Function of Two Random Variables
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
MA2213 Lecture 8 Eigenvectors.
Elementary Linear Algebra Anton & Rorres, 9th Edition
Matrices and vector spaces
Outline Introduction Signal, random variable, random process and spectra Analog modulation Analog to digital conversion Digital transmission through baseband.
Trigonometric Identities
Quantum One.
Quantum One.
Quantum Two.
GROUPS & THEIR REPRESENTATIONS: a card shuffling approach
Lecture on Linear Algebra
7. Two Random Variables In many experiments, the observations are expressible not as a single quantity, but as a family of quantities. For example to record.
Daifeng Wang and Brian L. Evans {wang,
Symmetric Matrices and Quadratic Forms
Quantum Two.
Maths for Signals and Systems Linear Algebra in Engineering Lectures 13 – 14, Tuesday 8th November 2016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR)
Elementary Linear Algebra Anton & Rorres, 9th Edition
8. One Function of Two Random Variables
7. Two Random Variables In many experiments, the observations are expressible not as a single quantity, but as a family of quantities. For example to record.
9. Two Functions of Two Random Variables
16. Mean Square Estimation
7. Two Random Variables In many experiments, the observations are expressible not as a single quantity, but as a family of quantities. For example to record.
8. One Function of Two Random Variables
Presentation transcript:

Random Matrices, Integrals and Space-time Systems Babak Hassibi California Institute of Technology DIMACS Workshop on Algebraic Coding and Information Theory, Dec 15-18, 2003

Outline Overview of multi-antenna systems Random matrices Rotational-invariance Eigendistributions Orthogonal polynomials Some important integrals Applications Open problems

Introduction We will be interested in multi-antenna systems of the form: whereare the receive, transmit, channel, and noise matrices, respectively. Moreover, are the number of transmit/receive antennas respectively, is the coherence interval and is the SNR. The entries of are iid and the entries of are also, but they may be correlated.

Some Questions What is the capacity? What are the capacity-achieving input distributions? For specific input distributions, what is the mutual information and/or cut-off rates? What are the (pairwise) probability of errors? We will be interested in two cases. The coherent case, where is known to the receiver and the non-coherent case, where is unknown to the receiver. The following questions are natural to ask.

Random Matrices A random matrix is simply described by the joint pdf of its entries, An example is the family of Gaussian random matrices, where the entries are jointly Gaussian.

Rotational-Invariance An important class of random matrices are (left- and right-) rotationally-invariant ones, with the property that their pdf is invariant to (pre- and post-) multiplication by any and unitary matrices and. and If a random matrix is both right- and left- rotationally-invariant we will simply call it isotropically-random (i.r.). If is a random matrix with iid Gaussian entries, then it is i.r., as are all of the matrices:

Isotropically-Random Unitary Matrices A random unitary matrix is one for which the pdf is given by When the unitary matrix is i.r., then it is not hard to show that Therefore an i.r. unitary matrix has a uniform distribution over the Stiefel manifold (space of unitary matrices). It is also called the Haar measure.

A Fourier Representation If we denote the columns of by then Using the Fourier representation of the delta function It follows that we can write

A Few Theorems I.r. unitary matrices come up in many applications. Theorem 1 Let be an i.r. random matrix and consider the svd Then the following two equivalent statements hold: 1. are independent random matrices and and are i.r. unitary. 2. The pdf of only depends on Idea of Proof: and have the same distribution for any unitary and

Theorem 2 Let A be an i.r. Hermitian matrix and consider the eigendecomposition. Then the following two equivalent statements are true. 1. are independent random matrices and is i.r. unitary. 2.The pdf of A is independent of U: Theorem 3 Let A be a left rotationally-invariant random matrix and consider the QR decomposition, A=QR. Then the matrices Q and R are independent and Q is i.r. unitary.

Some Jacobians The decompositions and can be considered as coordinate transformations. Their corresponding Jacobians can be computed to be: and for some constant c. Note that both Jacobians are independent of U and Q.

Eigendistributions Thus for an i.r. Hermitian A with pdf we have Integrating out the eigenvectors yields: Theorem 4 Let A be an i.r. Hermitian matrix with pdf Then Note that, a Vandermonde determinant.

Some Examples Wishart matrices,, where G is Ratio of Wishart matrices, I.r. unitary matrix. Eigenvalues are on the unit circle and the distribution of the phases are:

The Marginal Distribution Note that all the previous eigendistributions were of the form: For such pdf’s the marginal can be computed using an elegant trick due to Wigner. Define the Hankel matrix Note that Assume that Then we can perform the Cholesky decomposition F=LL*, with L lower triangular.

Note that implies that the polynomials are orthonormal wrt to the weighting function f(.): Now the marginal distribution of one eigenvalue is given by But

Now upon expanding out and integrating over the variables the only terms that do not vanish are those for which the indices of the orthonormal polynomials coincide. Thus, after the smoke clears In fact, we have the following result. Theorem 5 Let A be an i.r. Hermitian matrix with Then the marginal distribution of the eigenvalues of A is

Orthogonal Polynomials What was just described was the connection between random matrices and orthogonal polynomials. For Wishart matrices, Laguerre polynomials arise. For ratios of Wishart matrices it is Jacobi polynomials, and for i.r. unitary matrices it is the complex exponential functions (orthogonal on the unit circle). Theorem 5 gives a Christoffel-Darboux sum and so The above sum gives a uniform way to obtain the asymptotic distribution of the marginal pdf and to obtain results such as Wigner’s semi-circle law.

Remark The attentive audience will have discerned that my choice of the Cholesky factorization of F and the resulting orthogonal polynomials was rather arbitrary. It is possible to find the marginal distribution without resorting to orthogonal polynomials. The result is given below.

Coherent Channels Let us now return to the multi-antenna model where we will assume that the channel H is known. We will assume that where are the correlation matrices at the transmitter and receiver and G has iid CN(0,1) entries. Note that can be assumed diagonal wlog. According to Foschini&Telatar: 1.When

2. When 3. When 4. In the general case: Cases 1-3 are readily dealt with using the techniques developed so far, since the matrices are rotationally-invariant. Therefore we will do something more interesting and compute the characteristic function (not just the mean). This requires more machinery, as does Case 4, which we now develop.

A Useful Integral Formula Using a generalization of the technique used to prove Theorem 5, we can show the following result. Theorem 6 Let functions be given and define the matrices Then where

Theorem 6 was apparently first shown by Andreief in A useful generalization has been noted in Chiani, Win and Zanella (2003). Theorem 7 Let functions be given. Then where for the tensor we have defined and the sums are over all possible permutations of the integers 1 to m.

An Exponential Integral Theorem 8 (Itzyskon and Zuber, 1990) Let A and B be m-dimensional diagonal matrices. Then where Idea of Proof: Use induction. Start by partitioning

Then rewrite so that the desired integral becomes

The last integral is over an (m-1)-dimensional i.r. matrix. And so if use the integral formula (at the lower dimension) to do the integral over U, we get An application of Theorem 6 now gives the result.

Characteristic Function Consider The characteristic function is (assuming M=N) Successive use of Theorems 6 and 8 give the result.

Non-coherent Channels Let us now consider the non-coherent channel. where H is unknown and has iid CN(0,1) entries. Theorem 9 (Hochwald and Marzetta, 1998) The capacity- achieving distribution is given by S = UD, where U is T-by-M i.r. unitary and D is an independent diagonal. Idea of Proof: Write S=UDV*. V* can be absorbed in H and so Is not needed. Optimal S is left rotationally-invariant.

Mutual Information Determining the optimal distribution on D is an open problem. However, given D, one can compute all quantities of interest. The starting point is The expectation over U is now readily do-able to give p(X|D). (A little tricky since U is not square, but doable using Fourier Representation of delta functions and Theorems 6 and 8.)

Other Problems Mutual information for almost any input distribution on D can be computed. Cut-off rates for coherent and non-coherent channels for many input distributions (Gaussian, i.r. unitary, etc.) can be computed. Characteristic function for coherent channel capacity in general case can be computed. Sum rate capacity of MIMO broadcast channel in some special cases can be computed. Diversity of distributed space-time coding in wireless networks can be determined.

Other Work and Open Problems I did not touch at all upon asymptotic analysis using the Stieltjes transform. Open problem include determining the optimal input distribution for the non-coherent channel and finding the optimal power allocation for coherent channels when there is correlation among the transmit antennas.