Madras Tradeoff
Madras Tradeoff
A PROJECT REPORT
BACHELOR OF TECHNOLOGY
in
ELECTRICAL ENGINEERING
and
MASTER OF TECHNOLOGY
in
COMMUNICATION SYSTEMS
by
Dr. K. Giridhar
ii
Acknowledgements
I express my deepest thanks to my guide Dr.K.Giridhar for his constant guidance and
advice throughout the work. I would like to thank him for giving me the freedom to
pursue my ideas and encouraging me all the way.
I wish to thank Dr.Srikrishna Bhashyam for bearing me through the endless tech-
nical discussions. I also wish to thank DR. V.V. Rao, Dr.R. Aravind, Dr. De-
vendra Jalihal and other faculty members for the courses they have offered which
kindled my interest in communications and broadened my knowledge base. I thank
Dr.S.A. Choudum and Dr.S.H. Kulkarni for their courses which deepened my inter-
est in mathematics. I would like to thank the TeNet group for new Intel lab, who
provided me with excellent computing facilities.
I would like to thank Ranga, Raghavendra, Lakshmi and other members of Intel
lab for a creative , stimulating and fun environment. I am thankful to Klutto for his
help during the initial stages of my project and for the great discussions with him. I
would specially thank Srinivas for his constant help, motivation and listening to me
patiently. I also am grateful to Srinivas for helping me in the literature survey and
trips to MathScience library for searching books. A thousand thanks to Vamsi who
helped me to rejuvenate my interest in math. Hours and hours of discussions about
mathematical problems and especially geometry are times which I would never forget.
I also wish to thank Shashi for his help and lots of error control coding discussions.
I thank all my Dual Degree friends for the great and enjoyable times we had. I wish
to thank my room neighbor and good friend Sambit for, bearing with me for three
years, never refusing me financial ,technical and computational aid, and for all the
great times I had paining him. I wish to thank all the sixth wing juniors for a great
time I had with them. I thank Ravish, Subbu, Pavan Ram, Bamba, Jaykumar, Ankit,
Sweety, Biju and all my other sixth wing friends who made my stay at IIT memorable
and enjoyable.
I wish to thank my family, who did not force me to go to medical school and
encouraged me all the way. I am very grateful to my grandmother and my peda-
attha for their love and affection towards me. Finally I would like to thank my
parents and my younger brother Shyam for their never ending love and support.
iv
Abstract
Usage of multiple antennas at the transmitter and receiver increase the rate and
reliability of communication over the wireless channels. In order to achieve both rate
and diversity, the input code matrix should follow certain criterion. In this thesis,
design of high rate space time codes, decoding complexity and criterion to achieve
optimal rate diversity tradeoff will be studied.
We first investigate Linear Dispersion Codes (LDC) which achieve ergodic capacity
of the channel. LDC for different number of transmit and receive antennas will be
designed and compared with LDC based on frame theory using simulations.
We then analyze decoding in multiple antenna scenario using Sphere Decoding
algorithm. We propose a new initial radius and it’s update method based on channel
knowledge and noise variance rather than later alone. The reduction in complexity
due to new radius is verified by simulations.
We develop criterion to be satisfied by the space time codes, so as to achieve
optimal rate-diversity tradeoff. Error probability of non-Gaussian codes as a function
of singular values of the code and channel matrix is analyzed. Using this analysis,
conditions for code to achieve optimal tradeoff are derived. Finally, we present full
rate and full diversity space time codes using extension fields.
Contents
Acknowledgements iii
Abstract v
List of Figures x
Notation xiii
1 Introduction 1
1.1 Channel and System Model . . . . . . . . . . . . . . . . . . . . . . . 2
1.2 Thesis Contribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.3 Thesis Outline . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
2 Theoretical Background 6
2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
2.2 Design Metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
2.2.1 Channel Capacity . . . . . . . . . . . . . . . . . . . . . . . . . 6
2.2.2 Diversity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
2.2.3 Rate Diversity Tradeoff . . . . . . . . . . . . . . . . . . . . . . 11
4 Sphere Decoding 24
4.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
4.2 Sphere Decoding . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
4.2.1 The Sphere Decoding Algorithm . . . . . . . . . . . . . . . . . 27
4.3 Complexity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
4.3.1 Covering Radius . . . . . . . . . . . . . . . . . . . . . . . . . 30
4.4 Modified Decoding Rule . . . . . . . . . . . . . . . . . . . . . . . . . 30
4.4.1 Calculation of Major Diagonal . . . . . . . . . . . . . . . . . . 32
4.4.2 Method A . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
4.4.3 Method B . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
4.4.4 Comparing D/2 with R0 . . . . . . . . . . . . . . . . . . . . . 34
4.5 Simulation Results and Conclusion . . . . . . . . . . . . . . . . . . . 35
4.5.1 Infinite Lattice . . . . . . . . . . . . . . . . . . . . . . . . . . 35
4.5.2 Finite Lattice . . . . . . . . . . . . . . . . . . . . . . . . . . . 35
4.6 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37
vii
5 Rate-Diversity Tradeoff 40
5.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
5.2 Diversity-Multiplexing Tradeoff . . . . . . . . . . . . . . . . . . . . . 40
5.2.1 Channel Model . . . . . . . . . . . . . . . . . . . . . . . . . . 40
5.2.2 Definitions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
5.2.3 Tradeoff Curve . . . . . . . . . . . . . . . . . . . . . . . . . . 42
5.2.4 Error events . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42
5.2.5 Tradeoff Achieved by Alamouti Code . . . . . . . . . . . . . . 44
5.3 Tilted QAM Codes . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45
5.3.1 Criterion for a Non Gaussian code to achieve optimal tradeoff
for (M, M ) MIMO system . . . . . . . . . . . . . . . . . . . . 46
5.3.2 Examples of Codes that achieve optimal Tradeoff . . . . . . . 51
5.3.3 Error Analysis for Non Gaussian Codes . . . . . . . . . . . . . 52
5.4 Full rate and Full Diversity codes from Extension fields . . . . . . . . 58
5.4.1 Code Construction . . . . . . . . . . . . . . . . . . . . . . . . 59
5.4.2 Simulation Results . . . . . . . . . . . . . . . . . . . . . . . . 60
5.5 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62
Appendix A 64
A.1 Extension Fields . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64
Appendix B 66
B.1 Proof . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
Appendix C 68
C.1 LDC for (3, 3),(3, 1) and (4, 1) systems . . . . . . . . . . . . . . . . . 68
C.2 Frame theory based LDC for (3, 3) and (4, 1) systems . . . . . . . . . 73
viii
Bibliography 75
ix
List of Figures
3.1 Error Performance of LDC, Frame based LDC for (2,2) system at R =
2bps/Hz . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
3.2 Error Performance of LDC, Frame based LDC for (2,2) system at R =
4bps/Hz using QAM . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
3.3 Comparison of different Frame based Designs. . . . . . . . . . . . . . 22
3.4 Comparison of different LDC for different constraints. . . . . . . . . . 23
xi
List of Tables
xii
Notation
AT if A is real
A† :
AH if A is complex
∗ : complex conjugate
R(A) : real(A)
I(A) : imaginary(A)
⊗ : Kronecker Product
=
˙ : exponential equality (Defined Later)
˙
< : exponentialy less than (Defined Later)
˙
> : exponentialy greater than (Defined Later)
min(a, b) : minimum of a and b
max(a, b) : maximum of a and b
(x)+ : max(0, x)
tr(A) : trace of matrix A
vec(A) ; column stacking of matrix A
E(x) : expectation of random variable x
(M, N ) : MIMO system with M transmitt antenna and N receive antenna
Chapter 1
Introduction
1
complexity of decoding, sphere decoding algorithm which is of cubic complexity at
high SNR can be used. However sphere decoding is computationally complex at lower
SNR.
Zheng and Tse [8] established that there is a tradeoff between robustness and rate
of transmission. They provided with an upper bound on diversity given the rate of
transmission. Practical codes and the criterion to achieve this bound have been found
only for a two antenna systems.
In this thesis, our goal is to understand and design LDC and its variants namely
Frame theory based LDC. We bring down the complexity of sphere decoding by
appropriate choice of initial radius depending on channel conditions. Criterion for a
non-Gaussian code to achieve optimal tradeoff will be presented and error analysis of
non-Gaussian codes will be studied.
V1
h11
x1 y1
h12 h21
V2
x2
y2
VN
hM N
xm yN
Figure 1.1 shows a multiple input multiple output channel with M input antenna
and N output antennas. At any instant M signals satisfying an average power con-
2
dition are transmitted. In this thesis, the channel is modeled as a flat, Rayleigh, and
block fading, with channel knowledge available at receiver. Noise is modeled as Ad-
ditive White Gaussian Noise (AWGN). This model of the channel is used throughout
the thesis, unless specified otherwise. The mathematical model for the above channel
when the channel is assumed to be quasistatic for T channel usages is
pρ
Y= M
HX + V (1.1)
Y, V ∈ CN ×T , H ∈ CN ×M , X ∈ CM ×T
where H is the multiple antenna channel, X denotes the input transmission signal
matrix, V denotes the additive white gaussian noise, Y is the received signal matrix
and ρ denotes the required SNR. hij is modeled as a Rayleigh fading variable with
variance 0.5 in each dimension. vij is modeled as a complex Gaussian with variance
0.5 in each dimension.
In this thesis, we studied the design problem of codes for near channel capacity com-
munication. We studied the problem of decoding these codes using sphere decoding
and its sensitivity to initial choice of radius. The criterion required for non Gaussian
codes to achieve optimal rate diversity tradeoff has been studied.
In this thesis Linear Dispersion Codes (LDC) which achieve ergodic capacity of
channel, were obtained by optimization of capacity function. The performance of
these codes were compared under various power constraints and for various configu-
rations of transmit and receive antenna. LDC based on frame theory were constructed
and their performance was compared with conventional LDC.
LDC are high rate codes, and ML decoding at these rates is very computationally
expensive. Sphere decoding algorithm can be used to decode LDC when the input
symbols are drawn from lattice constellations. Sphere decoding has a very high com-
plexity at low SNR, since the initial radius is chosen by using noise variance. In this
3
thesis, initial radius that is based on channel knowledge is proposed. The new radius
is computationally easy to calculate, unlike other previous choices of initial radius
available in literature. A new method of updating the radius on decoding failure
is also proposed. These ideas were validated by comparing them with conventional
sphere decoding algorithm using simulations.
Tse [8] quantified the optimal rate diversity tradeoff and the criterion for Gaussian
codes to achieve this optimal tradeoff. Yao [9] proposed condition for non-Gaussian
codes to achieve this optimality over two transmit and receive antenna. In this thesis,
we derived the criterion for Non-Gaussian codes to achieve optimal tradeoff for a
general M transmit and M receive antenna configurations. We proved that the codes
which were claimed to achieve this optimality through simulation results theoretically
satisfy the proposed criterion. We also analyzed probability of error for non-Gaussian
codes as function of singular values of channel and code. We used this analysis to
re derive criterion required by a non gaussian code to achieve optimal rate-diversity
tradeoff. We also propose space time codes from extension fields which achieve full
diversity and full rate. These codes can be thought as a generalization of diagonal
algebraic space time codes.
4
for a good initial radius of decoding. We also propose a simple expression for radius
based on the channel. We then proceed to propose a updating scheme for the radius
based on intuition.We finally present simulation results comparing all the schemes.
Chapter 5: In this chapter rate diversity tradeoff will be presented in detail.
Tilted QAM codes are presented. Theoretical criterion required to achieve optimal
tradeoff for any M transmit and M receive antenna is then obtained. Error analysis
of non gaussian codes with regard to rate-diversity tradeoff is also presented. Finally
a set of codes are constructed from extension fields that provide full rate and full
diversity.
5
Chapter 2
Theoretical Background
2.1 Introduction
In this chapter basic theoretical aspects of MIMO systems will be reviewed. Various
metrics for evaluation of space time codes will be considered.
In this section we will review some important metrics that can be used in evaluating
a code performance
Channel capacity for additive white Gaussian noise(AWGN) channels was first derived
by Shannon in his paper ”A Mathematical Theory of Communication”. Space time
channels also exhibit fading and have an extra degree of freedom, spatial dimension.
In this section we shall review capacity of MIMO channels under various conditions
of channel configurations and availability of Channel State Information (CSI).
When perfect Channel knowledge is available at the transmitter, water filling tech-
nique can be used to achieve capacity. Water filling method assumes independent
gaussian inputs. The channel is broken into equivalent sub channels along the dif-
ferent eigenmodes of the channel. The variance of the inputs is taken to be the
6
difference between a constant and the power of the eigenmode. So if λi denote the
singular values of the channel, the variances of the input are chosen as follows.
E(x2i ) = (µ − λ−1
i )
+
where a+ denotes max(o, a), and µ is chosen to meet the power constraint. The
capacity of this channel is given by
X
C(µ) = (ln(µλi ))+
i
This capacity is the upper bound of all feedback capacities. So this can be used as a
measure of how well a feedback scheme is performing with regarding to rate.
= I(x; y|H)
ρ
It can be easily shown that that the choice of Q that maximizes Ψ(Q) is Q = M
I,
which gives the capacity as
ρ
C(ρ) = E[log det(IN + HH† )] (2.2)
M
7
This capacity is also known as ergodic capacity. The expectation can be evalu-
ated analytically by using Wishart distribution and the exact expression is given by
Teletar [13]. Foschini [14] showed that at high SNR.
C ≈ min(M, N ) log(SN R)
Ergodic capacity of various MIMO channels has been plotted in Figure 2.1. We
observe that the capacity indeed scales with log(SN R) as min(M, N ). Every code
Ergodic Capacity Versus SNR for different (N,M)
25
2x2
1X1
3x3
2x1
3x2
20
15
bps/Hz
10
0
0 5 10 15 20 25 30
SNR dB
X induces an equivalent channel and the ergodic capacity of this equivalent channel
can be calculated. This gives an idea of how much maximum rate that the code can
support on an average.
Consider the case when channel H is chosen randomly at the beginning of all time
and held constant for all the uses of the channel. The Shannon capacity of the
channel is zero, since there is a non zero probability that the realized H is incapable
of supporting the desired rate however long we take the code length. In such a case
8
we can talk of a different metric at each rate R called outage probability. Outage
probability of a channel can be defined as the percentage of channel usages which are
unable to support the required rate at a given SNR. Mathematically this translates
to
ρ
Pout (R, ρ) = P (log det(IN + HH† ) < R)
M
We observe that Pout is a function of both ρ and R. In practice, coding is performed
over a finite channel realizations, and transmitter has no knowledge of channel. So
when the realized channel capacity is below R, the receiver cannot decode even with
powerful codes. We can define this as the outage event. This is a very good metric
of evaluation when coding is done over finite channel realizations. Outage capacity
can be defined as the maximum rate Rmax that can be transmitted for a given ρ
and pout . We also understand that pout will be a lower bound on the actual error
probability [8]. Figure 2.3 illustrates the outage probability of (M, N ) channels when
R = M bps/Hz. We observe that the outage curve corresponding to (M, N ) channel
has a slope of M N . Figure 2.3 illustrates the channel outage capacity for different
(M, N ) for a outage probability of 0.01.
Outage Probability Vs SNR for R = M bits
0
10
1x1
2x2
4x4
−1
10
Outage Probability
−2
10
−3
10
−4
10
6 8 10 12 14 16 18 20
SNR dB
9
Outage Capacity at 1% Outage Probability
20
2x2
1x1
18
16
14
12
Outage Capacity
10
0
0 5 10 15 20 25 30 35 40
SNR dB
2.2.2 Diversity
1. T ≥ M
10
2.2.3 Rate Diversity Tradeoff
An (M, N ) system has min(M, N ) degrees of freedom, and hence one can transmit
independent data in these streams to increase the rate or can repeat the data across
the streams to achieve diversity. Hence we observe that there is a fundamental tradeoff
R
between rate and diversity. If we define r = log(SN R)
then the tradeoff between r and
diversity is given by [8]
d(r) = (M − r)(N − r)
The details of this tradeoff will be dealt in Chapter 4. The performance of the code
can be evaluated on the basis of the maximum diversity it can achieve for a given
rate. This is a useful metric to compare codes which have same diversity gain at a
fixed rate.
11
Chapter 3
3.1 Introduction
In this chapter we will analyze Linear Dispersion Codes (LDC) and LDC based on
frame theory. We will give LDC constructions for various (M, N ). We then compare
them with codes based on frame theory.
Linear Dispersion Codes are ergodic capacity achieving codes. The capacity of equiv-
alent channel, using LDC can be made to scale with SNR as min(M, N ) log(SN R).
Also these codes unlike V-BLAST, can be designed for any (M, N ).
Definition: A linear dispersion code is one for which
Q
X
X= (sq Cq + s∗q Dq ) (3.1)
q=1
Cq , Dq ∈ CM ×T ; sq , s∗q ∈ C
where s1 , . . . , sq are information symbols drawn from the signal constellation and
Cq , Dq are called dispersion matrices. Many of the existing space time codes can be
put in them framework of linear dispersion codes. Q is a design parameter and will
be discussed later. For Alamouti code, Q = 2 and the dispersion matrices are given
12
by
1 0 0 1
C1 = C2 =
0 0 0 0
0 0 0 0
D1 = D2 =
0 −1 1 0
In this subsection we will formulate the equivalent channel for a given LDC matrix.
This formulation converts the input and output matrices to equivalent vectors, which
make the calculation of capacity and decoding of codes easier. We derive an alternate
way of representing the channel which makes simulation of channel easy for any
(M, N ).
Decomposing sq into real and imaginary parts we get
sq = αq + jβq , q = 1, . . . , Q
where Aq = Cq + Dq and Bq = Cq − Dq .
Writing the received matrix Y
r Q
ρ X
Y= H (αq Aq + jβq Bq ) + V (3.2)
M q=1
13
0 0 0 0
where y = [YR† YI† ]† v = [VR† VI† ]† ,s = [ᾱ† β̄ † ]† and the equivalent channel H is
given by
0 0
HR −HI X1R −X2I
H= 0 0
(3.4)
HI HR X1I X2R
where AR , AI denotes the real part and imaginary of matrix A respectively. H is of
dimension 2N T × 2Q.
We observe that the set of Equations (3.3) is not under determined as long as
Q ≤ NT
Also a larger Q implies more degrees of freedom, thereby increasing the capacity
of equivalent channel. A smaller Q implies more repetition and higher diversity or
coding gain. Hence Q is optimally chosen as Q = min (M, N )T , since the maximum
rate can scale only as min (M, N ) log(SN R). Dispersion matrices Aq , Bq are chosen
such that, capacity of equivalent channel approaches actual channel capacity.
The equivalent channel capacity is given by.
1 ρ
Cld (ρ, T, M, N ) = max E log det(I2N T + HH† ) (3.5)
Aq ,Bq 2T M
The power constraints are given by any of the following equations
PQ †
1. q=1 (trAq Aq + trB†q Bq ) = 2T M
TM
2. trA†q Aq = trB†q Bq = Q
,q = 1, . . . , Q
T
3. A†q Aq = B†q Bq = I ,q
Q M
= 1, . . . , Q
1. The capacity achieved by LDC is always less than equal to the channel capacity.
This is because we are imposing a structure on the code, which results in loss
of degrees of freedom.
14
2. Any of the three power constraints given above can be used, but the conditions
become more rigid from one to three. Hence we are effectively reducing the de-
grees of freedom from constraint one to constraint three. This gives a reduction
in capacity, but the SER performance of the code becomes better.
Using constraint three implies, finding Aq , Bq such that they are unitary and also
maximize capacity. This optimization is difficult to achieve for any size of Aq , Bq ,
because of the large number of constraint involved (4QT M non linear constraints).
However this optimization problem can be solved easily for (2, 2) and (2, 1) systems
because of the following general representation for 2 × 2 unitary matrices [17]. X is
unitary if and only if
1 0 a −b
X= , |a|2 + |b|2 = 1, θ ∈ R, a, b ∈ C
iθ
0 e b̄ ā
In this subsection LDC codes for different number of transmitters and recievers will
be presented. These codes were obtained by solving the optimization problem using
15
MATLAB. The code will be listed in terms of Cq , Dq , q = 1, . . . , Q. The code is given
by
Q
X
X= (sq Cq + s∗q Dq )
q=1
16
−0.1236 − 0.3242i −0.0638 + 0.6207i 0.1689 + 0.4010i −0.0090 − 0.2639i
D3 = D4 =
0.0686 − 0.1890i 0.11714 + 0.0337i −0.2364 − 0.3637i −0.02843 − 0.0400i
Codes for other values of (M, N ) are given in Appendix C. In Table 3.1 capacities
obtained by the equivalent channel for these codes are compared to the actual ergodic
capacity of the channel.
LDC achieve capacity, but are not unique. So an exhaustive search and performance
simulation of these codes for SER is required to pick a code having a good SER
17
(M, N ) Channel Capacity (Bits) Ldc Capacity (Bits)
Constraint1 Constraint2 Constraint3
(2, 1) 6.2678 6.221 6.204 5.9508
(2, 2) 11.34 11.36 11.28 11.25
(3, 1) 6.415 6.29 6.24
(3, 3) 16.706 16.6633
(4, 1) 6.473 6.454
performance. Codes based on Frame Theory reduce the search space and provide
with codes that have better error performance. These codes were first proposed by
Heath and Paulraj [15]. In this section we will review codes based on frame theory
and obtain codes for some antenna configurations using numerical methods.
The code X is defined as in Equation (3.1) with Dq = 0. The effect of removing s∗q
on channel capacity is negligible [1]. Writing the equivalent channel .
Q
X
X= s q Cq
q=1
Substituting in the channel model, and taking the vec operation we get
ρ 0
y= H Xs + v
M
where X = [vec(C1 ), . . . , vec(CQ )] and others as previously defined. So the capacity
of equivalent channel is given by
1 ρ †
max E log det(IN T + H0 X X † H0 )
tr(X X †
<T ) T M
18
of a vector space cannot be greater than its dimension. In such a case making X
X † X = IQ (3.6)
would make the capacity grow asymptotically in proportion to Q/T [15]. We observe
that when Q < M T , the matrix X appears tall and is called as a tall matrix. Any
tall matrix satisfying equation (3.6) is called a tight frame. This above condition
guarantees that, the rate of code can scale to the ergodic capacity. To optimize the
error performance the rank of the code word matrices have to be optimized. When
Q = M T , the following criterion ensure that a code achieves channel capacity and
has a good error performance.
1. X X † = IM T
2. For T > M ensure that Cq C†q = Iwhile for T ≤ M ensure that C†q Cq = I for
q = 1, . . . , Q
When Q < M T , make X as tight frame, to achieve high rate. Codes obtained under
this condition should be optimized by considering other properties, i.e. making the
code word error matrix full rank or maximizing the worst case determinant of code
word error matrix.
3.3.2 Examples
The algorithm for constructing tight frames which satisfy both the above criterion
are given in [15] [16]. Using the algorithm we have come up with designs for various
(M, N ). In this section we will present codes for 2 × 2. Other constructions are
provided in Appendix C. The codes are presented in terms of X .
(M, N, Q, T) = (2, 2, 4, 2) : code1
19
0.104103 − 0.233339i −0.162140 + 0.241640i 0.439005 − 0.042893i −0.344111 − 0.192506i
0.354478 + 0.243024i 0.351128 − 0.205011i 0.009591 − 0.235250i −0.190313 − 0.241475i
X =
−0.231071 + 0.362383i −0.402476 − 0.057737i −0.161674 − 0.171161i −0.278022 − 0.131274i
0.187131 + 0.173973i −0.241182 − 0.162820i 0.359317 − 0.255846i 0.264814 + 0.292138i
In this section performance of LDC and LDC based on Frame theory will be compared.
In Figure 3.1 error performance of LDC is compared with Frame based LDC and
uncoded system for two transmitt and two recieve antennas. The code used for LDC is
(2, 2, 1, 4) code, which was given in subsection 3.2.5. Code 1 given in subsection 3.3.2
was used for frame theory based code. Both these codes transmit two symbols per
channel usage. BPSK was used for simulation. At SER of 1e4 LDC performs better
than uncoded system by 3dB and frame theory code perform better by 1dB over
LDC. So LDC codes perform better than spatial multiplexing systems for same rate
and have equal decoding complexity. Frame based codes perform better than LDC
because of their more constrainted design.
In Figure 3.2 the above two codes were simulated for R = 4 bps/Hz using QAM.
We observe that the difference between LDC and Frame based LDC decrease. This
is because we see the effect of diversity is more at lower rates, i.e at higher rates the
probability of encountering bad codewords is less, and hence designing codes based
on more constraints at high rates in not of much use.
In Figure 3.3 code1 and code2 given in subsection 3.3.2 were compared for (2,2)
system at 4bps/Hz. we observe that the performance difference is negligible.
20
BER plots for LDC, Frame based LDC , Uncoded, for (T ,R )=(2,2), Rate =2bps/Hz
x x
0
10
LDC
Uncoded
Frame Theory Based LDC
−1
10
−2
10
BER
−3
10
−4
10
−5
10
−6
10
0 2 4 6 8 10 12 14 16 18
SNR dB
Fig.3.1: Error Performance of LDC, Frame based LDC for (2,2) system at
R = 2bps/Hz
In figure 3.3 LDC codes for (2, 1, 2, 2) are compared for constraint1 and constraint
2 for R = 2bps/Hz. We observe that they have similar BER performance. Hence
we can infer that codes designed from constraint 1 and 2 perform almost similar.
Therefore one can design codes based on constraint 1 which makes optimization of
capacity function easy.
21
Comparison of LDC, Frame Theory based LDC for (T ,R )=(2,2), R=4 bps/Hz
x x
0
10
Frame based LDC
LDC
Uncoded
−1
10
SER
−2
10
−3
10
−4
10
0 2 4 6 8 10 12 14 16 18
SNR dB
Fig.3.2: Error Performance of LDC, Frame based LDC for (2,2) system at
R = 4bps/Hz using QAM
−1
10
SER
−2
10
−3
10
−4
10
0 2 4 6 8 10 12 14 16 18
SNR dB
22
Comparison of Constraint1 and Constraint2 for 2x1, R=2bps/Hz
0
10
Constraint1
Constraint2
−1
10
SER
−2
10
−3
10
0 2 4 6 8 10 12 14 16 18 20
SNR dB
23
Chapter 4
Sphere Decoding
4.1 Introduction
In this chapter we study Sphere Decoding and its complexity when applied to MIMO
systems. We will show that the initial radius of the sphere decoder is a critical
parameter and propose a new initial radius based on the channel knowledge.
MIMO wireless systems are characterized by high capacity, and many space time
codes try to achieve this limit. For many of these codes, the maximum-likelihood
(ML) approach results in a decoder which has a computational complexity that is
exponentially proportional to the number of transmit antennas and symbol constel-
lation size. In general, this problem can be posed as, finding the closest lattice point
in 2N -real dimensions when the point y ∈ CN .
The Sphere Decoding (SD) algorithm was first proposed in [20] to reduce the
ML search space, by confining the search only to the lattice points in a hypersphere
centered at the received point. Sphere Decoding algorithm primarily addresses the
problem of finding the lattice points inside the hypersphere given the lattice, the
received point, and the radius of the sphere.
The channel model that will be considered in this chapter is
r
ρ
Y= HX + V (4.1)
M
24
pρ
the familiar quasi-static and flat fading assumption. The term M
signifies the fact
that the total signal power is equally divided among the M transmit antennas. H
and V are comprised of independent, identically distributed complex-Gaussian entries
C(0, 1). By stacking the real part R(.) and imaginary part I(.) of these entries, we
define
h i†
x= R(X) †
I(X) † ∈ R2M ×1
h i†
y= R(Y) †
I(Y) † ∈ R2N ×1 ,
h i†
v= R(V) †
I(V) † ∈ R2N ×1 ,
y = Hx + v (4.3)
We only consider signal constellations which are subsets of the QAM constellation.
The basic idea behind Sphere Decoding is very simple. We attempt to search only
those lattice points which lie inside a hypersphere centered at the received point rather
than searching over the entire lattice. The closest lattice point inside the hypersphere
is also the closest lattice point in the entire lattice.
The problem of finding lattice points inside a hypersphere is analogous to ML
decoding which is of exponential complexity. Mathematically this translates to solving
1
the inequality k y − Hs k≤ R0 for s ∈ Z n . Sphere Decoding algorithm mainly
addresses the problem of finding the points inside the hypersphere given the radius
1
For this consider H, y, s to be real for illustration.
25
r
y 8
z 11
10 l2
x 7
6 5 4 3 1
2
l1
2R
26
4.2), we first find out the admissible values of x-coordinate. Suppose we have found
that x = {1, 2, 3, 4, 5, 6} satisfy the criterion, we then find out for each point the
admissible values for y-coordinate. For instance, for x = 2, the admissible values of
y-coordinate are {7, 8, 9, . . .}. Then for each of these y-coordinates we find out the
admissible values for z- coordinate. For y = 8, we find that the admissible values of
z are {10, 11}. Hence the points (2, 8, 10), (2, 8, 11) are inside the hypersphere. This
method constructs a tree where the nodes in level k correspond to lattice points inside
the hypersphere of radius r and dimension k. Therefore sphere decoding algorithm is
primarily a combination of depth first search (DFS) algorithm and a method to find
out the interval lengths in each dimension.
To find the interval lengths such that the interval length at dimension k is a
function only of points and interval lengths in lower dimensions , H is factorized and
the upper triangular property of R is used.
R
H = Q
0(2N −2M )×2M
Algorithm:
Input: Q = [Q1 Q2 ], R, x, y = Q1 ∗ x, R0 .
27
3. (Increase sk ) sk = sk + 1. If sk ≤ U B(sk ) go to 5, else to 4.
4.3 Complexity
The complexity of Sphere Decoding algorithm depends on the size of the tree gener-
ated, which is equal to the total number of lattice points in hyperspheres of radius R0
and dimension k = 1, . . . , 2M . The size of the tree depends on the entries of matrix
H. If H is a deterministic matrix, then the worst case algorithmic complexity comes
into picture and is shown (reference 40) to be
2
1 1
q b4R 0 dc + 2M − 1
(2(2M )3 + 3(2M )2 − 5(2M )) + (2b r02 dc + 1) + 1
6 2 2
b4R0 dc
(4.4)
−2 −2
where d = max (r1,1 , . . . , r2M,2M ).We observe that the upper bound on the number
of computations in (4.13) can be exponential. But in practice the received vector is
random due to the channel nature and noise.
y = Hx + v (4.5)
The entries of v are independent N (0, σ 2 /2) and the entries of H are N (0, 1/2). We
see that k v k2 = k y − Hs k2 is a χ2 random variable with 2N degrees of freedom.
Thus we may choose the radius to be a scaled variance of the noise,
in such a way that we may find a point inside the hypersphere with a high probability.
Z αN N −1
λ
e−λ dλ = 1 − (4.7)
0 Γ(m/2)
28
and choose to be close to zero. If we do not find any points inside the hypersphere
we make = 2 , recalculate the radius and rerun the algorithm.
When the channel and noise are random, average complexity rather than the peak
complexity becomes more meaningful. It is shown in [2][3] that the average complexity
becomes cubic at high SNR.
6.5
Complexity exponent e
6
2
Complexity = 4e
5.5 1
4.5
3.5
0 5 10 15 20 25 30
SNR
Figure 4.3 shows how the complexity exponent varies with SNR. We observe that
the average complexity is very high at low SNR and approaches a cubic complexity
for higher SNR. The choice of initial radius based entirely on noise variance, makes
the initial radius large for low SNR. Therefore we encounter large number of lattice
points inside the hypersphere and complexity increases. When the channel matrix
is not well conditioned, i.e when some of the eigenvalues are small, the transformed
lattice tends to collapse in the direction of these eigenmodes and results in very large
number of points inside the sphere. When the channel matrix is bad, there is a high
probability that our decoding will be in error (due to collapsed lattice) along with
high complexity. Therefore it becomes intuitive to choose the initial radius based on
29
both channel and noise.
The Voronoi cell is a convex hull formed by the intersection of hyperplanes. Covering
radius C is the radius of the smallest sphere containing the Voronoi cell. We observe
that C is the best possible initial radius. The received point falls in one of the Voronoi
cells and is at a maximum distance of C from a lattice point. So choosing the initial
radius to be C guarantees at-least one lattice point. The calculation of C is shown
to be a NP-hard problem. Algorithms like Diamond-Cutting algorithm, calculate the
covering radius but in exponential memory and time. So choosing initial radius based
on C is not optimal since we have to calculate C for every channel usage.
Here, α is chosen such that, one can find a lattice point with probability 1 − where
0 ≤ 1 [2] and
αN
λN −1 −λ
Z
1− = e dλ (4.10)
0 Γ(N )
30
where, Γ(.) is the gamma function. This radius R0 is completely based on the noise
variance and hence can lead to large number of points inside the sphere for low SNRs.
Instead, we propose to choose the new initial radius R1 as
D
R1 = min , R0 (4.11)
2
where D is the major diagonal of an individual lattice cell. This radius ensures
that the point is decoded without failure with probability one for an infinite lattice.
Unlike the covering radius of [2], this major diagonal is easy to calculate. Hence the
complexity of sphere decoder with new radius is less than or equal to the conventional
sphere decoder. R0 is a function of . R0 can be made small, which implies a large
epsilon, and a larger decoding failure. Hence we would want to be about 0.9 initially.
For infinite lattices the received point falls in one of the lattice cells and hence
the hypersphere centered around this point with radius R1 will include at least one
lattice point. But in practical situations, where the lattice is finite, the received point
may fall outside the lattice and the hypersphere may not include any lattice point
leading to a decoding failure. In such situations the radius should be updated and the
given point should be decoded. One method of updating would be to shift back to
the radius that depends on noise variance, if R1 fails. This is intuitive since, the point
is outside the lattice and is at a significant distance from any lattice point, implying
a high noise situation. The complexity would be sum of the decoding failure and the
present run. When the point lies outside the lattice, the hypersphere based on R1
would not include many points because of the finite nature of lattice, and hence the
number of nodes in the tree would be low and the complexity of decoding complexity
would be low. In this thesis, this updating procedure is termed as Method A.
If one could get an ideal radius, i.e the radius that would yield only one point
inside the hypersphere, it would lead to the lowest complexity. Intuitively one can
start with zero radius and slowly increase the radius in finite steps until we obtain a
lattice point. But such a procedure would lead to a large number of decoding failures.
The cumulative complexity of these decoding failures would be very high. So instead
31
of starting with zero radius, one could start with a radius that would yield a lattice
point with high probability, and if a failure occurs, the radius is updates in small
increments. The size of the increment is critical for the aggregate complexity. This
updating procedure is termed as Method B.
Given that the edges of the input lattice cell are the vectors ej = {0, 0, . . . , 1, . . . , 0, 0}
where 1 is in j th position and ej ∈ {0, 1}2M , the output lattice cell edges under the
linear transformation H are the columns of H. The major diagonal D is then given
by
X
D = max
a :a=
hi − hj , 0 ≤ j ≤ 2M (4.12)
1≤i≤2M
i6=j
4.4.2 Method A
When the radius based on D/2 fails for a finite lattice, it implies that the point lies
outside the lattice at a distance greater than D/2. It is intuitive then to fall back
to R0 and re-decode the received point. The updating procedure for method A is
therefore to shift back to R0 if R1 fails.
32
Start
k=1
R2(0) = R1
P=0; K=1
Is Yes
Lattice Point Stop
Present ?
No
R2 (k) ≥ R0 R2 (k) = R0
and Yes p=1
P==0
No
R2 (k) = R2 (k − 1) + δR2 (0)
k=k+1
4.4.3 Method B
33
complexity Versus δ for 2x2 and 16QAM
8
Complexity exponent e
7.5
e
Complexity Exponent e 7 C= 4
6.5
4.5
0 0.5 1 1.5 2 2.5 3
δ
From equation (4.12) we see that major diagonal is the norm of a 2N dimensional
vector. Each vector is obtained by adding 2M , 2N dimensional vectors. Therefore
the final vector is a 2N dimensional vector with variance of each term being 2M or
√
equivalently a 2N dimensional vector scaled by 2M . Hence D 2 is a χ square distri-
bution with 2N degrees of freedom. The conventional choice of radius is according
to (4.10). It is easy to see that the probability D 2 ≤ R02 is given by
2αN σ 2
2 2
P (D ≤ R0 ) = P X ≤ √ (4.13)
2M
where X is a Chi-square random variable of dimension 2N . Here P (D 2 ≤ R02 ) de-
creases either with σ 2 , or 1/M or both . Hence, taking D/2 as the new radius gives
an advantage only in the lower SNR range. The probability is also increasing func-
tion of α and from 4.7 we observe that decreases with increasing α. However to
get a better probability of success at the first decoding attempt α is generally large.
However, P (R1 ≤ R0 ) = 1. Hence, the new radius based sphere decoder will always
34
have a complexity that is less than equal to that of the sphere decoding rule in [2].
13
12.5
Complexity exponent e
12
11.5
11
10.5
10
9.5
−4 −2 0 2 4 6 8 10 12 14
Noise Variance
64-ary QAM constellations have been used for simulations. The signal energy at each
receiver is normalized to unity and the initial radius R0 for the conventional SD is
chosen to make = 0.9 [3]. The scaling factor of noise variance i.e 2αN to determine
R0 for 2 × 2 system, for various SNR that give = 0.9 are given in Table 4.2.
The complexity exponent of the average number of lattice points falling inside the
sphere is plotted versus SNR in Figure 4.8 for 2 × 2 MIMO system, using the SD
35
σn2 1 0.8 0.7 0.6 0.5 0.25 0.1 0.05 0.01 0.005 0.001
2αN 2 2 1.8 1.8 1.8 1.6 1.6 1.8 3 3.6 4.2
6.5
Complexity exponent e
6
2 e
Complexity = 4
5.5 1
4.5
3.5
0 5 10 15 20 25 30
SNR
of [2] defined by method A. Observe that the new radius requires only a quarter of
the computations when compared to the conventional SD at lower SNR. The average
number of points is close to the conventional SD curve, for high SNR. This is because,
for finite lattices, there is a large probability that the points may lie outside the lattice
at a distance greater than the major diagonal. At high SNRs, the original radius
becomes smaller than the major diagonal, and therefore both the algorithms tend to
use only R0 . The computational complexity of SD algorithm is also important since
the new radius may fail to decode for finite lattice case. When it happens, we revert
back to the old radius causing a overhead in the algorithmic complexity. In Fig.4.10,
the exponent of the average complexity have been plotted as a function of SNR for
the 2 × 2 system, using method B for updating the radius. Even though the average
number of times the SD has to be repeated for a single received point by method B
36
Average no. of points versus SNR for (2× 2) 64−QAM
5
1. New Radius
2. Old radius
4.5
e
3
Avg. no. points = 4
2
2.5 1
1.5
0.5
0
0 5 10 15 20 25 30
SNR
is more than that of method A, the average complexity for the whole procedure is
significantly lower. The choice of δ is very important. For this simulation, we have
chosen δ = 0.2 and no attempt has been made in this work to optimize the choice of
δ for these simulations.
4.6 Summary
In this chapter we studied Sphere Decoding and understood the importance of initial
radius. We have proposed a new method to choose the initial radius . By incorpo-
rating the new radius in the sphere decoding algorithm, we could achieve significant
reduction in complexity in the lower SNR regime. We also proposed two updating
strategies, for a decoding failure.
We investigated the case of two-transmit and two-receive antenna system in detail
using simulations. While the simulations in this study are mostly limited to 2 × 2
case, this idea works in a similar fashion for higher dimensional case. We have shown
that method B of updating gives a significant reduction in complexity compared to
37
Average no. of points versus SNR for (2×2) 64−QAM
5
1.New radius
2.Old radius
4.5
4
e
Avg. no = 4
3.5
Average no.of points e
3
2
2.5
2 1
1.5
0.5
0
0 5 10 15 20 25 30
SNR
6.5
Complexity Exponent e
2
6 Complexity = 4e
1
5.5
4.5
3.5
0 5 10 15 20 25 30
SNR
38
method A. We also showed the importance of choosing the parameter δ. In conclusion,
this work demonstrates the fact that radius of the sphere decoder and its updating
strategy can significantly impact decoding complexity.
39
Chapter 5
Rate-Diversity Tradeoff
5.1 Introduction
Multiple antennas provide us with high data rates as well as diversity, i.e. robustness
against channel fading. However a fundamental tradeoff exists between both. This
tradeoff was defined and studied by Zheng and Tse [8]. This tradeoff was studied
for Gaussian codes and optimal tradeoff was derived for T ≥ M + N − 1. Yao [9]
proposed tilted QAM codes which achieve the optimal tradeoff for a 2 × 2 system
with T = 2. She proved non-Gaussian codes, which satisfy minimum determinant
criterion, achieve optimal tradeoff for 2 × 2 system. In this chapter, rate-diversity
tradeoff for non Gaussian codes is analyzed and the criterion that a code should follow
in order to achieve the optimal tradeoff will be derived. We will also analyze the error
probability of non-Gaussian codes as a function of singular values of the code.
In this section, we shall introduce the channel model, definition of diversity, multi-
plexing gain and introduce the main result.
We consider a MIMO system with M transmit and N receive antenns. The channel is
given by H where hij ∈ C are independently Rayleigh distributed with unit variance
and zero mean. Channel state information is available only at the receiver and not at
40
the transmitter. We also assume that the channel is quasi-static over T time intervals
(i.e.. channel uses). The input output relation is given by
r
ρ
Y= HX + V (5.1)
M
where X ∈ C M ×T , V is the additive noise with i.i.d entries vij ∈ CN (0, 1). ρ is the
average signal to noise ratio at each receive antenna.
A rate R bps/Hz codebook C has | C |= 2RT codewords {X(1), . . . , X(| C |)} each of
which is a M × T matrix. The transmitted signal X is normalized to have the average
transmit power at each antenna in each symbol period to be 1.
5.2.2 Definitions
The channel capacity of an (M, N ) MIMO system grows as C ≈ min {M, N } log(SN R).
Therefore, at any SNR, to achieve a significant fraction of capacity, the rate of the
code should also increase as log(SN R). If the rate of the transmitted code is fixed
for all SNR, the gap between capacity and transmission rate increases with SNR and
becomes infinity as SN R → ∞. Hence, to achieve a significant fraction of the capac-
ity for all SNR, the rate of the transmit code should also increase, i.e a different rate
codebook C for each SN R.
In this contexti, it is better to visualize the input transmit matrix X not as a symbolic
function of input symbols, but rather a one-one function between input information
set and the code word matrix set. The size of the input information set varies with
SN R.
We define a scheme as a family of codes {C(SN R)} of block length T , one at each
SN R level [8].
A scheme {C(SN R)} is said to achieve spatial multiplexing gain r if,
R(SN R)
lim =r (5.2)
SN R→∞ log(SN R)
41
The scheme achieves a diversity d if,
The tradeoff curve between spatial multiplexing gain r and the optimal diversity
at r denoted by d∗ (r) is given by the piecewise linear function connecting points
(r, d∗ (r)), k = 0, 1, . . . , min{M, N }, where,
when T ≥ M + N − 1.
The optimal tradeoff curve is obtained by considering the error probability and show-
ing that it is of the same order as that of outage probability for Gaussian codes
when T ≥ M + N − 1. In this section, error mechanism of MIMO channels will be
considered. By considering the outage probability Pout as a function of R2 we get
42
(0,mn)
(2,(m−2)(n−2))
(r,(m−r)(n−r))
(minm,n,0)
N
Y
= P [ (1 + ρλi ) < ρr ]
i=1
So the outage probability at rate r is approximately equal to the probability that there
are more than N − r singular values, each less than ρ−1 at high SNR, or equivalently
outage probability at r is equal to the probability that the rank of the channel matrix
is less than r. It can be shown that this probability is equal to ρ−(M −r)(N −r) . So the
outage probability is equal to
Outage probability at a given rate and SNR can be shown to be a lower bound to the
actual error probability (not the pairwise error [8] probability) i.e
Also
Pe (ρ) = Pout (r)P (error | outage) + P (error, no outage)
43
Hence if P (error, no outage) is of the same order of P (outage) then by equation(5.5)
and equation(5.14) we get
P (error) ≈ Pout (r)
For Gaussian codes, when T ≥ M + N − 1, both the probabilities are of same order
and we obtain optimal tradeoff.
Orthogonal designs provide us full diversity [10]. In this section the tradeoff curve for
orthogonal designs will be computed for two transmit antenna [8]. For two transmit
antenna Alamouti scheme will be considered. The equivalent received signal at the
receiver is given by r
ρ k H k2F
yi = xi + vi , i = 1, 2 (5.7)
2
where k H kF is the Frobenius norm of H. k H k2F is chi square distributed with
2N M degrees of freedom. Therefore, for arbitrarily small ,
P (k H k2F ≤ ) ≈ M N
We can show that the probability of error is also bounded by outage probability.
Hence to compute the tradeoff curve it is sufficient to evaluate the outage probability.
ρ k H k2F
r
Pout (r) = P 1 + ≤ρ
2
+
= P k H k2F ≤ ρ−(1−r)
+
= ρ−M N (1−r)
Therefore the optimal tradeoff curve for alamouti is given by M N (1 − r)+ . Tradeoff
achieved by Alamouti scheme is compared to optimal tradeoff in Figure 5.2.
44
Comparison of Alamouti and Optimal Rate−Diversity Tradeoff M=N=T=2
Alamouti
Optimal Tradeoff
3
Diversity
0
0 1 2 3
Multiplexing Rate r
45
as follows.
x11 x12
X =
x21 x22
x11 cos θ1 − sin θ1 s11
=
x22 sin θ1 cos θ1 s22
x21 cos θ2 − sin θ2 s21
=
x12 sin θ2 cos θ2 s12
θ1 and θ2 are chosen so as to maximize the worst case determinant of X. (2θ1 , 2θ2 ) =
(arctan(1/2), arctan(2)) maximize the determinant over the complete integer lattice,
and
1
min | det(X) |≥ √ (5.8)
∀X 2 5
for QAM like constellations of all sizes. When the channel is not in outage, this code
can be shown to achieve optimal tradeoff [9] over all QAM due to its minimum de-
terminant property. When not in outage the minimum distance between the received
codewords can be shown to be greater than half the noise variance. This implies that
probability of error when not in outage is in the same order that of outage leading
to optimal tradeoff even when T < M + N − 1. Therefore any code which satisfies
minimum determinant criterion achieves optimality for a 2 × 2 system.
5.3.1 Criterion for a Non Gaussian code to achieve optimal tradeoff for
(M, M ) MIMO system
In this section we shall derive the criterion required to achieve the optimal tradeoff
curve for a general (M, M ) antenna configuration. Also, we assume that T ≥ M .
This is a valid assumption as the number of signaling intervals should be greater than
number of transmit antennas to achieve full diversity [10]. Consider a m2 -QAM
.
constellation from Z(i) to transmit at a rate R = r log 2 (ρ) such that m2 = 2R/M =
.
ρr/M , where = denotes asymptotic sense with respect to ρ, i.e. any value which is not
46
a function of ρ can be considered to be 1, since it does not grow with ρ. Therefore
any constant multiplying ρ can be absorbed into ρ.
˙ 2 = 2R/M = ρr/M
Es =m (5.9)
Es . r/M −1
σv2 = =ρ (5.10)
ρ
min{k HX k2 }>σ
˙ v2 (5.11)
∀X
(5.11) and (5.13) we can observe that (5.11) is equivalent to finding a condition on
ci such that
X
min ci λ2i >ρ
˙ r/M −1 (5.14)
X
47
Equation (5.12) is symmetric with respect to λi ∀ i. Hence there should exist a
minimum distance from the origin, i.e. min( λ2i ) exists and it occurs when λ1 =
P
When ρ → ∞ we get
X
min πj2 > 1
X
j
The above inequality says that the instantaneous energy (not the average power) of
the input matrix X should be greater than 1 at high SNR. In the above derivation, an
P 2
assumption that the worst case of channel occurs when λi is minimum was made,
but the worst case may occur at other combinations of λi . We will show that the
above condition equation (5.17) is a sufficient condition to achieve optimality. When
equation(5.12) holds true then
X
λ2i > M (ρr/M −1 − ρ−1 ) (5.18)
i
48
10002 (x y)2 + 1000 (x2+y2)+1 − 10001.5 = 0
2
0.5
d
min
0 d>dmin
y
−0.5
−1
Non Outage Region
−1.5
x=y
−2
−2 −1.5 −1 −0.5 0 0.5 1 1.5 2
x
since any distance from origin is greater than when λi are equal Figure 5.3.
Also For eq5.14 to hold true
X X
min ci λ2i >
˙ min min {ci } λ2i > ρr/M −1
X X
ρr/M −1
⇒ ˙
min min{ci } > (5.19)
X M (ρr/M −1 − ρ−1 )
We know that, min{ci } > min{πi2 }
49
channel is exactly out of outage, i.e on the boundary point. Depending on how much
deep we are into the non outage region the codeword difference will be much more
greater than the noise variance. In Appendix A we have shown that this condition
makes the code behave like T → ∞.
P 2
In the above derivation we got the condition that πi > 1 for all QAM, but a
P 2
condition like πi > C for all QAM where C is some constant is sufficient because
all the above results are in asymptotic sense and any C which is not function of ρ
is equal to 1 in the above sense. In the above results the point r = 0 will cause the
bound to go to infinity. This is because of the following reason. In analysis of rate
diversity tradeoff the absolute rate R = r log(ρ) which becomes zero at r = 0. This
implies that there is only one matrix in the code book. To avoid this problem one
can define the rate R = r log(ρ) + log(). We observe that the contribution by log()
term is constant and becomes negligible at higher SNR. The terms ρr/M −1 becomes
ρr/M −1 1/M in the above results, and the singularity at r = 0 is removed.
P 2
The codes designed using the criterion πi ≥ 1 come close to achieving the
optimal tradeoff curve, but they will tend to perform better than the conventional
codes, since we are reducing the error probability when not in outage. Summarizing
the results of this section
A Linear Space Time code X satisfying the following conditions
50
5.3.2 Examples of Codes that achieve optimal Tradeoff
In this subsection we shall analyze some existing codes, for their tradeoff character-
istics
Tilted QAM Codes: Yao [9] showed that Tilted QAM codes achieve optimal
P 2
tradeoff curve. In this subsection we will show that this code satisfies πi ≥ 1 for all
QAM. Tilted QAM Codes were shown to have a determinant greater than a constant
for all QAM. Tilted QAM code is given by
cos(θ1 )s1 − sin(θ1 )s2 sin(θ2 )s3 + cos(θ2 )s4
X= (5.20)
cos(θ2 )s3 − sin(θ2 )s4 sin(θ1 )s1 + cos(θ1 )s2
where (θ1 , θ2 ) = (0.5 arctan(2), 0.5 arctan(0.5)) are the optimal angles to maximize
P 2
the determinant. Since πi =k X k2 , Frobenius norm of the code matrix can be
considered. Frobenius norm of the above matrix can be found out to be
we observe that for any QAM constellation k X k is greater than 1 since, k X k6= 0
and should be an integer for si ∈ Z(i). We observe that the above code satisfies the
P 2
condition ˙ and det(X) > 0 for all QAM.
πi >1
Linear Threaded Algebraic Space-Time Codes: TAST codes [19] were pro-
posed by Damen and were later extended to LTAST codes. Damen claims that LTAST
codes achieve optimal tradeoff curve, which were verified by simulations, but did not
have a theoretical proof. In this section we will show that these coded satisfy the
determinant and optimality criterion and hence achieve optimal tradeoff. A generic
51
TAST code is defined as follows
1
u11 . . . φ M u21
1 2
φ u22 . . . φ u32
M M
DM,M,M (u) = .. .. ..
(5.21)
. . .
M −1
φ M uM M . . . u1M
where φ is chosen to guarantee full diversity and optimize the coding gain. φ is gen-
erally chosen as follows
φ = eiλ , λ 6= 0 ∈ R
ui = M i s i
since |φc | = 1. Also it is shown by Damen [18] that det(DM,M,1 ) > 0 for all QAM.
Hence as this code satisfies all the required conditions this code should achieve optimal
tradeoff
In this section, the error performance of non-Gaussian codes with regard to rate-
diversity tradeoff will be analyzed. In [10] error analysis was made when the code
is fixed, and in [8] tradeoff curve was established assuming gaussian codes. In this
section we assume that the code is non-Gaussian and is varying with ρ, and derive
the tradeoff criterion.
52
When Gaussian codes are used and T ≥ M + N − 1 the tradeoff curve is obtained.
In Gaussian codes as the individual entities can be anything the properties of codes
depend on T which is the only parameter. In non gaussian codes or codes from finite
alphabet, the properties of codes can be characterized not only by T but also param-
eters like worst case minimum distance between the codes, minimum determinant.
These all are functions of the discrete singular values of the code.All the results in
this section assume N = M .
We derive the error probability as a function of the eigenvalues of HH† and
XX† . Consider X1 transmitted, and the channel realization is H. Let HH† =
UΣU† , Σ = diag(λ1 , . . . , λM ) denote the eigenvalue decomposition of HH† , and
XX† = VΠV† , Π = diag(π1 , . . . , πM ) denote the eigenvalue decomposition of XX† .
Then the probability that X1 is wrongly decoded as X2 is
As before if we assume that the code is linear and input alphabet belong to QAM,
then X1 − X2 = X .
det E(A, B)
Z
†
dΘeT rΘAΘ B δ(ΘΘ† − Im ) = Γ(m + 1) . . . Γ(1) (5.23)
det V (A) det V (B)
53
Where
e a1 b 1 . . . e a1 b m
. .. ..
E(A, B) = .. . .
e am b 1 . . . e am b m
and v(A), V (B) are the vandermonde matrices of A, B respectively. Using the fact
that the probability density function of unitary matrices of order m is
Γ(m) . . . Γ(1)
f (Θ) = δ(ΘΘ† − Im )
π m(m+1)/2
where Γ() denotes the gamma function. Averaging the error probability gives us
−ρλ1 π1 −ρλ1 πM
e ... e
.. .. ..
C
P [X(1) → X(2)|Σ] ≤ det . . .
Q Q
i<j (ρπi − ρπj ) i<j (λi − λj )
e−ρλM π1 . . . e−ρλM πM
1 1
1+ρλ1 π1
... 1+ρλ1 πM
.. .. ..
C
≤ det . . .
Q Q
i<j (ρπi − ρπj ) i<j (λi − λj )
1 1
1+ρλM π1
... 1+ρλM πM
(5.24)
1
where C is a constant. we make the approximation e−x ≤ 1+x
, ∀x ≥ 0 .
1 1
1+ρλ1 π1
... 1+ρλ1 πM Q (ρπi − ρπj ) Q (λi − λj )
.. .. ..
i<j i<j
det . . . =
QM
i,j=1 (1 + ρλi πj )
1 1
1+ρλM π1
... 1+ρλM πM
1
P [X(1) → X(2)|Σ] ≤ QM (5.25)
i,j=1 (1 + ρλi πj )
54
1
The above approximation can be made only for higher SNR, since e−αx ≤ 1+αx
, ∀x, α ≥
1 1
0 but e−αx − e−αy 1+αx
− 1+αy
∀α ≥ 0 but can be made true as α → ∞. The
two bounds given by eq 5.24 and eq 5.25 have been averaged over H and plotted. We
Comparison of Upper bound and its Approximation averaged over H
0
10
Exponential Approximation
Upper Bound
−1
10
−2
10
Probability of Error
−3
10
−4
10
−5
10
−6
10
−7
10
0 5 10 15 20 25 30 35 40
SNR
observe that the bound becomes correct for SNR over 10dB. We also observe that
the slope is preserved, which is of more importance for tradeoff analysis. Instead of
averaging over the code ensemble, the values of πi which give rise to the worst deter-
minant will be considered. This can be justified by removing the 1 in deneominator
and observing that denominator is a function of determinant. After applying union
bound we get
ρT r
P [error|Σ] ≤ QM
i,j=1 (1 + ρλi πj )
For a code to achieve optimal tradeoff, the probability of error when not in outage
should be of the same order of outage probability. So to obtain a criterion for achieving
optimal error probability we average it over the event of non outage (A)c . We define
λi = ρ−αi as defined in [8]. We also define πi = ρ−βi since πi are no longer constant
55
but vary with size of codebook which varies with ρ. We then get
Using the method followed in [8], the error probability when not in outage is
Z
p(error|no outage) ≤ ρ−d(r,α,β) dα
(A)c
where !
M
X m X
X m
d(r, α, β) = (2i − 1)αi + (1 − αi − βj )+ − T r (5.26)
i=1 i=i j=1
and
X
A = {αi ∈ R+ , α1 ≥ α2 ≥ . . . ≥ αm ≥ 0, and (1 − αi )+ ≤ r} (5.27)
56
r dmin (r, α, β) β1 β2
0 4 -1 1
1 1 0 0
2 0 1 -1
Table 5.3: Values of βi for optimal rate diversity tradeoff for 2 × 2 system,
T =2
r dmin (r, α, β) β1 β2 β3
0 10 0 0 -1
1 4 0 0 0
2 1 0 0 0
3 0 0 0 0
Table 5.4: Values of βi for optimal rate diversity tradeoff for 3 × 3 system,
T =3
r dmin (r, α, β) β1 β2 β3 β4
0 16 -1 0 0 0
1 10 0 1 -1 0
2 4 0 0 0 0
3 1 0 0 0 0
4 0 0 0 1 -1
Table 5.5: Values of βi for optimal rate diversity tradeoff for 4 × 4 system,
T =4
57
√
r=0: β1 + β2 ≤ −0.5 ⇒ π 1 π2 ≥ ρ
r = 1, 2 : β 1 + β2 ≤ 0 ⇒ π 1 π2 ≥ 1
Except for the point r = 0 we have obtained the condition given by yao [9] for
optimality of 2 × 2 system. At r = 0 we obtain the above result saying that det(X) ≥
√
ρ because of the following reason. In rate diversity tradeoff r = 0 ⇒ R = 0, which
theoretically means there is only one or no matrix to transmit,i.e size of the codebook
is 1. Hence the above result is very valid since we can put the total power in the
single matrix. Intuitively r = 0 corresponds to any constant rate transmission for
all ρ. Since we already know that such a constant rate code achieves full diversity
when det(X) > 0 , we can impose the same condition here for r = 0. This condition
combined with the above conditions for other r we get that any 2 × 2 code achieves
optimal tradeoff if
det(X) ≥ 1 ∀X (5.28)
5.4 Full rate and Full Diversity codes from Extension fields
58
5.4.1 Code Construction
Let s denote the information vector of size q × 1 and fi (s) denote a linear function of
information symbols. Then the input code matrix is defined as
X = U ΣV † (5.29)
f1 (s) and f2 (s) are chosen over the appropriate extension field so that they are non
zero and linear over the gaussian integer ring. See Appendix 1 for information about
extension fields.
√
For example, in a 2 × 2 system we can choose f1 (s) = s1 + 2s2 and f2 (s) =
√
s1 − 2s2 . Choosing U and V as 2 × 2 Hadamard matrices, we get the code X as
√
s 2s2
√1 (5.31)
2s2 s1
We observe that X is linear and has a determinant s21 − 2s22 which is non zero for
√
si ∈ Q(i) . Here we have chosen Q( 2) as the extension field. we see that if
s1 , s2 ∈ Z(i) then f1 (s), f2 (s) 6= 0. we also observe that if f1 (s) and f2 (s) are linear
combination of the complete information vector as in the previous example then
0
det (X) 6= 0 ⇒ det (X − X ) 6= 0, where X0 is onother code matrix
fi (s) is chosen as the dot product of the information vector and the basis elements
of an algebraic extension field over Q(i) whose degree over Q is q. i.e.
59
is important and has to be chosen so that the approximation error of integers by the
basis elements is large, yielding a better product distance. The rate of the above
q
code is given by R = T
log2 (|S|) where S is the constellation set. These codes can be
viewed as a more general form of Diagonal Algebriac Space Time Codes (DAST) [11].
In DAST codes instead of linear functionals rotations of information symbols were
used. The rotation was being chosen to maximize product distance. Also matrix U
is always chosen as a Hadamard matrix, while matrix V is chosen to be an identity
M
matrix. The rate of DAST codes is specified is R = M
log2 (|S|).
In this subsection we will study the effect of U, V on ergodic capacity, outage ca-
pacity and BER performance. We will also study the effect of extension field on the
performance of the code by simulation. We will also compare the performance of the
code with alamouti scheme and DAST codes. For simulation BPSK was used with
the code given by eq 5.31. In Figure 5.5 ,the proposed method was compared with
Alamouti and DAST codes for a 2 × 1 system with R = 1bps/Hz. We observe that
the new code has a diversity 2. The new code performs worse than both Alamouti
and DAST because the minimum determinant of the new code is smaller than the
other.
60
BER Performance for 2×1 system, R =1b/s/Hz
0
10
Alamouti Code
DAST Code
Proposed Code
−1
10
−2
10
BER
−3
10
−4
10
−5
10
−6
10
0 5 10 15 20 25 30
SNR dB
can observe from Figure 5.6 that the code from Q(21/4 ) has a diversity of 1 while the
other has a diversity 4
Although Σ has been specified, the choice of U, V are not specified. So the code
was simulated with different unitary matrices for capacity and SER performance,
with a view to choose the best possible rotation. The code was simulated for capacity
√ √
with f1 (s) = s1 + 2s2 and f2 (s) = s1 − 2s2 . In one case Hadamard matrices were
used for rotation. In the second case an arbitrary rotation matrix was chosen. The
matrix chosen was
−0.72388 −0.6899
U=V=
−0.6899 0.72388
Figure 5.6 show the ergodic capacity as a function of SNR for both the cases. We
observe that the choice of rotation matrix is immaterial. Both the rotations give
exactly the same ergodic capacity. Figure 5.7 shows SER performance of the code for
the above two rotations using BPSK for a 2 × 2 system. We observe that both the
rotations perform identical. So from the above two simulations, one can infer that
the choice of rotation matrix is not vital.
61
Comparison of New Method with different extension fields
0
10
jπ/4
Extension Field Q(e )
1/4
Extension Field Q(2 )
−1
10
−2
10
SER
−3
10
−4
10
−5
10
4 6 8 10 12 14 16 18
SNR dB
5.5 Summary
In this chapter, we studied rate-diversity tradeoff and Tilted QAM codes in detail.
We proposed new criterion on the singular values of the code matrix, so as to achieve
optimal tradeoff for non Gaussian codes for any (M, M ). We also proved that tilted
QAM codes, and TAST codes which achieve optimal tradeoff, indeed satisfy the
criterion. We also analyzed the error probability of non Gaussian codes. we obtained
a upper bound on the error probability as function of singular values of the channel
and the code matrix. We then used this result in rate-diversity framework to obtain
the exact condition on singular values of code matrix, to obtain the optimal tradeoff.
The results that are obtained from this framework match the criterion given by Yao [9]
for a 2 × 2 system. We also obtained a similar condition for a 3 × 3 system. We
proposed codes which obtain full diversity using extension fields. These codes can be
tought as a generalization of DAST codes.
62
Capacity of 2x2 for proposed scheme for various rotations
10
7
Capacity bps/Hz
2
Hadamard Rotation
Arbit Rotation
1
0 5 10 15 20 25 30
SNR dB
SER performance with Hadamard and Arbitary Rotation for 2x2 System
0
10
Hadamard Rotation
Arbit Rotation
−1
10
−2
10
BER
−3
10
−4
10
−5
10
−6
10
0 5 10 15 20 25 30
SNR
63
Appendix A
In this appendix, we summarize important results from the extension fields which are
required.
where bi ∈ F .
√
Example Consider the set of rationals Q. Suppose we want to include 2 and
still keep the field structure, Then we should include all the elements of the form
√ √
a + b 2, a, b ∈ Q. 2 is algebraic integer since it is the root of x2 − 2. This new
√
field is represented by Q( 2). Every element of this extension field is of the form
√
a + b 2, a, b ∈ Q. Hence this can be also thought as a vector field with basis vectors
√ √
{1, 2} and the dimension of this field [Q( 2) : Q] = 2. Also we observe that, the
element 0 is generated only when a = b = 0.
√ √
Example Consider Q( 3). The basis vectors are {1, 3}, and the dimension of the
field 2.
√ √
Example Consider Q( 2, 3) , by the above theorem the basis of this vector space
√ √ √
over Q is {1, 2, 3, 6}, and the dimension 4.
iπ
Example Let φ = e 4 ∈ C. Then φ is a root of the irreducible polynomial µφ (X) =
X 4 + 1 ∈ Q[X]. The degree of φ over Q is 4 and one has φ4 = −1 ∈ Q. The minimum
polynomial of φ over Q[iX] is given by µφ (X) = X 2 − i, and φ has a degree of 2 over
Q(i).
65
Appendix B
B.1 Proof
ML decoder can gaurantee to decode if the magnitude of noise is smaller than half
the minimum distance between codewords. In this appendix we will show that the
P 2
condition πi ≥ 1 is indeed sufficient to achieve optimality. We generalize the
method given in Yao [9] for the proof.
Proof:
If the mutual information is 2α times the desired rate, then
QM 2
i=1 (1 + ρλi )
=α≥1
ρr
If i π 2 ≥ 1 , Then it can be shown by exactly following the inverse of the proof that
P
(1−αi )+ −r
α=ρ
˙
1
(1−αi )+ −r)
˙ M(
⇒ τ =ρ
k V k2
P (eror|H) ≤ P ( > τ /2)
σv2
Z ∞
=
˙ u2M T /2−1 e−u du
u=τ /2
= Γ(M T − 1, τ /2)
where Γ(n, x) is a incomplete function of n−dimensions . Approximating the gamma
function
˙ M T −1 e−τ /2
p(error|H)=τ (B.1)
1
(1−αi )+ −r)
and τ > ρ M ( , when channel is not in outage. In Gaussain code of length
T Tse [8] showed that when channel is not in outage
˙ −T ( (1−αi )+ −r)
p(error|H)≤ρ
67
Appendix C
69
(M, N, T, Q) = (3, 3, 3, 9), Power constraint 1 :
70
−0.0159 − 0.2227i 0.0387 + 0.0416i −0.2616 − 0.0507i
D1 = −0.0888 + 0.2190i 0.1439 + 0.0226i −0.0538 − 0.1059i
−0.0286 + 0.0060i −0.0977 + 0.1141i 0.0913 − 0.1135i
71
(M, N, T, Q) = (4, 1, 4, 4), Power constraint 1 :
72
C.2 Frame theory based LDC for (3, 3) and (4, 1) systems
73
(M, N, T, Q) = (4, 1, 4, 4)
74
Bibliography
[1] B. Hassibi and B.Hochwald, “High-rate codes that are linear in space and time,“
IEEE Trans. Info Theory, vol.48, pp.1804-1824, July 2002.
[2] Babak Hassibi and Haris Vikalo, “On Sphere Decoding Algorithm. 1. Expected
Complexity”.
[3] Babak Hassibi and Haris Vikalo, “On Sphere Decoding Algorithm. 2. General-
izations, Second-order Statistics, and Applications to Communications“.
[4] Haris Vikalo, “Sphere Decoding Algorithms for Digital Communications“, PhD
thesis, Stanford University, 2003.
[6] J.Conway and N.Sloane, “Sphere Packings, Lattices and Graphs“. Springer-
verlag, 1993.
[8] Lizhong Zheng and David N.C. Tse, ”Diversity and Multiplexing: A Fundamen-
tal Tradeoff in Multiple Antenna Channels”
[9] Huan Yao, “Efficient Signal, Code, and Receiver Designs for MIMO Communi-
cation Systems“, PhD thesis, Massachusetts Institute of Technology, 2003 .
75
[10] V. Tarokh, N. Seshadri, and A.Caladerbank, “Space-time codes for high data
rate wireless communications: Performance criterion and code construction,“
IEEE Trans. Info. Theory, vol. 44, pp.744-65, 1998
[17] Kenneth Hoffman, Ray Kunze. “Linear Algebra“, second edition, Prentice Hall
of India.
[18] M.O. Damen, H. EL Gammal and N.C. Beaulieu,“ Linear Threaded Algebriac
Space-Time Constellations“, IEEE Trans. Inform. Theory, vol.49, pp. 2372-2388,
Oct. 2003.
76
[20] U. Fincke and M. Pohst,”Improved methods for calculating vectors of short
length in a lattice, including a complexity analysis,” Mathematics of Compu-
tation, Vol. 44, pp.463-471, April 1985.
77