Добавил:
Опубликованный материал нарушает ваши авторские права? Сообщите нам.
Вуз: Предмет: Файл:

Advanced Wireless Networks - 4G Technologies

.pdf
Скачиваний:
47
Добавлен:
17.08.2013
Размер:
21.94 Mб
Скачать

778 NETWORK INFORMATION THEORY

each channel, we shall transmit a time-parameterized stream of symbols. In other words, the channel will be time-slotted. As a consequence, the operation of coding at a node will be time-slotted as well.

19.5.4 Memoryless communication network

Given a communication network (G, S) and a positive integer τ , the associated memoryless communication network denoted as (G(τ ), S) is defined as follows. The set of nodes in G(τ ) includes the node S and all the pairs of the type [X, t], where X is a nonsource node in G and t ranges through integers 1 – τ . The channels in the network (G(τ ), S) belong to one of the three types listed below. For any nonsource nodes X and Y in (G, S):

(1)for t τ , the multiplicity of the channel from S to [X, t] is the same as that of the channel S X in the network (G, S);

(2)for t < τ , the multiplicity of the channel from [X, t] to [Y, t + 1] is the same as that of the channel XY in the network (G, S);

(3) for t < τ , the multiplicity of the channel from [X, t] to [X, τ ] is equal to max f lowG (X) = m f G(X).

Lemma 2

The memoryless communication network (G(τ ), S) is acyclic.

Lemma 3

There exists a fixed number ε, independent of τ , such that for all nonsource nodes X in (G, S), the maximum volume of a flow from S to the node [X, τ ] in (G(τ ), S) is at least τ ε times mfG(X). For proof see Li et al. [47].

Transmission of data symbols over the network (G(τ ), ryless’ transmission of data streams over the network (G,

S) may be interpreted as ‘memo- S) as follows:

(1)A symbol sent from S to [X, t] in (G(τ ), S) corresponds to the symbol sent on the channel S X in (G, S) during the time slot t.

(2)A symbol sent from [X, t] to [Y, t + 1] in (G(τ ), S) corresponds to the symbol sent on the channel X Y in (G, S) during the time slot t + 1. This symbol is a linear combination of symbols received by X during the time slot t and is unrelated to symbols received earlier by X.

(3)The channels from [X, t] to [X, t] for t < τ signify the accumulation of received information by the node X in (G, S) over time.

Since this is an acyclic network, the LCM on the network (G(τ ), S) can be physically realized in the way mentioned above. The physical realization can then be interpreted as a memoryless transmission of data streams over the original network (G, S).

NETWORK CODING

779

19.5.5 Network with memory

In this case we have to slightly modify the associated acyclic network.

Definition 1

Given a communication network (G, S) and a positive integer τ , the associated communication network with memory, denoted as (G[τ ], S), is defined as follows. The set of nodes in G[τ ] includes the node S and all pairs of the type [X, t], where X is a nonsource node in G and t ranges through integers 1 to τ . Channels in the network (G[τ ], S) belong to one of the three types listed below. For any nonsource nodes X and Y in (G, S);

(1)for t τ , the multiplicity of the channel from S to [X, t] is the same as that of the channel S X in the network (G, S);

(2)for t < τ , the multiplicity of the channel from [X, t] to [Y, t + 1] is the same as that of the channel X Y in the network (G, S);

(3) for t < τ , the multiplicity of channels from [X, t] to [X, t + 1] is equal to t ×

mf G(X).

(4)The communication network (G[τ ], S) is acyclic.

(5)Every flow from the source to the node X in the network (G(τ ), S) corresponds to

aflow with the same volume from the source to the node [X, t] in the network (G[τ ], S).

(6)Every LCM v on the network (G(τ ), S) corresponds to an LCM u on the network (G[τ ], S) such that for all nodes X in G: dim[u([X, τ ])] = dim[v(X)].

19.5.6 Construction of a generic LCM on an acyclic network

Let the nodes in the acyclic network be sequentially indexed as X0 = S, X1, X2, . . . , Xn such that every channel is from a smaller indexed node to a larger indexed node. The following procedure constructs an LCM by assigning a vector v(XY ) to each channel XY , one channel at a time.

{

for all channels XY

v(X Y ) = the zero vector; // initialization for ( j = 0; j n; j + +)

{

arrange all outgoing channels X j Y from X j in an arbitrary order;

take one outgoing channel from X j at a time

{

let the channel taken be X j Y ;

choose a vector w in the space v(X j ) such that w v(U Z ) : U Z ξ for any collection ξ of at most d 1 channels with

780 NETWORK INFORMATION THEORY

v(X j ) v(U Z ) : U Z ξ ; v(X j Y ) = w;

}

v(X j+1) = the linear span by vectors

v(X X j+1) on all incoming channels X X j+1 to X j+1;

}

}

The essence of the above procedure is to construct the generic LCM iteratively and make sure that in each step the partially constructed LCM is generic.

19.5.7 Time-invariant LCM and heuristic construction

In order to handle delays, we can use an element a(z) of F[(z)] to represent the z-transform of a stream of symbols a0, a1, a2, . . . , at , . . . that are sent on a channel, one symbol at a time. The formal variable z is interpreted as a unit-time shift. In particular, the vector assigned to an outgoing channel from a node is z times a linear combination of the vectors assigned to incoming channels to the same node. Hence a TILCM is completely determined by the vectors that it assigns to channels. On the communication network illustrated in Figure 19.18, define the TILCM v as

v(S X) = (1 0)T, v(SY ) = (0 1)T, v(X Y ) = (z z3)T v(Y W ) = (z2 z)T and v(W X) = (z3 z2)

By formal transformation we have for example

v(X Y ) = (z z3) = z(1 z3)(1 0)T + z(z3 z2)T

=z[(1 z3)v(S X) + v(W X)]

Thus, v(X Y ) is equal to z times the linear combination of v(S X) and v(W X) with coefficients 1 z3 and 1, respectively. This specifies an encoding process for the channel X Y that does not change with time. It can be seen that the same is true for the encoding process of every other channel in the network. This explains the terminology ‘time-invariant’ for an LCM.

To obtain further insight into the physical process write the information vector as [a(z) b(z)], where

a(z) =

a j z j and b(z) =

b j z j

 

j0

j0

belong to F[(z)]. The product of the information (row) vector with the (column) vector assigned to that channel represents the data stream transmitted over a channel

[a(z) b(z)] · v(S X) = [a(z) b(z)] · (1 0)T

 

= a(z) (a0, a1, a2, a3, a4, a5, . . . , at , . . .)[a(z)

b(z)] · v(SY )

= b(z) (b0, b1, b2, b3, b4, b5, . . . , bt , . . .)[a(z)

b(z)] · v(XY )

= za(z) + z3b(z) (0, a0, a1, a2 + b0, a3 + b1, a4 + b2, . . . , at1 + bt3, . . .)[a(z) b(z)] · v(Y W )

= z2a(z) + zb(z) (0, b0, a0 + b1, a1 + b2, a2 + b3, a3 + b4, . . . ,

NETWORK CODING

781

at2 + bt1, . . .)[a(z) b(z)] · v(W X)

=z3a(z) + z2b(z) (0, 0, b0, a0 + b1, a1 + b2, a2 + b3, . . . , at3 + bt2, . . .)

Adopt the convention that at = bt = 0 for all t < 0. Then the data symbol flowing over the channel X Y , for example, at the time slot t is at1 + bt3 for all t 0. If infinite loops are allowed, then the previous definition of TILCM v is modified as follows:

v(S X) = (1 0)T , v(SY ) = (0 1)T

v(XY ) = (1 z3)1(z z3)T, v(Y W ) = (1 z3)1(z2 z)T

and

v(W X) = (1 z3)1(z3 z2)

The data stream transmitted over the channel X Y , for instance, is represented by

a j z j b j z j · v(X Y ) = (a j z j+1 + b j z j+3)/(1 z3)

j0

j0

j0

 

=

a j z j+1 + b j z j+3

z3i =

a j z3i+ j+1 + b j z3i+ j+3

j0

i0

j0 j0

=

zt (at1 + at4 + at7 + · · · + bt3 + bt6 + bt9 + · · ·)

t1

 

 

 

That is, the data symbol at1 + at4 + at7 + · · · + bt3 + bt6 + bt9 + · · · is sent on the channel X Y at the time slot t. This TILCM v, besides being time invariant in nature, is a ‘memoryless’ one because the following linear equations allows an encoding mechanism that requires no memory:

v(X Y ) = zv(S X) + zv(W X)

 

v(Y W ) = zv(SY ) + zv(XY )

(19.60)

and

v(W X) = zv(Y W )

19.5.7.1 TILCM construction

There are potentially various ways to define a generic TILCM and, to establish desirable dimensions of the module assigned to every node. In this section we present a heuristic construction procedure based on graph-theoretical block decomposition of the network. For the sake of computational efficiency, the procedure will first remove ‘redundant’ channels from the network before identifying the ‘blocks’ so that the ‘blocks’ are smaller. A channel in a communication network is said to be irredundant if it is on a simple path starting at the source. otherwise, it is said to be redundant. Moreover, a communication network is said to be irredundant if it contains no redundant channels. In the network illustrated in Figure 19.19, the channels Z X, T X and T Z are redundant.

The deletion of a redundant channel from a network results in a subnetwork with the same set of irredundant channels. Consequently, the irredundant channels in a network define an

782 NETWORK INFORMATION THEORY

Y

S

Wss

X T

Z

Figure 19.19 Redundant channels in a network.

irredundant subnetwork. It can be also shown that, if v is an LCM (respectively, a TILCM) on a network, then v also defines an LCM (respectively, a TILCM) on the subnetwork that results from the deletion of any redundant channel. In addition we say that two nodes are equivalent if there exists a directed path leading from one node to the other and vice versa. An equivalence class under this relationship is called a block in the graph. The source node by itself always forms a block. When every block ‘contracts’ into a single node, the resulting graph is acyclic. In other words, the blocks can be sequentially indexed so that every interblock channel is from a smaller indexed block to a larger indexed block.

For the construction of a ‘good’ TILCM, smaller sizes of blocks tend to facilitate the computation. The extreme favorable case of the block decomposition of a network is when the network is acyclic, which implies that every block consists of a single node. The opposite extreme is when all nonsource nodes form a single block exemplified by the network illustrated in Figure 19.19. The removal of redundant channels sometimes serves for the purpose of breaking up a block into pieces. For the network illustrated in Figure 19.19, the removal of the three redundant channels breaks the block {T, W, X, Y, Z } into the three blocks

{T }, {W, X, Y } and {Z }.

In the construction of a ‘good’ LCM on an acyclic network, as before, the procedure takes one node at a time according to the acyclic ordering of nodes and assigns vectors to outgoing channels from the taken node. For a general network, we can start with the trivial TILCM v on the network consisting of just the source and then expand it to a ‘good’ TILCM v that covers one more block at a time.

The sequential choices of blocks are according to the acyclic order in the block decomposition of the network. Thus, the expansion of the ‘good’ TILCM v at each step involves only incoming channels to nodes in the new block. A heuristic algorithm for assigning vectors v(X Y ) to such channels XY is for v(XY ) to be z times an arbitrary convenient linear combination of vectors assigned to incoming channels to X. In this way, a system of linear equations of the form of Ax = b is set up, where A is a square matrix with the dimension equal to the total number of channels in the network and x is the unknown column vector whose entries are v(XY ) for all channels XY . The elements of A and b are polynomials in z. In particular, the elements of A are either ±1, 0, or a polynomial in z containing the factor z. Therefore, the determinant of A is a formal power series with the constant term (the zeroth power of z) being ±1, and, hence is invertible in F(z). According to Cramer’s rule, a unique solution exists. This is consistent with the physical intuition because the whole network is completely determined once the encoding process for each channel is specified. If this unique solution does not happen to satisfy the requirement for being a ‘good’ TILCM, then

CAPACITY OF WIRELESS NETWORKS USING MIMO TECHNOLOGY

783

the heuristic algorithm calls for adjustments on the coefficients of the linear equations on the trial-and-error basis.

After a ‘good’ TILCM is constructed on the subnetwork formed by irredundant channels in a given network, we may simply assign the zero vectors to all redundant channels.

Example

After the removal of redundant channels, the network depicted by Figure 19.19 consists of four blocks in the order of {S}, {W, X, Y }, {Z } and {T }. The subnetwork consisting of the first two blocks is the same as the network in Figure 19.18. When we expand the trivial TILCM on the network consisting of just the source to cover the block {W, X, Y }, a heuristic trial would be

v(S X) = (1 0)T v(SY ) = (0 1)T

together with the following linear equations:

v(X Y ) = zv(S X) + zv(W X) v(Y W ) = zv(SY ) + zv(XY )

and

v(W X) = zv(Y W ).

The result is the memoryless TILCM v in the preceding example. This TILCM can be further expanded to cover the block {Z } and then the block {T }.

19.6 CAPACITY OF WIRELESS NETWORKS USING MIMO TECHNOLOGY

In this section an information theoretic network objective function is formulated, which takes full advantage of multiple input multiple output (MIMO) channels. The demand for efficient data communications has fueled a tremendous amount of research into maximizing the performance of wireless networks. Much of that work has gone into enhancing the design of the receiver, however considerable gains can be achieved by optimizing the transmitter.

A key tool for optimizing transmitter performance is the use of channel reciprocity. Exploitation of channel reciprocity makes the design of optimal transmit weights far simpler . It has been recognized that network objective functions that can be characterized as functions of the output signal-to-interference noise ratio (SINR) are well suited to the exploitation of reciprocity, since it permits us to relate the uplink and downlink objective functions [48–50]. It is desirable, however, to relate the network objective function to information theory directly, rather than ad-hoc SINR formulations, due to the promise of obtaining optimal channel capacity. This section demonstrates a link between the information theoretic, Gaussian interference channel [51] and a practical network objective function that can be optimized in a simple fashion and can exploit channel reciprocity. The formulation of the objective function permits a water filling [52] optimal power control solution, that can exploit multipath modes, MIMO channels and multipoint networks.

Consider the network of transceivers suggested by Figure 19.20. Each transceiver is labeled by a node number. The nodes are divided into two groups. The group 1 nodes transmit in a given time-slot, while the group 2 nodes receive. In alternate time-slots, group

784 NETWORK INFORMATION THEORY

Figure 19.20 Network layout.

2 nodes transmit, while group 1 nodes receive, according to a time division duplex (TDD) transmission scheme.

The channel is assumed to be channelized into several narrow band frequency channels, whose bandwidth is small enough so that all multipath is assumed to add coherently. This assumption allows the channel between any transmit antenna and any receive antenna to be modeled as a single complex gain and is consistent with an orthogonal frequency division multiplexing (OFDM) modulation scheme. All transceivers are assumed to have either multiple antennas or polarization diversity, for both transmit and receive processing. The link connecting nodes may experience multipath reflections. For this analysis we assume that the receiver synchronizes to a given transmitter and removes any propagation delay offsets. The channel, for a given narrow band, between the transmitter and receiver arrays is therefore modeled as a complex matrix multiply.

Let 1 be an index into the group 1 transceivers, and 2 an index into the group 2 transceivers. When node 2 is transmitting during the uplink, we model the channel between the two nodes as a complex matrix H12( 1, 2) of dimension M1( 1) × M2( 2), where M1( 1) is the size of the antenna array at node 1, and M2( 2) is the size of the array at node 2. Polarization diversity is treated like additional antennae. In the next TDD time slot, during the downlink, node 1 will transmit and the channel from 1 to 2 is described by the complex M2( 1) × M1( 2) matrix H21( 2, 1). For every node pair ( 1, 2) that forms a communications link in the network, we assign a MIMO channel link number, indexed by k or m, in order to label all such connections established by the network. Obviously not every node pair is necessarily assigned a link number, only those that actually communicate. We also define the mapping from the MIMO link number to the associated group 1 node, 1(k) and the associated group 2 node 2(k) by the association of k with the link [ 1(k), 2(k)]. Because each channel is MIMO, a given node will transmit multiple symbols over possibly more than one transmission mode. The set of all transmission modes over the entire network is indexed by q and p. This index represents all the symbol streams that are transmitted from one node to another and therefore represents a lower level link number. The low level link numbers will map to its MIMO channel link, via the mapping k(q). Because our network will

CAPACITY OF WIRELESS NETWORKS USING MIMO TECHNOLOGY

785

exploit channel reciprocity, we assume that every uplink symbol stream indexed by q has an associated downlink symbol stream assigned to the same index. The q th uplink symbol stream is spread by an M1(q) × 1 complex vector g1(q), where M1(q) M1{ 1[k(q)]}. Similarly g2(q) is the associated transmit vector for the downlink.

For each node, we group the transmit vectors into a single matrix,

G2(k) g2(q1), g2(q2), . . . , g2(qMc (k))

(19.61)

and

 

G1(k) g1(q1), g1(q2), . . . , g1(qMc (k))

(19.62)

where k(qi ) = k and there are Mc(k) transmission modes associated with MIMO link k. With these conventions the signal model can be written as:

x1(n; k) = i1(n; k) + H12(k, k)G2(k)d2(n; k)

(19.63)

x2(n; k) = i2(n; k) + H21(k, k)G1(k)d1(n; k)

(19.64)

where x1(n; k) is the received complex data vector at sample

n, and node 1(k),

H12(k, m) H12[ 1(k), 2(m)] is the M1[ 1(k)] × M2[ 2(m)] complex MIMO channel matrix for downlink transmission, n is a time/frequency index, that represents an independent reuse of the channel, either due to adjacent frequency channels (e.g. adjacent OFDM channels) or due to multiple independent time samples, i1(n; k) is the interference vector seen at node 1(k) due to the other transmitting nodes as well as due to background radiation, and d2(n; k) is the Mc(k) × 1 downlink information symbol vector, transmitted for sample n. The analogous model for the uplink case is shown in (4). The interference vector can be written as

ir (n; k) = Hrt (k, m)Gt (m)dt (n; m) + εr (n, k)

(19.65)

m=k

 

where εr (n, k), r = {1, 2}, is the background radiation noise vector seen by the receiving node r (k) and t = {2, 1} is the transmission timeslot indicator. The convention is adopted that t = 2 if r = 1, otherwise t = 1 when r = 2.

19.6.1 Capacity metrics

In this section we will be interested in the asymptotic maximum achievable throughput under certain network constraints:

(1)We will assume that signal energy from all the nodes in the network, other than the one from the link being processed, is treated as interference. This is due to difficulty of synchronizing a node to multiple interference sources and reduces the complexity of the receiver.

(2)The interference ir (n; k) is modeled as complex circular Gaussian noise. The superposition of many distant interferers will appear Gaussian by the central limit theorem.

Considering therefore the channel model described by Equations (19.63) and (19.65), we can write the mutual information between the source data vector and the received data vector

786 NETWORK INFORMATION THEORY

as [52, 53]:

I

xr k

);

dt k

)]

=

log

2

I + Rir ir

k

Rst st

k

= Crt k

Gt

)

(19.66)

[

(

(

 

 

1

( )

 

(

)

( ;

 

where, Gt ↔ {gt (q)} represents all transmit weights stacked into a single parameter vector and where we neglect the n dependency for the random vectors xr (n; k) and dt (n; k). The interference and the signal covariance matrices are defined as:

 

(k, m) + Rεr εr

(k)

(19.67)

Rir ir (k) = Art (k, m)ArHt

m=k

 

 

 

 

 

 

(19.68)

Art (k, m) = Hrt (k, m)Gt (m)

 

 

 

 

(19.69)

Rst st (k) = Art (k, k)ArHt (k, k),

 

and where Rεr εr (k) is the covariance of the background noise vector εr (n, k). The covariance of the source statistics is assumed governed by the transmit weights Gt (k), therefore the covariance of dt (k) is assumed to be the identity matrix.

Let us now consider the introduction of complex linear beam-forming weights at the receiver. For each low level link q, we assign a receive weight vector wr (q), r {1, 2} for use at receiver r [k(q)]. The weights, in theory, are used to copy the transmitted information symbol,

ˆ

 

H

(19.70)

dt (n; q) = wr (q)xr (n; k)

= wrH(q)ir (n; k) + wrH(q)Art (k, k)dt (n; k)

(19.71)

k(q) = k where,

 

 

 

dt (n; k) dt (q1), dt (q2), . . . dt (qMc (k)) T

(19.72)

and k(q j ) = k. The vector version of this can be written as:

ˆ

H

H

(k)Art (k, k)dt (n; k)

(19.73)

dt (n; k) = Wr

(k)ir (n; k) + Wr

where

 

 

 

 

 

Wr (k) wr (q1), wr (q2), · · · wr (qMc (k))

(19.74)

and k(q j ) = k. The structure of the transceiver channel model is illustrated in Figure 19.21 for the uplink case, highlighting the first Mc(k) low-level links, all assumed to be associated with the same MIMO-channel link. The MIMO channel link index k and sample index n are omitted for simplicity. The downlink case is the same after the 1 and 2 indices are swapped.

Owing to the additional processing [52], the mutual information either remains the same or is reduced by the application of the linear receive weights,

ˆ

(19.75)

I [dt (k); dt (k)] I [xr (k); dt (k)]

CAPACITY OF WIRELESS NETWORKS USING MIMO TECHNOLOGY

787

^

^

^

d1(1)

d1(2)

d1(Mc)

w2H(1)

w2H(2)

w2H(Mc)

 

 

i2

 

H21

 

g1(1) g1(2) g1(Mc)

d1(1) d1(2) d1(Mc)

Figure 19.21 Uplink transceiver channel model.

In addition to inequality, Equation (19.75), it can be also shown [54, 55] that, if the source symbols dt (q) are mutually independent then,

I

ˆ

ˆ

(19.76)

dt (k); dt (k)

I dt (q); dt (q)

q: k(q)=k

The lower bound of in Equation (19.76) is referred to as the decoupled capacity and can be written as:

Drt (k; Wr , Gt )

ˆ

(19.77)

I [dt (q); dt (q)]

q: k(q)=k

where the collection of all receive weights is stacked into the parameter vector Wr ↔ {wr (q)}. Equations (19.75) and (19.76) demonstrate that

Crt (k; Gt ) Drt (k; Wr , Gt )

(19.78)

The decoupled capacity is implicitly a function of the transmit weights gt (q) and the receive weights wr (q). Optimizing over these weights can always acheive the full capacity of the Gaussian interference channel and hence the upper bounds in Equation (19.78).

First we write ˆt ( ; ) as, d n q

ˆ

H

H

 

dt (n; q) = wr

(q)ar (q) + wr (q)ir (q),

 

ir (n; q)

Hrt [k, k( p)]gt ( p)dt (n; p) + εr (n, k)

 

 

p =q

 

 

 

ar (q) Hrt (k, k)gt (q), k = k(q)

(19.79)