• Keine Ergebnisse gefunden

An Achievable Rate Region for a Three-Phase Relay Channel

2.3 Achievable Rates for a System with More Than Two Phases

2.3.1 An Achievable Rate Region for a Three-Phase Relay Channel

The goal is to transmit a messagew1from node 1 to node 2 andw2from node 2 to node 1 using the medium between the two nodes and the relay a total ofn∈Ntimes. We start the discussion by adapting the definitions to this setup.

We assume three phases where 1 ≥ α, β, γ ≥ 0, α+β+ γ = 1 indicate timesharing be-tween the phases: In the first phase, node 1 transmits the codewordX1n1 of lengthn1to the relay and node 2 using a channel p1(yR,1,y2,1|x1) n1 := n1(n) ∈ N times with nn1 → α as n → ∞. These nodes will receive the signals YR,1n1 and Y2,1n1 respectively. In the second phase, node 2 transmits the codewordX2n2 of lengthn2to the relay and node 1 using a channel p1(yR,2,y1,2|x2) n2 := n2(n) ∈ Ntimes with nn2 → β asn → ∞. The received signals areYR,2n2 andY1,2n2 respec-tively. In the third phase the relay node transmits XRn3 to node 1 and node 2 using a channel p2(y1,3,y2,3|xR)n3 := n3(n) ∈Ntimes with nn3 → γ asn → ∞. The terminal nodes receive the

signalsY1,3n3 andY2,3n3 respectively. All channels are assumed to be memoryless and the channels in the three phases are assumed to be independent. Therefore we have a joint probability dis-tribution p(yR,1,y2,1,yR,2,y1,2,y1,3,y2,3|x1,x2,xR) = p1(yR,1,y2,1|x1)p2(yR,2,y1,2|x2)pR(y1,3,y1,3|xR) which defines the considered relay channel as follows:

Definition 2.6. A discrete memoryless three-phase two-way relay channel is defined by a family np(n) :Xn11 × Xn22 × XnR3 → YnR,11 × Yn2,11 × YnR,22 × Yn1,22 × Yn1,33 × Yn2,33 o

n1N,n2N,n3N

withn1+n2+n3= n. The family consists of probability transition functions given by p(n)(ynR,11 ,yn2,11,ynR,22 ,yn1,22 ,yn1,33 ,yn2,33 |xn11,xn22,xnR3) :=

n1

Y

i=1

p1(yR,1,(i),y2,1,(i)|x1,(i))

n2

Y

i=1

p2(yR,2,(i),y1,2,(i)|x2,(i))

n3

Y

i=1

pR(y1,3,(i),y2,3,(i)|xR,(i)) for probability functionsp1 :X1 → YR,1×Y2,1, p2:X2→ YR,2×Y1,2andpR :XR → Y1,3×Y2,3. Definition 2.7. A (M(n)1 ,M2(n),n1,n2,n3)-code for the three-phase two-way relay channel under a decode-and-forward protocol consists of an encoder at node one

xn11 :W1→ Xn11 withW1 =[1,2, . . . ,M1(n)], an encoder at node two

xn22 :W2→ Xn22 withW2 =[1,2, . . . ,M2(n)], an encoder at the relay node

xnR3 :W1× W2 → XnR3, a decoder at node one and node two

g1 :Yn1,22 × Yn1,33 × W1 → W2 g2 :Yn2,11 × Yn2,33 × W2 → W1 and a decoder at the relay node

gR :YnR,11 × YnR,22 → W1× W2.

Definition 2.8. When w := w(w1,w2) = [w1,w2] ∈ W := W1 × W2 is the message pair transmitted by the two terminal nodes, the message w2 is decoded in error ifg1(yn1,22 ,yn1,33,w1), w2or ifgR(ynR,11 ,ynR,22 ),( ˜w1,w2) for some ˜w1 ∈ W1. The probability of this error event is denoted

by

λ1(w) := Prh

g1(Y1,2n2,Y1,3n3,w1),w2∨gR(ynR,11 ,ynR,22 ),( ˜w1,w2)|w(w1,w2) has been senti . Accordingly the corresponding error event for the messagew1is denoted by

λ2(w) := Prh

g2(Y2,1n1,Y2,3n3,w2),w1∨gR(ynR,11 ,ynR,22 ),(w1,w˜2)|w(w1,w2) has been senti . Note that the definition for the error is with respect to the messages rather than with respect to the decoder. The reason for this is that we need to capture the constraint of decoding at the relay in the definition of achievable rates.

Definition 2.9. The average probability of decoding error is given by µ(n)1 := 1

|W|

X

w∈W

λ1(w) for messagew2and

µ(n)2 := 1

|W|

X

w∈W

λ2(w) for messagew1.

Definition 2.10. Letµ(n)1 andµ(n)2 be the average probabilities of decoding error for messagew2 andw1, respectively. The rate pair [R1,R2] is said to be achievable for the three-phase two-way relay channel under a decode-and-forward protocol if there exists a sequence of M1(n),M(n)2 ,n1, n2,n3-codes with logMn1(n) →R1and lognM2(n) →R2such thatµ(n)1 , µ(n)2 →0 asn→ ∞.

Theorem 2.7. An achievable rate region for the three-phase two-way relay channel using a decode-and-forward protocol is the set of all rate pairs [R1,R2] satisfying

R1< minαI(X1;YR,1);αI(X1;Y2,1)+γI(XR;Y2,3)

R2< minβI(X2;YR,2);βI(X2;Y1,2)+γI(XR;Y1,3) (2.12) for some joint probability distribution p1(yR,1,y2,1|x1)p2(yR,2,y1,2|x2)pR(y1,3,y2,3|xR)p(x1,x2,xR) and someα, β, γ≥ 0 withα+β+γ= 1.

Remark 2.9. Due to the factorization of the channel, there is no rate loss if we restrict the probability distribution of the input p(x1,x2,xR) such thatX1,X2 andXR are independent.

Remark 2.10 (Necessity of the feedback constraint). The restriction of the nodes not to use any feedback mechanism may not seem necessary for the considered setup. To see that it is indeed necessary, consider a toy setup: Suppose that there is no channel from node 1 to the relay, while all other channels are error and interference free and offer some rate to transmit.

For this setupI(X1;YR,1)=0 and thereforeR1is zero as well. Without the restriction it would of

course be possible to transmit some data from node 1 to node 2 in the first phase. In the second phase node 2 could forward the data received from node 1 to the relay. Therefore the decode-and-forward requirement would be satisfied. If we do not allow such cooperation between the nodes, than using a cutset bound [30] slightly adapted to the considered setup the given region can be shown to be optimal. The adaption is needed, as the proof in [30] assumes, that the messages in the network are independent, while here we transmit the same message to the relay and to the terminal node. Furthermore the encoders in [30] may use all the received signals for the encoding.

Remark 2.11(Convexity of the rate region). Looking at the rate region it is not immediately clear whether the region is convex or not. For fixed timesharing parameters the region is ob-viously convex. It remains to show that there exist parameters and probability distributions such that if [R(1)1 ;R(1)2 ] and [R(2)1 ;R(2)2 ] are achievable with possibly different timesharing pa-rameters, then also a convex combination of both is within the achievable rate region. Indeed the proof is not that difficult, so we will only sketch it for the given region once. It turns out, that the the weighted addition can be encapsulated in some auxiliary variable Q. Note that all terms in the above rate region can be conditioned on Q, if we change p(x1,x2,xR) to p(x1,x2,xR|q)p(q). This will not change the region. Furthermore using the observation in Re-mark 2.9, we can add three variables Q1,Q2, Q3 to the expressions and change p(x1,x2,xR) to p(x1|q1)p(x2|q2)p(x3|q3)p(q1)p(q2)p(q3). The minimum operation can now be split up into two inequalities. We receive forR1being a convex combination ofR(1)1 andR(1)1

R1= aR(1)1 +(1−a)R(1)1 the resulting inequalities

R1 ≤aα(1)I(X(1)1 ;YR,1(1)|Q(1)1 )+(1−a)α(2)I(X(2)1 ;YR,1(2)|Q(2)1 ) and

R1 ≤aα(1)I(X1(1);Y2,1(1)|Q(1)1 )+aγ(1)I(XR(1);Y2,3(1)|Q(1)3 )+

(1−a)α(2)I(X1(2);Y2,1(2)|Q(2)1 )+(1−a)γ(2)I(XR(2);Y2,3(2)|Q(2)3 ).

Now

(1)I(·|Q(1)1 )+(1−a)α(2)I(·|Q(2)1 )= (aα(1)+(1−a)α(2))I(·|Q˜1),

where ˜Q1with|Q˜1|= |Q(1)1 |+|Q(1)1 |is used to include the weighted sum in the expectation over a appropriately constructed probability distribution. Note that we do not change the channel if we switch to ˜Q1, but only the input distribution to the channel. The above steps can be performed for the other inequalities and for R2 where the same variables ˜Q1, ˜Q2, ˜Q3 can be used. We used three auxiliary variables, as the needed random variable might be different for the different

phases. We see that the resulting vector [R1,R2] is in the region for the timesharing parameters α = (aα(1) +(1 −a)α(2)), β = (aβ(1) +(1− a)β(2)), and γ = (aγ(1) +(1− a)γ(2)). Similar arguments can be used for the proof of the convexity of other regions in this thesis.

Proof. Forγ = 0 the result follows immediately by interpreting marginal channels of the two broadcast channels as a compound channel [60, 29] usedn1andn2times respectively. Therefore in what follows we assumeγ >0. LetR1,R2, p(x1,x2,xR),α, β, γ≥ 0 withα+β+γ =1 be given such that the inequalities in (2.12) are strict. The achievability of the closure is a consequence of the definition of achievability and will not be repeated here. It can be proved analogous to the arguments in the proof of Theorem 2.2 in Section 2.1.3.

Random Codebook Generation Letn1 = ⌊αn⌋,n2 = ⌊βn⌋andn3 = n−n1−n2 ≤ ⌈γn⌉+1.

We generate M1(n) = 2⌊nR1 independent codewords X1n1(w1) of length n1 drawn according to Qn1

i=1p(x1,(i)) where p(x1) = P

x2∈X2

P

xR∈XR p(x1,x2,xR) is the marginal probability distribution.

Similarly, we generate M2(n) = 2⌊nR2 independent codewords Xn22(w2) of length n2 drawn ac-cording toQn2

i=1p(x2,(i)) andM(n)1 M2(n)independent codewordsXRn3(w),w= [w1,w2] of lengthn3 drawn according toQn3

i=1p(xR,(i)). The random code is revealed to both terminal nodes and the relay.

Encoding Depending on the message to transmitw1 node 1 sends the corresponding code-word xn1(w1) using the channeln1 times. Node 2 sendsxn2(w2) to transmit the messagew2. To send the decoded pairw=[w1,w2] withwk ∈ Wk,k∈ {1,2}, the relay sends the corresponding codeword xnR3(w).

Decoding The receiving nodes will use typical set decoding. For a strict definition of the decoding sets we choose parameter for the typical sets as ǫ1 < αI(X1,YR,1)−R1, ǫ2 < βI(X2,YR,2)−R2, ǫ3 < γI(XR,Y2,3)+αI(X 1,Y2,1)−R1, and ǫ4 < γI(XR,Y2,3)+αI(X 1,Y2,1)−R1. For the first two phases the relay decides that w1 andw2 are transmitted, if xn11(w1) and xn22(w2) are the only codewords jointly typical with the received signals ynR,11 andynR,22 respectively, i.e.

xn11(w1),ynR,11

∈ Tǫ(n11)(X1,YR,1) and

xn22(w2),ynR,22

∈ Tǫ(n22)(X2,YR,2). Knowingw2, the decoder at node 2 decides that w1 was transmitted, if this is the uniquew1such that

xn11(w1),yn2,11

∈ Tǫ(n31)(X1,Y2,1) and simultaneously xnR3(w1,w2),yn2,33

∈ Tǫ(n43)(XR,Y2,3). Decoding at receiver 1 works in an analogous way. To keep the definition of the decoder consistent the decoders map to the default message w1 = 1 and w2 =1 if no or more than one codewords is found in their respective decoding sets.

Analysis of the Probability of Error Now we bound the probability of error for the decod-ing. We give the proof for the transmission of message w1 to receiver 2. The proof for the messagew2 follows from analogous arguments. We have

µ(n)2 = P(1)e,2(2)e,2+P(1)e,2(1−P˜(2)e,2)+(1−P(1)e,2) ˜P(2)e,2

where ˜P(2)e,2is the average probability of the event, that the decoding at the terminal node fails and P(1)e,2is the average probability for a decoding error in the decoding ofw1at the relay. Therefore we can boundµ(n)2 form above as

µ(n)2 ≤P(1)e,2+P(2)e,2

whereP(2)e,2is the average probability for a decoding error at node 2 given that the relay decoded without error. We can split the analysis of the overall error probability into the analysis of the error in decoding at the relay and of the error in the decoding at the terminal node. As we use the random coding argument, in the analysis, we will average over the random codebook and considerExn1

1 ,xn22,xnR3

(n)k o

Exn1 1 ,xn22,xnR3

nP(1)e,k+P(2)e,ko

. We show that we haveExn1

1 ,xn22,xnR3

(n)1 o

→ 0.

We will then conclude, that there exists at least one codebook with small average probability of error for the decoding ofw1andw2.

Decoding at the relay First consider the decoding at the relay. The relay is in error if either xn11(w1),ynR,11

<Tǫ(n11)(X1,YR,1) or if there exists a ˆw1 , w1with

xn11( ˆw1),ynR,11

∈ Tǫ(n11)(X1,YR,1).

Using the union bound it is sufficient to show that both these error events occur with arbitrarily small probability asn→ ∞.

By the law of large numbers the probability that

xn11(w1),ynR,11

< Tǫ(n11)(X1,YR,1) for se-quences

xn11(w1),ynR,11

drawn according to a joint probability distribution can be made arbitrarily small by choosingn(and as a consequencen1) big.

The probability of

xn11( ˆw1),ynR,11

∈ Tǫ(n11)(X1,YR,1) for ˆw1 , w1 averaged over all codewords and the random codebook can be bounded as follows:

X

ynR,11∈YnR,11

Exn11











 p

ynR,11 |xn11(w1)

|W1|

X

ˆ w1=1 ˆ w1,w1

χT(n1)

ǫ1 (X1,YR,1)(xn11( ˆw1),ynR,11 )













=(|W1| −1) X

ynR,11∈YnR,11

X

xn11∈Xn11

p(xn11)p(ynR,11T(n1)

ǫ1 (X1,YR,1)(xn11,ynR,11 )

≤ 2n(R1+3αǫ1−αI(X1,YR,1))+I(X1,YR,1). The last step follows from the properties of the typical set analogous to the procedure in the proof of Theorem 2.2 and the choice ofαn−1≤ n1 ≤αn. Now forn→ ∞

2n(R1+3αǫ1−αI(X1,YR,1))+I(X1,YR,1) →0 as we choose ǫ1 < αI(X1,YR,1)−R1. We conclude thatExnR

nP(1)e,ko

can be made arbitrarily small forn large.

Decoding at the terminal node For the calculation of the probabilityE{P(2)e,2}we assume that the relay has decoded w1 andw2 without error. Furthermore node 2 received someyn2,11 drawn

according to p

yn2,11 |xn11(w1) .

The error that may occur at node 2 is characterized by several possible events:

• E1: xn11(w1) is not jointly typical withyn2,11 ,

• E2: xnR3(w1,w2) is not jointly typical withyn2,33 ,

• E3: xn21( ˜w1) is jointly typical withyn2,11 for some ˜w1 , w1, or

• E4: xnR3( ˜w1,w2) is jointly typical withyn2,33 for some ˜w1, w1.

For these events we can calculate the probabilityPEi where we take into account the random codebook generation as well as the joint randomness in the system. We can bound the average probability of error for the decoding from above by

E{P(2)e,2} ≤ PE1 +PE2 + M1PE3PE4,

where the average is over the random codebook as well as over the transmitted symbols. The last term follows form the observation that an error occurs if E3andE4happen simultaneously.

The factor M1attributes the fact, that this may happen for each wrong message ˜w1 , w1. ClearlyPE1 → 0 ifn1 → ∞and PE2 → 0 if n3 → ∞which follows from the definition of strong typicality and the law of large numbers. Analogous to the proceeding above we have

PE3 ≤2n(3αǫ3−αI(X1,Y2,1))+I(X1,Y2,1) and

PE4 ≤2n(3γǫ4−γI(XR,Y2,3))+2ǫ4 and therefore

M2PE3PE4 ≤2n(R1−αI(X1,Y2,1)+3αǫ3−γI(XR,Y2,3)−3γǫ4)+I(X1,Y2,1)+2ǫ4. We can now conclude thatEn

P(2)e,2o

→0 asn→ ∞by the choice ofǫ3 < γI(XR,Y2,3)+αI(X 1,Y2,1)−R1 andǫ4< γI(XR,Y2,3)+αI(X 1,Y2,1)−R1.

This proves thatExn1

1 ,xn22,xnR3

(n)1 o

→ 0 forn → ∞. SimilarlyExn1

1 ,xn22,xnR3

(n)2 o

→0. Therefore

Exn11,xn22,xnR3

(n)1(n)2 o

→0 and we can conclude that there is at least one codebook such thatµ(n)1 and µ(n)2 can be made arbitrarily small by choosingnlarge enough. The closure of the region is achievable using similar arguments as in the proof of Theorem 2.2. Therefore the claim is

proved.