Arbitrarily varying channel – Wikipedia

before-content-x4

An arbitrarily varying channel (AVC) is a communication channel model used in coding theory, and was first introduced by Blackwell, Breiman, and Thomasian. This particular channel has unknown parameters that can change over time and these changes may not have a uniform pattern during the transmission of a codeword.

n{displaystyle textstyle n}
after-content-x4

uses of this channel can be described using a stochastic matrix

Wn:Xn×{displaystyle textstyle W^{n}:X^{n}times }

SnYn{displaystyle textstyle S^{n}rightarrow Y^{n}}

, where

X{displaystyle textstyle X}

is the input alphabet,

Y{displaystyle textstyle Y}

is the output alphabet, and

Wn(y|x,s){displaystyle textstyle W^{n}(y|x,s)}

is the probability over a given set of states

S{displaystyle textstyle S}

, that the transmitted input

after-content-x4
x=(x1,,xn){displaystyle textstyle x=(x_{1},ldots ,x_{n})}

leads to the received output

y=(y1,,yn){displaystyle textstyle y=(y_{1},ldots ,y_{n})}

. The state

si{displaystyle textstyle s_{i}}

in set

S{displaystyle textstyle S}

can vary arbitrarily at each time unit

i{displaystyle textstyle i}

. This channel was developed as an alternative to Shannon’s Binary Symmetric Channel (BSC), where the entire nature of the channel is known, to be more realistic to actual network channel situations.

Capacities and associated proofs[edit]

Capacity of deterministic AVCs[edit]

An AVC’s capacity can vary depending on the certain parameters.

R{displaystyle textstyle R}

is an achievable rate for a deterministic AVC code if it is larger than

0{displaystyle textstyle 0}

, and if for every positive

ε{displaystyle textstyle varepsilon }

and

δ{displaystyle textstyle delta }

, and very large

n{displaystyle textstyle n}

, length-

n{displaystyle textstyle n}

block codes exist that satisfy the following equations:

1nlogN>Rδ{displaystyle textstyle {frac {1}{n}}log N>R-delta }

maxsSne¯(s)ε{displaystyle displaystyle max _{sin S^{n}}{bar {e}}(s)leq varepsilon }

, where

N{displaystyle textstyle N}

is the highest value in

Y{displaystyle textstyle Y}

and where

e¯(s){displaystyle textstyle {bar {e}}(s)}

is the average probability of error for a state sequence

s{displaystyle textstyle s}

. The largest rate

R{displaystyle textstyle R}

represents the capacity of the AVC, denoted by

c{displaystyle textstyle c}

.

As you can see, the only useful situations are when the capacity of the AVC is greater than

0{displaystyle textstyle 0}

, because then the channel can transmit a guaranteed amount of data

c{displaystyle textstyle leq c}

without errors. So we start out with a theorem that shows when

c{displaystyle textstyle c}

is positive in an AVC and the theorems discussed afterward will narrow down the range of

c{displaystyle textstyle c}

for different circumstances.

Before stating Theorem 1, a few definitions need to be addressed:

  • An AVC is symmetric if

Theorem 1:

c>0{displaystyle textstyle c>0}

c>0{displaystyle textstyle c>0}

c=maxPI(P){displaystyle displaystyle c=max _{P}I(P)}

.

Proof of 1st part for symmetry: If we can prove that

I(P){displaystyle textstyle I(P)}

is positive when the AVC is not symmetric, and then prove that

c=maxPI(P){displaystyle textstyle c=max _{P}I(P)}

, we will be able to prove Theorem 1. Assume

I(P){displaystyle textstyle I(P)}

were equal to

0{displaystyle textstyle 0}

. From the definition of

I(P){displaystyle textstyle I(P)}

, this would make

Xr{displaystyle textstyle X_{r}}

and

Yr{displaystyle textstyle Y_{r}}

independent random variables, for some

Sr{displaystyle textstyle S_{r}}

, because this would mean that neither random variable’s entropy would rely on the other random variable’s value. By using equation

PXrSrYr{displaystyle textstyle P_{X_{r}S_{r}Y_{r}}}

, (and remembering

PXr=P{displaystyle textstyle P_{X_{r}}=P}

,) we can get,

So now we have a probability distribution on

Yr{displaystyle textstyle Y_{r}}

that is independent of

Xr{displaystyle textstyle X_{r}}

. So now the definition of a symmetric AVC can be rewritten as follows:

sSW(y|s)PSr(s)=sSW(y|s)PSr(s){displaystyle displaystyle sum _{sin S}W'(y|s)P_{S_{r}}(s)=sum _{sin S}W'(y|s)P_{S_{r}}(s)}

since

U(s|x){displaystyle textstyle U(s|x)}

and

W(y|x,s){displaystyle textstyle W(y|x,s)}

are both functions based on

x{displaystyle textstyle x}

, they have been replaced with functions based on

s{displaystyle textstyle s}

and

y{displaystyle textstyle y}

only. As you can see, both sides are now equal to the

PYr(y){displaystyle textstyle P_{Y_{r}}(y)}

we calculated earlier, so the AVC is indeed symmetric when

I(P){displaystyle textstyle I(P)}

is equal to

0{displaystyle textstyle 0}

. Therefore,

I(P){displaystyle textstyle I(P)}

can only be positive if the AVC is not symmetric.

Proof of second part for capacity: See the paper “The capacity of the arbitrarily varying channel revisited: positivity, constraints,” referenced below for full proof.

Capacity of AVCs with input and state constraints[edit]

The next theorem will deal with the capacity for AVCs with input and/or state constraints. These constraints help to decrease the very large range of possibilities for transmission and error on an AVC, making it a bit easier to see how the AVC behaves.

Before we go on to Theorem 2, we need to define a few definitions and lemmas:

For such AVCs, there exists:

– An input constraint
– A state constraint

Assume

g(x){displaystyle textstyle g(x)}

is a given non-negative-valued function on

X{displaystyle textstyle X}

and

l(s){displaystyle textstyle l(s)}

is a given non-negative-valued function on

S{displaystyle textstyle S}

and that the minimum values for both is

0{displaystyle textstyle 0}

. In the literature I have read on this subject, the exact definitions of both

g(x){displaystyle textstyle g(x)}

and

l(s){displaystyle textstyle l(s)}

(for one variable

xi{displaystyle textstyle x_{i}}

,) is never described formally. The usefulness of the input constraint

Γ{displaystyle textstyle Gamma }

and the state constraint

Λ{displaystyle textstyle Lambda }

will be based on these equations.

For AVCs with input and/or state constraints, the rate

R{displaystyle textstyle R}

is now limited to codewords of format

x1,,xN{displaystyle textstyle x_{1},dots ,x_{N}}

that satisfy

g(xi)Γ{displaystyle textstyle g(x_{i})leq Gamma }

, and now the state

s{displaystyle textstyle s}

is limited to all states that satisfy

l(s)Λ{displaystyle textstyle l(s)leq Lambda }

. The largest rate is still considered the capacity of the AVC, and is now denoted as

c(Γ,Λ){displaystyle textstyle c(Gamma ,Lambda )}

.

Lemma 1: Any codes where

Λ{displaystyle textstyle Lambda }

is greater than

Λ0(P){displaystyle textstyle Lambda _{0}(P)}

cannot be considered “good” codes, because those kinds of codes have a maximum average probability of error greater than or equal to

N12N1nlmax2n(ΛΛ0(P))2{displaystyle textstyle {frac {N-1}{2N}}-{frac {1}{n}}{frac {l_{max}^{2}}{n(Lambda -Lambda _{0}(P))^{2}}}}

, where

lmax{displaystyle textstyle l_{max}}

is the maximum value of

l(s){displaystyle textstyle l(s)}

. This isn’t a good maximum average error probability because it is fairly large,

N12N{displaystyle textstyle {frac {N-1}{2N}}}

is close to

12{displaystyle textstyle {frac {1}{2}}}

, and the other part of the equation will be very small since the

(ΛΛ0(P)){displaystyle textstyle (Lambda -Lambda _{0}(P))}

value is squared, and

Λ{displaystyle textstyle Lambda }

is set to be larger than

Λ0(P){displaystyle textstyle Lambda _{0}(P)}

. Therefore, it would be very unlikely to receive a codeword without error. This is why the

Λ0(P){displaystyle textstyle Lambda _{0}(P)}

condition is present in Theorem 2.

Theorem 2: Given a positive

Λ{displaystyle textstyle Lambda }

and arbitrarily small

α>0{displaystyle textstyle alpha >0}

β>0{displaystyle textstyle beta >0}

δ>0{displaystyle textstyle delta >0}

nn0{displaystyle textstyle ngeq n_{0}}

and for any type

P{displaystyle textstyle P}

with conditions

Λ0(P)Λ+α{displaystyle textstyle Lambda _{0}(P)geq Lambda +alpha }

and

minxXP(x)β{displaystyle displaystyle min _{xin X}P(x)geq beta }

, and where

PXr=P{displaystyle textstyle P_{X_{r}}=P}

, there exists a code with codewords

x1,,xN{displaystyle textstyle x_{1},dots ,x_{N}}

, each of type

P{displaystyle textstyle P}

, that satisfy the following equations:

1nlogN>I(P,Λ)δ{displaystyle textstyle {frac {1}{n}}log N>I(P,Lambda )-delta }

maxl(s)Λe¯(s)exp(nγ){displaystyle displaystyle max _{l(s)leq Lambda }{bar {e}}(s)leq exp(-ngamma )}

, and where positive

n0{displaystyle textstyle n_{0}}

and

γ{displaystyle textstyle gamma }

depend only on

α{displaystyle textstyle alpha }

,

β{displaystyle textstyle beta }

,

δ{displaystyle textstyle delta }

, and the given AVC.

Proof of Theorem 2: See the paper “The capacity of the arbitrarily varying channel revisited: positivity, constraints,” referenced below for full proof.

Capacity of randomized AVCs[edit]

The next theorem will be for AVCs with randomized code. For such AVCs the code is a random variable with values from a family of length-n block codes, and these codes are not allowed to depend/rely on the actual value of the codeword. These codes have the same maximum and average error probability value for any channel because of its random nature. These types of codes also help to make certain properties of the AVC more clear.

Before we go on to Theorem 3, we need to define a couple important terms first:

Wζ(y|x)=sSW(y|x,s)PSr(s){displaystyle displaystyle W_{zeta }(y|x)=sum _{sin S}W(y|x,s)P_{S_{r}}(s)}


I(P,ζ){displaystyle textstyle I(P,zeta )}

is very similar to the

I(P){displaystyle textstyle I(P)}

equation mentioned previously,

I(P,ζ)=minYrI(XrYr){displaystyle displaystyle I(P,zeta )=min _{Y_{r}}I(X_{r}land Y_{r})}

, but now the pmf

PSr(s){displaystyle textstyle P_{S_{r}}(s)}

is added to the equation, making the minimum of

I(P,ζ){displaystyle textstyle I(P,zeta )}

based a new form of

PXrSrYr{displaystyle textstyle P_{X_{r}S_{r}Y_{r}}}

, where

Wζ(y|x){displaystyle textstyle W_{zeta }(y|x)}

replaces

W(y|x,s){displaystyle textstyle W(y|x,s)}

.

Theorem 3: The capacity for randomized codes of the AVC is

c=maxPI(P,ζ){displaystyle displaystyle c=max_{P}I(P,zeta )}

.

Proof of Theorem 3: See paper “The Capacities of Certain Channel Classes Under Random Coding” referenced below for full proof.

See also[edit]

References[edit]

after-content-x4