Digital Communications 1. Safwan El Assad

Чтение книги онлайн.

Читать онлайн книгу Digital Communications 1 - Safwan El Assad страница 11

Digital Communications 1 - Safwan El Assad

Скачать книгу

in an alphabet A = { } of a discrete source: a finite set of predetermined symbols, for example:

      We will use the term discrete source of information Sn to indicate the system selecting, from the set of symbols from As = S, the symbols to transmit:

      The choice of sm,n is carried out according to a given probability law, which can be steady temporally or variable over time.

      2.2.1. Simple source (memoryless)

      This is a source for which the probability of the occurrence of a symbol does not depend on the previous symbols:

      [2.1]

      [2.2] images

      2.2.2. Discrete source with memory

      It is a source for which the probability of occurrence of a symbol depends on the symbols already emitted (statistical dependence) and on instants 1, 2, ..., n where they have been emitted:

      [2.3] images

      If the source is in addition stationary then the dependence is only on the ranks and not on the moments when the symbols have been transmitted, so:

      [2.4] images

      2.2.3. Ergodic source: stationary source with finite memory

      The ergodicity of an information source implies that a temporal average calculated over an infinite duration is equal to the statistical average.

      2.2.4. First order Markovian source (first order Markov chain)

      First order Markov sources play an important role in several domains, for example, in (Caragata et al. 2015), the authors used such a source to model the cryptanalysis of a digital watermarking algorithm.

      It is characterized by:

      [2.5] images

      with:

images

      The probability images = pl,k is called transition probability from state l to state k, and:

      [2.6] images

      The probability that at time n the source is in the state k is:

      where Tt is the transposed matrix of transition probabilities.

      Moreover, if the source is, stationary, then:

images

      [2.10] images

      where P0 is the matrix of probabilities governing the generation of symbols by the source at the initial instant n = 0.

      The realization of an event x of probability p(x) conveys a quantity of information h(x) related to its uncertainty. h(x) is an increasing function of its improbability 1/p(x):

images

      If an event x is certain, then p(x) = 1, the uncertainty is zero and therefore the quantity of information h(x) brought by its realization is null.

      Moreover, let us consider the realization of a pair of independent events x and y. Their joint realization leads to the amount of information it brings being the sum of the quantities of information brought by each of them:

      [2.13] images

      It

Скачать книгу