Presentation is loading. Please wait.

Presentation is loading. Please wait.

Name Iterative Source- and Channel Decoding Speaker: Inga Trusova Advisor: Joachim Hagenauer.

Similar presentations


Presentation on theme: "Name Iterative Source- and Channel Decoding Speaker: Inga Trusova Advisor: Joachim Hagenauer."— Presentation transcript:

1 Name Iterative Source- and Channel Decoding Speaker: Inga Trusova Advisor: Joachim Hagenauer

2 Name- 2 - 1. Introduction 2. System model 3. Joint Source-Channel Decoding(JSCD) 4. Iterative Source-Channel Decoding(ISCD) 5. Simulation Results 6. Conclusions Content

3 Name- 3 - Introduction PROBLEMS EXIST Limited block length for source and channel coding Data-bits issued by a source encoder contain residual redundancies Infinite block-length for achieving “perfect” channel codes Output bits of a practical channel decoder are not error free Application of the separation theorem of information theory is not justified in practice!

4 Name- 4 - Introduction GOAL: To improve the performance of communication systems without sacrificing resources SOLUTION: Joint source-channel coding & decoding (JSCCD) Several auto correlated source signals are considered Source samples are 1. quantized 2. their indexes appropriately mapped into bit vectors 3. bits are interleaved & channel-encoded

5 Name- 5 - Introduction AREA OF INTEREST: Joint source-channel decoding (JSCD) Key idea of JSCD: To exploit the residual redundancies in the data bits in order To improve the overall quality of the transmission The turbo principle (iterative decoding between components) is a general scheme, which we apply to JSCD

6 Name- 6 - System Model Initial data AWGN channel is assumed for transmission A set of input source signals has to be transmitted at each time index k Only one of the inputs, the samples, is considered are quantized by the bit vector with and, denoting the set of all possible N-bit vectors (1)

7 Name- 7 - System Model Figure 1: System Model

8 Name- 8 - System Model As coherently detected binary modulation (phase shift keying) is assumed Than conditional pdf of the received value at the channel output, given that code bit has been transmitted, is given by (3)

9 Name- 9 - System Model Where, the variance energy that is used to transmit each channel-code bit One-sided power spectral density of the channel noise Note: The joint conditional pdf for a channel word to be received, given that codeword is transmitted, is the product of (3) over all code-bits, since the channel noise is statistically independent

10 Name- 10 - System Model IF are autocorrelated THAN show dependencies AND are modeled by first-oder stationary Markov-process, which is described by transition probabilities ASSUMPTIONS Transition probabilities and probability-distributions of the bitvectors are known Bitvectors are independent of all other data, which is transmitted in parallel by bitvector

11 Name- 11 - Joint Source-Channel Decoding GOAL: Distortion of the decoder output signal min JSCD for a fixed transmitter Optimization criterion is given by the conditional expectation of the mean square error: (4)

12 Name- 12 - Joint Source-Channel Decoding In (4) is the quantizer reproduction value corresponding to the bitvector, which is used by the source encoder to quantize is a set of channel output words which were received up to the current time k D min results in the minimum mean – square estimator (5) (6)

13 Name- 13 - Joint Source-Channel Decoding Bitvector a-posteriori probabilities (APPs), using the Bayes- rule, are given by Where is the bitvector a-priori probability is a normalizing constant Since (7)

14 Name- 14 - Joint Source-Channel Decoding A-priori probabilities are given by At k=0 the unconditional probability distribution is used in stead of the “old” APPs Drawback : From (7) the term is very hard to compute analytically (8)

15 Name- 15 - Iterative Source-Channel Decoding Goal: To find more feasible, less complex way to compute at least a good approximation Solution: Iterative Source-Channel Decoding (ISCD) We write: (9)

16 Name- 16 - Iterative Source-Channel Decoding Now, Bitvector probability densities are approximated by the product over the corresponding bit probability densities With the bits (10)

17 Name- 17 - Iterative Source-Channel Decoding If we insert (10) into the formula (7) which defines bitvector a-posteriori probabilities we obtain: The bit a-posteriori probabilities can be efficiently computed by the symbol-by-symbol APP algorithm for a binary convolution channel code with a small number of states. (11)

18 Name- 18 - Iterative Source-Channel Decoding Note: ALL the received channel words up to the current time are used for the computation of the bit APPs, because the bit-based a-priori information For a specific bit (12)

19 Name- 19 - Iterative Source-Channel Decoding Let interpret the fraction in (11) as the extrinsic information that we get from the channel decoder: Note: Superscript “(C)” is used to indicate that is the extrinsic information produced by the channel decoder. (13)

20 Name- 20 - Iterative Source-Channel Decoding As a result we have: A modified channel-term (btw brackets ) that includes the reliabilities of the received bits and, additionally, the information derived by the APP-algorithm from the channel-code. Drawback: Bitvector APPs are only approximations of the optimal values, since the bit a-priori information didn’t contain the mutual dependencies of the bits within bitvectors

21 Name- 21 - Iterative Source-Channel Decoding How to improve the accuracy of the bitvector APPs? Idea: Iterative decoding of turbo codes: From the intermediate results for the bitvector APPs (13), new bit APPs are computed by (14)

22 Name- 22 - Iterative Source-Channel Decoding Bit extrinsic information from the source decoder: Note: Computed extrinsic information is used as the new a- priori information for the second and further runs of the channel decoder. (15)

23 Name- 23 - Iterative Source-Channel Decoding SUMMARY OF ISCD: Step 1 At each time k, compute the initial bitvector a-priori probabilities by: (8)

24 Name- 24 - Iterative Source-Channel Decoding Step 2: Use the results from step 1 in to compute the initial bit a-priori information for the APP channel decoder. Step 3: Perform APP channel decoding (12)

25 Name- 25 - Iterative Source-Channel Decoding Step 4: Perform source decoding by inserting the extrinsic bit information from APP channel decoding into to compute new (temporary) bitvector APPs Step 5: If this is the last iteration proceed with step 8, otherwise continue with step 6 (13)

26 Name- 26 - Iterative Source-Channel Decoding Step 6: Use the bitvector APPs of step 4 in to compute extrinsic bit information from the source redundancies (14) (15)

27 Name- 27 - Iterative Source-Channel Decoding Step 7: Set the extrinsic bit information from Step 6 equal to the new bit a-priori information for the APP channel decoder in the next iteration ; proceed with Step 3 Step 8: Estimate the receiver output signals by using the bitvector APPs from Step 4 (6)

28 Name- 28 - Iterative source channel decoding Figure 2: Iterative Source-Channel Decoding according to the Turbo Principle

29 Name- 29 - Iterative source channel decoding Computation of the bitvector APPs by (13) requires bit probabilities which can be computed from from the output L-values: With inversion: (16) (17) are fixed real numbers

30 Name- 30 - Iterative source channel decoding SIMPLIFICATION 1: Reminder: formula (13) bitvector APPs computation: Let’s insert (17) into (13) and turn the product over the exponential functions into summations in the exponents: (18) (13)

31 Name- 31 - Iterative source channel decoding Benefits of using (18) instead of (13): Normalizing constant Ak doesn’t depend on the variable I k,n L-values from the APP channel decoder can be integrated into the Optimal-Estimation algorithm for APP source decoding without converting the individual L-values back to bit probabilities Strong numerical advantages

32 Name- 32 - Iterative source channel decoding SIMPLIFICATION 2: The computation of new bit APPs within the iteration is still carried out by (14) Reminder: But, instead of (15) for the new bit extrinsic information Reminder: (14) (15)

33 Name- 33 - Iterative source channel decoding Extrinsic L-values are used: Benefits of using (19) in stead of (15): Division is turned into a simple subtraction in the L-value domain THUS, In ISCD the L-values from the APP channel decoder are used and the probabilities are not required (19)

34 Name- 34 - Quantizer Bit Mapping Assumption: Input is a low-pass correlation The value of the sample xk will be close to xk-1 If: The channel code is strong enough L-values at the APP channel decoder output have large magnitudes a-priori information for the source decoder is perfect ISCD: APP source decoder tries to generate extrinsic information for a particular data bit, while it exactly knows all other bits

35 Name- 35 - Quantizer Bit Mapping Figure 3: Bit Mappings for a 3-bit Quantizer to be used in ISCD natural optimized Gray

36 Name- 36 - Simulation Results Simulation process: Correlation of independent Gaussian random samples by a first-order recursive filter (coefficient ) Source encoders: 5-bit Lloyd Max scalar quantizers 50 mutually independent bitvectors were generated, all transmitted at time index k. The bits were scrambled by a random-interleaver

37 Name- 37 - Simulation Results Simulation process (contd.): The bits were channel-encoded by a rate- ½ recursive systematic convolution code (RSC-code with memory 4, which were terminated after each block of 50 bitvectors (250 bits) AWGN-channel was used for transmission ISCD was performed at the decoder

38 Name- 38 - Simulation Results Figure 4: Performance of ISCD for various 5-Bit Mappings

39 Name- 39 - Conclusions Strong quality gains are achievable by: Application of the turbo principle in joint source-channel decoding Bitmapping of the quantizers is important for the performance Optimized bit mapping of the quantizers in ISCD allows to obtain strong quality improvements

40 Name- 40 - Thank you for your attention!


Download ppt "Name Iterative Source- and Channel Decoding Speaker: Inga Trusova Advisor: Joachim Hagenauer."

Similar presentations


Ads by Google