Source coding theory has as its goal the characterization of the optimal performance achievable in idealized communication systems which must code an information source for transmission over a digital communication or storage channel for transmission to a user. The user must decode the information into a form that is a good approximation to the original. A code is optimal within some class if it achieves the best possible fidelity given whatever constraints are imposed on the code by the available channel. In theory, the primary constraint imposed on a code by the channel is its rate or resolution, the number of bits per second or per input symbol that it can transmit from sender to receiver. In the real world, complexity may be as important as rate. The origins and the basic form of much of the theory date from Shan- non's classical development of noiseless source coding and source coding subject to a fidelity criterion (also called rate-distortion theory)  .
Shannon combined a probabilistic notion of information with limit theo- rems from ergodic theory and a random coding technique to describe the optimal performance of systems with a constrained rate but with uncon- strained complexity and delay. An alternative approach called asymptotic or high rate quantization theory based on different techniques and approx- imations was introduced by Bennett at approximately the same time . This approach constrained the delay but allowed the rate to grow large.
1 Information Sources.- 1.1 Probability Spaces.- 1.2 Random Variables and Vectors.- 1.3 Random Processes.- 1.4 Expectation.- 1.5 Ergodic Properties.- Exercises.- 2 Codes, Distortion, and Information.- 2.1 Basic Models of Communication Systems.- 2.2 Code Structures.- 2.3 Code Rate.- 2.4 Code Performance.- 2.5 Optimal Performance.- 2.6 Information.- 2.6.1 Information and Entropy Rates.- 2.7 Limiting Properties.- 2.8 Related Reading.- Exercises.- 3 Distortion-Rate Theory.- 3.1 Introduction.- 3.2 Distortion-Rate Functions.- 3.3 Almost Noiseless Codes.- 3.4 The Source Coding Theorem for Block Codes.- 3.4.1 Block Codes.- 3.4.2 A Coding Theorem.- 3.5 Synchronizing Block Codes.- 3.6 Sliding-Block Codes.- 3.7 Trellis Encoding.- Exercises.- 4 Rate-Distortion Functions.- 4.1 Basic Properties.- 4.2 The Variational Equations.- 4.3 The Discrete Shannon Lower Bound.- 4.4 The Blahut Algorithm.- 4.5 Continuous Alphabets.- 4.6 The Continuous Shannon Lower Bound.- 4.7 Vectors and Processes.- 4.7.1 The Wyner-Ziv Lower Bound.- 4.7.2 The Autoregressive Lower Bound.- 4.7.3 The Vector Shannon Lower Bound.- 4.8 Norm Distortion.- Exercises.- 5 High Rate Quantization.- 5.1 Introduction.- 5.2 Asymptotic Distortion.- 5.3 The High Rate Lower Bound.- 5.4 High Rate Entropy.- 5.5 Lattice Vector Quantizers.- 5.6 Optimal Performance.- 5.7 Comparison of the Bounds.- 5.8 Optimized VQ vs. Uniform Quantization.- 5.9 Quantization Noise.- Exercises.- 6 Uniform Quantization Noise.- 6.1 Introduction.- 6.2 Uniform Quantization.- 6.3 PCM Quantization Noise: Deterministic Inputs.- 6.4 Random Inputs and Dithering.- 6.5 Sigma-Delta Modulation.- 6.6 Two-Stage Sigma-Delta Modulation.- 6.7 Delta Modulation.- Exercises.
Series: The Springer International Engineering and Computer Science
Number Of Pages: 188
Published: 31st October 1989
Publisher: SPRINGER VERLAG GMBH
Country of Publication: US
Dimensions (cm): 24.13 x 16.2
Weight (kg): 0.42