<< Chapter < Page | Chapter >> Page > |
So far, we have described the approach by Jalali and Weissman to compress . Suppose instead that is analog. Seeing that MCMC optimizes over a discrete alphabet, it would be convenient to keep doing so. However, because is finite, and assuming that square distortion is used, i.i., , we see that the distortion could be large.
Fortunately, Baron and Weissman showed [link] that the following quantizer has favorable properties,
where is an integer that grows with . That is, as grows the set of reproduction levels quantizers a wider internal more finely. Therefore, we have that the probability that some is an outlier, i.e., , vanishes with , and the effect of outliers on vanishes. Moreover, because the internal is sampled more finely as increases with , this quantizer can emulate any codebook with continuous levels, and so in the limit its RD performance converges to the RD function.
To evaluate the RD performance, we must first define the RD function. Consider a source that generates a sequence . A lossy encoder is a mapping , where is a codebook that contains a set of codewords in . Let be the smallest cardinality codebook for inputs of length generated by such that the expected per-symbol distortion between the input and the codeword is at most . The rate is the per-symbol log-cardinality of the codebook,
This is an operational definition of the RD function in terms of the best block code. In contrast, it can be shown that the RD function is equalto a minimization over mutual information, yielding a different flavor of definition [link] , [link] .
Having defined the RD function, we can describe the RD performance of the compression algorithm by Baron and Weissman for continuous valued inputs.
Theorem 8 Consider square error distortion, , let be a finite variance stationary and ergodic source with RD function , and use the MCMC algorithm with data independent reproduction alphabet and temperature that decays sufficiently slowly. Let be the approximation to after super-iterations. Then the length of context tree weighting (CTW) applied to converges as follows,
Remark 1 Let us make some comments.
Adaptive reproduction levels : The above algorithm is promising from a theoretical perspective, but is of limited practical interest. In order to approach the RD function closely, may need to be large, which slows down the algorithm.
Our approach to overcome the large alphabet is inspired by the work by Rose [link] , who showed that in many cases the optimal RD codebook uses a small discrete-valuedreproduction alphabet. This runs counter to our intuition that a continuous reproduction alphabet is needed to compress a continuous valued source. Therefore, we propose analgorithm that allows for reduction of the alphabet size while supporting adaptive reproduction levels.
Notification Switch
Would you like to follow the 'Universal algorithms in signal processing and communications' conversation and receive update notifications?