Back to Search Start Over

Mutual Information, the Linear Prediction Model, and CELP Voice Codecs

Authors :
Jerry D. Gibson
Source :
Information, Vol 10, Iss 5, p 179 (2019), Information, vol 10, iss 5, Information, Volume 10, Issue 5
Publication Year :
2019
Publisher :
MDPI AG, 2019.

Abstract

We write the mutual information between an input speech utterance and its reconstruction by a code-excited linear prediction (CELP) codec in terms of the mutual information between the input speech and the contributions due to the short-term predictor, the adaptive codebook, and the fixed codebook. We then show that a recently introduced quantity, the log ratio of entropy powers, can be used to estimate these mutual informations in terms of bits/sample. A key result is that for many common distributions and for Gaussian autoregressive processes, the entropy powers in the ratio can be replaced by the corresponding minimum mean squared errors. We provide examples of estimating CELP codec performance using the new results and compare these to the performance of the adaptive multirate (AMR) codec and other CELP codecs. Similar to rate distortion theory, this method only needs the input source model and the appropriate distortion measure.

Details

ISSN :
20782489
Volume :
10
Database :
OpenAIRE
Journal :
Information
Accession number :
edsair.doi.dedup.....4d5804f3257b5ecf61aace9705c1359f