Preface
Just a quick post about recovering a signal, which has been corrupted by quantization error. This is something I discovered while working on my GPSDO project, but has general applicability. This all might be common knowledge and trivial, but I enjoyed coming up with my own solution. Documenting it here so I don't forget about it.
Introduction
Consider a continuous observable, which is sampled by a low resolution process. By low resolution, I mean that there is a limited number of steps that can be differentiated. In a noiseless ideal case the response would look something like shown in figure 1.
![]() |
Figure 1. Ideal quantization of continuous observable. |
However, consider then that the observable is corrupted by additive noise. This causes the sampled signal to also exhibit noise. The result might look as shown in figure 2.
![]() |
Figure 2. Effect of additive noise on sampling. |
This noise in the sampled signal can be used as a way to increase resolution, where several samples are averaged to produce a single sample. Noise is sometimes deliberately injected into measurement signals to take advantage of this. With enough samples averaged, not only is the additive noise reduced, so is also the quantization error. This is illustrated in figure 3.
![]() |
Figure 3. Result of averaging noisy sampled data. |
Now, the example above exhibits a large amount of additive noise. Noise standard deviation was taken as 40% of quantization resolution. However, if the additive noise is small compared to the quantization resolution, this breaks down. See figure 4, where the noise standard deviation is just 15% of quantization resolution.
![]() | ||
Figure 4. Effect of small additive noise. |
Suddenly, the quantization error is no longer removed through averaging and a step-like pattern emerges.
An important thing of note is that there are again some continuous values for which there is only a single sampled value. It seems rather clear that in general no further resolution can be extracted for those values. However, for the following discussion it's more interesting to observe the values that do still exhibit two distinct sampled values. This is also the situation I have in my GPSDO project that motivated all this.
In the quantization transition regions averaging clearly gives more resolution, but the averaged values do not match the true values. The obvious question is whether this can be improved.
Bayes saves the day
Initially I took a Bayesian approach to the problem. When presented with the quantized value q, what can be inferred of the continuous value c? This is answered by Bayes' theorem. p(c|q)=p(q|c)p(c)p(q).
That is, the probability of the continuous variable having value c after we've observed quantized value q is proportional to the prior probability of observing c times the probability of observing quantized value q given that the continuous variable has value c, which is also called the likelihood. The probability of observing the quantized value q in the denominator can be thought of just a normalizing factor.
The formula expresses how we update our knowledge of the continuous variable, when presented with a measurement. The posterior distribution is proportional to the prior times the likelihood.
Before we can formulate the likelihood, we must properly define the additive noise and the quantization. To keep things simple, let's take almost the simplest quantization possible: the round function. Also let's take the noise as normally distributed and zero mean n∼N(0,σ).
The likelihood is the probability of sampling a value, given that the continuous variable has some value. With our choice of quantization, we observe quantized value q if the continuous value plus noise c+n is in [q−0.5,q+0.5). The probability can be evaluated as p(q|c)=p(q−0.5≤c+n≤q+0.5)=p(q−0.5−c≤n≤q+0.5−c). Which, since the noise is normally distributed, can be further evaluated as p(q|c)=Fn(q+0.5−c)−Fn(q−0.5−c)=12erf(q+0.5−c√2σ)−12erf(q−0.5−c√2σ).
To gain more intuition, let's look at what the likelihoods look like. Let's take the noise standard deviation as σ=0.1. This is shown in figure 5.
![]() |
Figure 5. Likelihood functions for observing quantized value 0 and 1. |
Near the continuous values 0 and 1 there is an almost 100% probability for observing quantized values 0 and 1, which causes the stairs we saw in figure 4. The significant overlap area of the probabilities is only a small fraction of the range, which leads to the average not matching the true value also seen in figure 4.
Then, say that we have observed a pattern 1, 0, 0, 0, 1, 0, 0, 0. What can we say about the continuous variable? Taking the prior distribution as uniform (i.e. no knowledge of the continuous variable), the Bayesian inference progresses as shown in figure 6.
![]() |
Figure 6. Inference of pattern 1, 0, 0, 0, 1, 0, 0, 0. |
This approach is very powerful, but also very heavy. And as an additional complexity, if the system is dynamic, we need to add uncertainty to the distribution between inference steps, to allow the inferred continuous variable to evolve. For the usual applications, this likely adds too much complexity to be practical.
This exercise, however, gave me insight in why the simple averages of the sampled values don't give the correct result, as well as how the likelihoods look like.
A practical solution
In practice, the sampled values alternate at most between 2 distinct values. While it's not strictly impossible to observe 3 or more values, for the situations of interest here, the probability is just extremely low. This allows us to simplify things by allowing only two values. The likelihood of observing a zero is then simplified as p(q=0|c)=Fn(0.5−c)=12+12erf(0.5−c√2σ),while the likelihood of observing a 1 is the complement of this. These likelihood functions are visualized in figure 7.
![]() |
Figure 7. Likelihoods of the two observable values. |
Since the two cases are complements of each other, inferring the continuous variable becomes much easier: probability of observing, say, the value 1, can be directly mapped to the value of the continuous variable. The probability, on the other hand (for the values 0 and 1), can be simply estimated as the average of observed samples. Figure 8 shows the mapping from probability to value.
![]() |
Figure 8. Continuous variable inference from probability of observed value. |
While it appears in figure 8 that a probability of 0 gets mapped the continuous variable value 0 and that the probability of 1 gets mapped to the continuous variable value 1, this is not quite the case. This is because there are still very tiny probabilities unaccounted for in figure 8. However, since these are practically impossible, we can round the remaining probability to zero.
Since estimating the probability of observing a 1 is the same thing as computing the average of ones and zeros, we can extend the concept of averaging the values to the other values as well. Turns out this way we can define a single curve that takes the average of the observations and outputs an estimate of the continuous variable. See figure 9.
![]() |
Figure 9. Continuous variable estimation from average of observations. |
This approach is very lightweight to compute, and allows handling dynamic situations by choosing the length of time that the average is computed over.
Let's revisit the situation of figure 4 using this approach. The results are presented in figure 10.
![]() |
Figure 10. Estimation method demonstration. |