Pyramid vector quantization explained

Pyramid vector quantization (PVQ) is a method used in audio and video codecs to quantize and transmit unit vectors, i.e. vectors whose magnitudes are known to the decoder but whose directions are unknown. PVQ may also be used as part of a gain/shape quantization scheme, whereby the magnitude and direction of a vector are quantized separately from each other. PVQ was initially described in 1986 in the paper "A Pyramid Vector Quantizer" by Thomas R. Fischer.[1]

One caveat of PVQ is that it operates under the taxicab distance (L1-norm). Conversion to/from the more familiar Euclidean distance (L2-norm) is possible via vector projection, though results in a less uniform distribution of quantization points (the poles of the Euclidean n-sphere become denser than non-poles).[2] No efficient algorithm for the ideal (i.e., uniform) vector quantization of the Euclidean n-sphere is known as of 2010.[3] This non-uniformity can be reduced by applying deformation like coordinate-wise power before projection, reducing mean-squared quantization error by ~10%.[4]

PVQ is used in the CELT audio codec (inherited into Opus) and the Daala video codec.

Overview

As a form of vector quantization, PVQ defines a codebook of quantization points, each of which is assigned an integer codeword from 0 to -1. The goal of the encoder is to find the codeword of the closest vector, which the decoder must decode back into a vector.

The PVQ codebook consists of all -dimensional points

\vec{p}

with integer-only coordinates whose absolute values sum to a constant (i.e. whose L1-norm equals). In set-builder notation:

S(N,K)=\left\{\vec{p}\inZN:\left\|\vec{p}\right\|1=K\right\}

where

\left\|\vec{p}\right\|1

denotes the L1-norm of

\vec{p}

.

As it stands, the set tesselates the surface of an -dimensional pyramid. If desired, we may reshape it into a sphere by "projecting" the points onto the sphere, i.e. by normalizing them:

S
sphere(N,K)=\left\{\vec{p
} : \vec \in S(N,K)\right\}

where

\left\|\vec{p}\right\|2

denotes the L2-norm of

\vec{p}

.

\vec{v}

at the cost of larger integer codewords that require more bits to transmit.

Example

Suppose we wish to quantize three-dimensional unit vectors using the parameter =2. Our codebook becomes:

CodewordPointNormalized point
0<-2, 0, 0><-1.000, 0.000, 0.000>
1<-1, -1, 0><-0.707, -0.707, 0.000>
2<-1, 0, -1><-0.707, 0.000, -0.707>
3<-1, 0, 1><-0.707, 0.000, 0.707>
4<-1, 1, 0><-0.707, 0.707, 0.000>
5<0, -2, 0><0.000, -1.000, 0.000>
6<0, -1, -1><0.000, -0.707, -0.707>
7<0, -1, 1><0.000, -0.707, 0.707>
8<0, 0, -2><0.000, 0.000, -1.000>
CodewordPointNormalized point
9<0, 0, 2><0.000, 0.000, 1.000>
10<0, 1, -1><0.000, 0.707, -0.707>
11<0, 1, 1><0.000, 0.707, 0.707>
12<0, 2, 0><0.000, 1.000, 0.000>
13<1, -1, 0><0.707, -0.707, 0.000>
14<1, 0, -1><0.707, 0.000, -0.707>
15<1, 0, 1><0.707, 0.000, 0.707>
16<1, 1, 0><0.707, 0.707, 0.000>
17<2, 0, 0><1.000, 0.000, 0.000>

(0.707 =

\sqrt{2}/2

rounded to 3 decimal places.)

Now, suppose we wish to transmit the unit vector <0.592, -0.720, 0.362> (rounded here to 3 decimal places, for clarity). According to our codebook, the closest point we can pick is codeword 13 (<0.707, -0.707, 0.000>), located approximately 0.381 units away from our original point.

Increasing the parameter results in a larger codebook, which typically increases the reconstruction accuracy. For example, based on the Python code below, =5 (codebook size: 102) yields an error of only 0.097 units, and =20 (codebook size: 1602) yields an error of only 0.042 units.

Python code

import itertoolsimport mathfrom typing import List, NamedTuple, Tuple

class PVQEntry(NamedTuple): codeword: int point: Tuple[int, ...] normalizedPoint: Tuple[float, ...]

def create_pvq_codebook(n: int, k: int) -> List[PVQEntry]: """ Naive algorithm to generate an n-dimensional PVQ codebook with k pulses. Runtime complexity: O(k**n) """ ret = [] for p in itertools.product(range(-k, k + 1), repeat=n): if sum(abs(x) for x in p)

k: norm = math.sqrt(sum(x ** 2 for x in p)) q = tuple(x / norm for x in p) ret.append(PVQEntry(len(ret), p, q))

return ret

def search_pvq_codebook(codebook: List[PVQEntry], p: Tuple[float, ...]) -> Tuple[PVQEntry, float]: """ Naive algorithm to search the PVQ codebook. Returns the point in the codebook that's "closest" to p, according to the Euclidean distance.) """ ret = None min_dist = None for entry in codebook: q = entry.normalizedPoint dist = math.sqrt(sum((q[j] - p[j]) ** 2 for j in range(len(p)))) if min_dist is None or dist < min_dist: ret = entry min_dist = dist

return ret, min_dist

def example(p: Tuple[float, ...], k: int) -> None: n = len(p) codebook = create_pvq_codebook(n, k) print("Number of codebook entries: " + str(len(codebook))) entry, dist = search_pvq_codebook(codebook, p) print("Best entry: " + str(entry)) print("Distance: " + str(dist))

phi = 1.2theta = 5.4x = math.sin(phi) * math.cos(theta)y = math.sin(phi) * math.sin(theta)z = math.cos(phi)p = (x, y, z)example(p, 2)example(p, 5)example(p, 20)

Complexity

The PVQ codebook can be searched in

O(KN)

. Encoding and decoding can likewise be performed in

O(KN)

using

O(K+N)

memory.[5]

The codebook size obeys the recurrence

V(N,K)=V(N-1,K)+V(N,K-1)+V(N-1,K-1)

with

V(N,0)=1

for all

N\ge0

and

V(0,K)=0

for all

K\ne0

.

A closed-form solution is given by[6]

V(N,K)=2N{}2F1(1-K,1-N;2;2).

where

{}2F1

is the hypergeometric function.

See also

Notes and References

  1. Thomas R. . Fischer . A Pyramid Vector Quantizer . . 32 . 4 . 568–583 . July 1986 . 10.1109/TIT.1986.1057198.
  2. Web site: Jean-Marc . Valin . Pyramid Vector Quantization for Video Coding . September 2013 . . April 4, 2021.
  3. Jean-Marc . Valin . Timothy B. . Terriberry . Christopher . Montgomery . Gregory . Maxwell . A High-Quality Speech and Audio Codec With Less Than 10 ms Delay . IEEE Transactions on Audio, Speech, and Language Processing . 18 . 1 . 58–67 . January 2010 . 10.1109/TASL.2009.2023186. 1602.05526 . 11516136 .
  4. Duda . Jarek . 1705.05285 . Improving Pyramid Vector Quantizer with power projection . math.OC . 2017 .
  5. Web site: Timothy B. . Terriberry . cwrs.c . Opus . 2009 . . April 6, 2021.
  6. Web site: Timothy B. . Terriberry . Pulse Vector Coding . December 2007 . . https://web.archive.org/web/20190930134119/https://people.xiph.org/~tterribe/notes/cwrs.html . September 30, 2019 . April 4, 2021.