Colour difference coding in computing (unfinished)

Billy Biggs <vektor@dumbterm.net>


This paper attempts to explain the different standards that exist for digital images in stored as colour difference values: Y'CBCR or YUV formats. We introduce the standards involved, give a technical background to these colourspaces, and then look at how the standards specify encoded images. We then present case studies and how we believe conversions should be performed.

1 Standards

Video is standardized by locality: each region of the world uses either a variant of NTSC, or a variant of PAL. Broadcasters standardize themselves even more specifically by the properties of the equipment they use and support.

In computing, we are faced with the challenge of writing software that can handle digital recordings and captured images from all possible TV and digital video standards, and then modify our output to be displayed on any of these plus the many variants of PC CRTs or LCD screen. To deal with this effectively, we must be aware of the many standards in place and how to recognize them.

Below is a list of important standards for understanding digital images. I describe them briefly here so I can refer to them later.

You can order copies of SMPTE standards from the SMPTE website, or maybe find them at your local library. ITU standards can also be ordered from their website, but my university library had copies of at least the original CCIR specifications.

2.1 ITU-R BT.470

The official title is "Conventional television systems", and it documents the 'standard' for composite video systems in use around the world: variants of NTSC, PAL and SECAM. This standard makes some interesting claims about video equipment calibration:

  1. Claims that while no phosphors exist that are close to the NTSC chromaticities, the United States of America recommends that equipment perform corrections on the signals to compensate.
  2. It notes that NTSC specifies Illuminant C as the white point, but that studio monitors are adjusted to a reference white of D65.
  3. It claims that studio monitors in Japan are set to a white temperature of 9300K.

2.2 SMPTE 170M

"This standard reflects modern practice in the generation of NTSC signals and, in some respects, differs from the original NTSC specification published in 1953." I have the 1999 revision of the 1994 standard.

2.3 ITU-R BT.601 (aka CCIR 601)

This standardizes digital sampling for television: coefficients and excursion for Y'CBCR but no chromaticities.

2.4 SMPTE 240M

This standard is for 1125-line HDTV systems. I have the 1999 revision of the 1995 standard.

2.5 ITU-R BT.709

2.6 JFIF JPEG images

2.7 MPEG1

2 Background

This section explains the aspects of digital image coding which lead to YUV encodings.

2.1 Colour perception and image coding

Human perception allows for about three degrees of freedom when representing colour. However, the eye is most sensitive to the area of the spectrum roughly in the center. This curve is called the luminous efficiency curve, or luminance, and measuring the luminance of an image can demonstrate how bright it will appear.


Luminous efficiency V(lambda)

Between 60 and 70 percent of the luminance signal comprises green information. In image coding, we take advantage of this efficiency, function by ensuring that this data is sampled at a high rate. The simplest way to do this is to "remove" the luminance information from the blue and green signals to form a pair of colour difference components (Poynton).

2.2 Gamma and image coding

The output of a CRT is not a linear function of the applied signal. Instead, a CRT has a power-law response to voltage: the luminance produced at the face of the display is approximately proportional to the applied voltage raised to a power known as gamma. In order to preduct the luminance of the output, we must consider this gamma function when storing our colours.

For image coding we can use gamma to our advantage. If images are stored as gamma-corrected values, instead of luminance values, then we don't need to do a lossy conversion when we output. For this reason, digital images in RGB format are almost always stored as gamma corrected R'G'B' values.

2.3 Chromaticities and RGB spaces

Monitors and other display devices usually use three primaries to create colours: red, green and blue. Over the years there have been different standards for what exact colour of red, green and blue should be used, which has changed to adapt to changes in phosphor technology and availablility.

3 Colour difference coding

3.1 Encoding pipeline

For compression, the final encoding pipeline is:

  1. The source image is sampled to an RGB space with chromaticities matching the intended display device.
  2. The image is gamma corrected for the gamma of the intended display device to create an R'G'B' image.
  3. For each pixel, we transform to get the luma, the gamma-corrected estimation of the luminance signal Y'. Using this value, we obtain the difference signals R'-Y' and B'-Y'.
  4. The result is quantized to the desired format.

4 Differences in encoding

Based on the above encoding pipeline, we can see where possibilities there are for differences in our output:

  1. Differences in RGB chromaticities
  2. Differences in gamma function
  3. Differences in conversion coefficients
  4. Differences in quantization levels

4.1 Differences in RGB chromaticities

Standard red (x,y) green (x,y) blue (x,y)
ITU-R BT.709 (0.640, 0.330) (0.300, 0.600) (0.150, 0.060)
ITU-R BT.470-2 System M (NTSC/USA) (0.67, 0.33) (0.21, 0.71) (0.14, 0.08)
ITU-R BT.470-2 System B, G (PAL) (0.64, 0.33) (0.29, 0.60) (0.15, 0.06)
SMPTE 240M (1987)
also SMPTE 170M
and SMPTE RP 145-1994
(0.630, 0.340) (0.310, 0.595) (0.155, 0.070)
Name white (x,y) Used in
D65 (0.3127, 0.3290) ITU-R BT.709
and SMPTE 170M
and SMPTE 240M
and ITU-R BT.470-2 System B, G (PAL)
White C (0.310, 0.316) ITU-R BT.470-2 System M (NTSC/USA)

4.2 Differences in gamma function

Standard Transfer function
ITU-R BT.709
and SMPTE 170M
and sRGB
SMPTE 240M (1987)

4.3 Differences in conversion coefficients

Standard Luma function
ITU-R BT.601
ITU-R BT.709
SMPTE 240M

4.4 Differences in quantization levels

In all of the ITU and SMPTE specifications, and in the MPEG2 spec, the analogue 0-1 values are converted to 8-bit components using the following formulae:

This gives excursions of 16-235 for Y', and 16-240 for both CB and CR.

However, the JFIF standard for JPEG specifies the following conversion:

To quote: where the E'y, E'Cb and E'Cb are defined as in CCIR 601. Since values of E'y have a range of 0 to 1.0 and those for E'Cb and E'Cr have a range of -0.5 to +0.5, Y, Cb, and Cr must be clamped to 255 when they are maximum value..

This bugs me: why not use 255 in the above equation, instead of losing the last quantization level?

5 References

  1. JFIF standard for common JPEG images
  2. Gamma FAQ by Charles Poynton
  3. Color FAQ by Charles Poynton
[ICO]NameLast modifiedSizeDescription

[PARENTDIR]Parent Directory   -  
[   ]Chroma_subsampling_n..>2003-01-05 17:35 163K 
[IMG]conversion-jpeg.png 2002-05-13 20:03 1.9K 
[IMG]conversion.png 2002-05-10 14:09 2.0K 
[IMG]gamma240m.png 2002-05-13 20:46 2.3K 
[IMG]gamma709.png 2002-05-10 13:46 2.3K 
[   ]jfif.txt.gz 1992-09-09 07:26 4.5K 
[IMG]luma240m.png 2002-05-13 20:08 1.4K 
[IMG]luma601.png 2002-05-13 20:08 1.4K 
[IMG]luma709.png 2002-05-13 20:08 1.6K 
[IMG]luminance.png 2002-05-10 08:47 1.9K