Multiple sub-Nyquist sampling encoding explained

MUSE (Multiple sub-Nyquist Sampling Encoding),[1] commercially known as Hi-Vision (a contraction of HIgh-definition teleVISION) was a Japanese analog high-definition television system, with design efforts going back to 1979.[2]

It used dot-interlacing and digital video compression to deliver 1125 line, 60 field-per-second (1125i60) signals to the home. The system was standardized as ITU-R recommendation BO.786[3] and specified by SMPTE 260M,[4] using a colorimetry matrix specified by SMPTE 240M.[5] As with other analog systems, not all lines carry visible information. On MUSE there are 1035 active interlaced lines, therefore this system is sometimes also mentioned as 1035i.[6] It employed 2-dimensional filtering, dot-interlacing, motion-vector compensation and line-sequential color encoding with time compression to "fold" an original 20 MHz bandwidth source signal into just 8.1 MHz.

Japan began broadcasting wideband analog HDTV signals in December 1988,[7] initially with an aspect ratio of 2:1. The Sony HDVS high-definition video system was used to create content for the MUSE system.[2] By the time of its commercial launch in 1991, digital HDTV was already under development in the United States. Hi-Vision was mainly broadcast by NHK through their BShi satellite TV channel.

On May 20, 1994, Panasonic released the first MUSE LaserDisc player.[8] There were also a number of players available from other brands like Pioneer and Sony.

Hi-Vision continued broadcasting in analog until 2007.

History

MUSE was developed by NHK Science & Technology Research Laboratories in the 1980s as a compression system for Hi-Vision HDTV signals.

Modulation research

Y

for luminance and

C

for chroma like NTSC and PAL) and uses a phase alternating by line with half-line offset carrier encoding of the wideband/narrowband chroma components. Only the very lowest part of the wideband chroma component overlapped the high-frequency chroma. The narrowband chroma was completely separated from luminance. PAF, or phase alternating by field (like the first NTSC color system trial) was also experimented with, and it gave much better decoding results, but NHK abandoned all composite encoding systems. Because of the use of satellite transmission, Frequency modulation (FM) should be used with power-limitation problem. FM incurs triangular noise, so if a sub-carrierred composite signal is used with FM, demodulated chroma signal has more noise than luminance. Because of this, they looked [12] at other options, and decided to use

Y/C

component emission for satellite. At one point, it seemed that FCFE (Frame Conversion Fineness Enhanced), I/P conversion compression system,[13] would be chosen, but MUSE was ultimately picked.[14]

Y

and

C

components was explored. The MUSE format which is transmitted today, uses separated component signalling. The improvement in picture quality was so great, that the original test systems were recalled.

Technical specifications

MUSE's "1125 lines" are an analog measurement, which includes non-video scan lines taking place while a CRT's electron beam returns to the top of the screen to begin scanning the next field. Only 1035 lines have picture information. Digital signals count only the lines (rows of pixels) that have actual detail, so NTSC's 525 lines become 486i (rounded to 480 to be MPEG compatible), PAL's 625 lines become 576i, and MUSE would be 1035i. To convert the bandwidth of Hi-Vision MUSE into "conventional" lines-of-horizontal resolution (as is used in the NTSC world), multiply 29.9 lines per MHz of bandwidth. (NTSC and PAL/SECAM are 79.9 lines per MHz) - this calculation of 29.9 lines works for all current HD systems including Blu-ray and HD-DVD. So, for MUSE, during a still picture, the lines of resolution would be: 598-lines of luminance resolution per-picture-height. The chroma resolution is: 209-lines. The horizontal luminance measurement approximately matches the vertical resolution of a 1080 interlaced image when the Kell factor and interlace factor are taken into account.

Key features of the MUSE system:

Colorimetry

The MUSE luminance signal

Y

encodes

YM

, specified as the following mix of the original RGB color channels:

\begin{align} YM=0.294R+0.588G+0.118B\end{align}

C

signal encodes

B-YM

and

R-YM

difference signals. By using these three signals (

YM

,

B-YM

and

R-YM

), a MUSE receiver can retrieve the original RGB color components using the following matrix:

\begin{align} \begin{bmatrix}G\B\R\end{bmatrix}  =  \begin{bmatrix} 1&-1/5&-1/2\\ 1&1&0\\ 1&0&1 \end{bmatrix} \begin{bmatrix} 1&0&0\\ 0&5/4&0\\ 0&0&1 \end{bmatrix} \begin{bmatrix} YM\\ B-YM\\ R-YM \end{bmatrix}  =  \begin{bmatrix} 1&-1/4&-1/2\\ 1&5/4&0\\ 1&0&1 \end{bmatrix} \begin{bmatrix}YM\B-YM\R-YM\end{bmatrix} \end{align}

The system used a colorimetry matrix specified by SMPTE 240M[5] [15] [16] (with coefficients corresponding to the SMPTE RP 145 primaries, also known as SMPTE-C, in use at the time the standard was created).[17] The chromaticity of the primary colors and white point are:[16] [5]

MUSE colorimetry (SMPTE 240M / SMPTE "C")
PrimariesCIE 1931 coordinates
xy
Red 0.630 0.340
Green 0.310 0.595
Blue 0.155 0.070
White Point D650.3127 0.3290

The luma (

EY

) function is specified as:[5]

\begin{align} EY=0.212ER+0.701EG+0.087EB\end{align}

The blue color difference (

EPB

) is amplitude-scaled (

EB-EY

), according to:[5]

\begin{align} EPB=1.826(EB-EY)\end{align}

The red color difference (

EPR

) is amplitude-scaled (

ER-EY

), according to:[5]

\begin{align} EPR=1.576(ER-EY)\end{align}

Signal and Transmission

MUSE is a 1125 line system (1035 visible), and is not pulse and sync compatible with the digital 1080 line system used by modern HDTV. Originally, it was a 1125 line, interlaced, 60 Hz, system with a 5/3 (1.66:1) aspect ratio and an optimal viewing distance of roughly 3.3H.

For terrestrial MUSE transmission a bandwidth limited FM system was devised. A satellite transmission system uses uncompressed FM.

The pre-compression bandwidth for

Y

is 20 MHz, and the pre-compression bandwidth for chrominance is a 7.425 MHz carrier.

The Japanese initially explored the idea of frequency modulation of a conventionally constructed composite signal. This would create a signal similar in structure to the

Y/C

composite video NTSC signal - with the

Y

(luminance) at the lower frequencies and the

C

(chrominance) above. Approximately 3 kW of power would be required, in order to get 40 dB of signal to noise ratio for a composite FM signal in the 22 GHz band. This was incompatible with satellite broadcast techniques and bandwidth.

To overcome this limitation, it was decided to use a separate transmission of

Y

and

C

. This reduces the effective frequency range and lowers the required power. Approximately 570 W (360 for

Y

and 210 for

C

) would be needed in order to get a 40 dB of signal to noise ratio for a separate

Y/C

FM signal in the 22 GHz satellite band. This was feasible.

There is one more power saving that appears from the character of the human eye. The lack of visual response to low frequency noise allows significant reduction in transponder power if the higher video frequencies are emphasized prior to modulation at the transmitter and then de-emphasized at the receiver. This method was adopted, with crossover frequencies for the emphasis/de-emphasis at 5.2 MHz for

Y

and 1.6 MHz for

C

. With this in place, the power requirements drop to 260 W of power (190 for

Y

and 69 for

C

).

Sampling systems and ratios

See main article: article and Chroma subsampling. The subsampling in a video system is usually expressed as a three part ratio. The three terms of the ratio are: the number of brightness (luma)

Y

samples, followed by the number of samples of the two color (chroma) components

Cb

and

Cr

, for each complete sample area. Traditionally the value for brightness is always 4, with the rest of the values scaled accordingly.

A sampling of 4:4:4 indicates that all three components are fully sampled. A sampling of 4:2:0, for example, indicated that the two chroma components are sampled at half the horizontal sample rate of luma - the horizontal chroma resolution is halved. This reduces the bandwidth of an uncompressed video signal by one-third.

MUSE implements a similar system as a means of reducing bandwidth, but instead of static sampling, the actual ratio varies according to the amount of motion on the screen. In practice, MUSE sampling will vary from approximately 4:2:1 to 4:0.5:0.25, depending on the amount of movement. Thus the red-green chroma component

Cr

has between one-half and one-eighth the sampling resolution of the luma component

Y

, and the blue-yellow chroma

Cb

has half the resolution of red-green.

Audio subsystem

MUSE had a discrete 2- or 4-channel digital audio system called "DANCE", which stood for Digital Audio Near-instantaneous Compression and Expansion.

It used differential audio transmission (differential pulse-code modulation) that was not psychoacoustics-based like MPEG-1 Layer II. It used a fixed transmission rate of 1350 kbp/s. Like the PAL NICAM stereo system, it used near-instantaneous companding (as opposed to Syllabic-companding like the dbx system uses) and non-linear 13-bit digital encoding at a 32 kHz sample rate.

It could also operate in a 48 kHz 16-bit mode. The DANCE system was well documented in numerous NHK technical papers and in a NHK-published book issued in the USA called Hi-Vision Technology.[18]

The DANCE audio codec was superseded by Dolby AC-3 (a.k.a. Dolby Digital), DTS Coherent Acoustics (a.k.a. DTS Zeta 6x20 or ARTEC), MPEG-1 Layer III (a.k.a. MP3), MPEG-2 Layer I, MPEG-4 AAC and many other audio coders. The methods of this codec are described in the IEEE paper:[19]

Real world performance issues

MUSE had a four-field dot-interlacing cycle, meaning it took four fields to complete a single MUSE frame. Thus, only stationary images were transmitted at full resolution. However, as MUSE lowers the horizontal and vertical resolution of material that varies greatly from frame to frame, moving images were blurred. Because MUSE used motion-compensation, whole camera pans maintained full resolution, but individual moving elements could be reduced to only a quarter of the full frame resolution. Because the mix between motion and non-motion was encoded on a pixel-by-pixel basis, it wasn't as visible as most would think. Later, NHK came up with backwards compatible methods of MUSE encoding/decoding that greatly increased resolution in moving areas of the image as well as increasing the chroma resolution during motion. This so-called MUSE-III system was used for broadcasts starting in 1995 and a very few of the last Hi-Vision MUSE LaserDiscs used it (A River Runs Through It is one Hi-Vision LD that used it). During early demonstrations of the MUSE system, complaints were common about the decoder's large size, which led to the creation of a miniaturized decoder.

Shadows and multipath still plague this analog frequency modulated transmission mode.

Japan has since switched to a digital HDTV system based on ISDB, but the original MUSE-based BS Satellite channel 9 (NHK BS Hi-vision) was broadcast until September 30, 2007.

Cultural and geopolitical impacts

Internal reasons inside Japan that led to the creation of Hi-Vision

MUSE, as the US public came to know it, was initially covered in the magazine Popular Science in the mid-1980s. The US television networks did not provide much coverage of MUSE until the late 1980s, as there were few public demonstrations of the system outside Japan.

Because Japan had its own domestic frequency allocation tables (that were more open to the deployment of MUSE) it became possible for this television system to be transmitted by Ku Band satellite technology by the end of the 1980s.

The US FCC in the late 1980s began to issue directives that would allow MUSE to be tested in the US, providing it could be fit into a 6 MHz System-M channel.

The Europeans (in the form of the European Broadcasting Union (EBU)) were impressed with MUSE, but could never adopt it because it is a 60 Hz TV system, not a 50 Hz system that is standard in Europe and the rest of the world (outside the Americas and Japan).

The EBU development and deployment of B-MAC, D-MAC and much later on HD-MAC were made possible by Hi-Vision's technical success. In many ways MAC transmission systems are better than MUSE because of the total separation of colour from brightness in the time domain within the MAC signal structure.

Like Hi-Vision, HD-MAC could not be transmitted in 8 MHz channels without substantial modification  - and a severe loss of quality and frame rate. A 6 MHz version Hi-Vision was experimented with in the US,[7] but it too had severe quality problems so the FCC never fully sanctioned its use as a domestic terrestrial television transmission standard.

The US ATSC working group that had led to the creation of NTSC in the 1950s was reactivated in the early 1990s because of Hi-Vision's success. Many aspects of the DVB standard are based on work done by the ATSC working group, however most of the impact is in support for 60 Hz (as well as 24 Hz for film transmission) and uniform sampling rates and interoperable screen sizes.

Device support for Hi-Vision

Hi-Vision LaserDiscs

On May 20, 1994, Panasonic released the first MUSE LaserDisc player.[8] There were a number of MUSE LaserDisc players available in Japan: Pioneer HLD-XØ, HLD-X9, HLD-1000, HLD-V500, HLD-V700; Sony HIL-1000, HIL-C1 and HIL-C2EX; the last two of which have OEM versions made by Panasonic, LX-HD10 and LX-HD20. Players also supported standard NTSC LaserDiscs. Hi-Vision LaserDiscs are extremely rare and expensive.

The HDL-5800 Video Disc Recorder recorded both high definition still images and continuous video onto an optical disc and was part of the early analog wideband Sony HDVS high-definition video system which supported the MUSE system. Capable of recording HD still images and video onto either the WHD-3AL0 or the WHD-33A0 optical disc; WHD-3Al0 for CLV mode (up to 10 minute video or 18,000 still frames per side); WHD-33A0 for CAV mode (up to 3 minute video or 5400 still frames per side).

The HDL-2000 was a full band high definition video disc player.

Video cassettes

W-VHS allowed home recording of Hi-Vision programmes.

See also

The analog TV systems these systems were meant to replace:

Related standards:

External links

Notes and References

  1. Web site: DBNSTJ : Realization of High-Definition Television by MUSE System. dbnst.nii.ac.jp.
  2. Book: Cianci, Philip J. . High Definition Television: The Creation, Development and Implementation of HDTV Technology . January 10, 2014 . McFarland . 9780786487974 . Google Books.
  3. Web site: ITU-R BO.786 . MUSE system for HDTV broadcasting-satellite services . 1992 . International Telecommunication Union.
  4. ST 240:1999 - SMPTE Standard - For Television — 1125-Line High-Definition Production Systems — Signal Parameters. https://web.archive.org/web/20220131084402/https://ieeexplore.ieee.org/document/7291461/references#references. dead. January 31, 2022. November 30, 1999. St 240:1999. 1–7. IEEE Xplore. 10.5594/SMPTE.ST240.1999. 978-1-61482-389-6 .
  5. Book: ANSI/SMPTE 240M-1995 - Signal Parameters 1125-Line High-Definition Production Systems . SMPTE . 1995.
  6. Book: Poynton, Charles. Digital Video and HD: Algorithms and Interfaces. January 3, 2003. Elsevier. 9780080504308. Google Books.
  7. Web site: MUSE LaserDisc . 2022-10-19 . ura.caldc.com.
  8. Web site: MUSE HI-DEF LaserDisc Players . LaserDisc UK Web Site . 10 October 2021 . https://web.archive.org/web/20160430061324/http://www.laserdiscarchive.co.uk/laserdisc_archive/muse_high_def_ld/Muse_high_def_ld.htm . 30 April 2016 . dead.
  9. 3. Signal and Transmission Equipment for High-Definition TV. Ishida. Jun-ichi. Yuichi. Ninomiya. December 19, 1982. The Journal of the Institute of Television Engineers of Japan. 36. 10. 882–888. CiNii. 10.3169/itej1978.36.10_882. free.
  10. High-Definition Television System for Future : Desirable Standard, Signal Form and Broadcasting System. Takashi. Fujio. December 19, 1980. ITE Technical Report. 4. 28. 19–24. CiNii. 10.11485/tvtr.4.28_19.
  11. High Definitional Television. Takashi. Fujio. December 19, 1981. The Journal of the Institute of Television Engineers of Japan. 35. 12. 1016–1023. CiNii. 10.3169/itej1978.35.1016. free.
  12. YC Separate Transmission of high Definition Television Signal by BSE. Taro. Komoto. Junichi. Ishida. Masaji. Hata. Keiichi. Yasunaga. December 19, 1979. ITE Technical Report. 3. 26. 61–66. CiNii. 10.11485/tvtr.3.26_61.
  13. High-Definition Television System. Takashi. FUJIO. December 19, 1984. ITE Technical Report. 8. 1. 33–39. CiNii. 10.11485/tvtr.8.1_33.
  14. Rowing a Boat to the HDTV New World. Takashi. FUJIO. August 19, 2006. The Journal of the Institute of Electronics, Information and Communication Engineers. 89. 8. 728–734. CiNii.
  15. Web site: SMPTE-240M Y'PbPr. www5.in.tum.de.
  16. Web site: Detailed Colorspace Descriptions. www.linuxtv.org.
  17. Charles A. Poynton, Digital Video and HDTV: Algorithms and Interfaces, Morgan–Kaufmann, 2003. online
  18. Book: NHK . High Definition Television - Hi Vision Technology . 1993 . 0-442-00798-1.
  19. A single-chip audio signal processor for HDTV receiver. K.. Naganawa. Y.. Hori. S.. Yanase. N.. Itoh. Y.. Asano. August 19, 1991. IEEE Transactions on Consumer Electronics. 37. 3. 677–683. 10.1109/30.85585. 62603128 .