Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

31
Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW

Transcript of Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

Page 1: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

Multimedia Basics (2)

Hongli luo

Fall 2010CEIT, IPFW

Page 2: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

Topics

Image data type Color Model :

RGB, CMY, CMYK, YUV, YIQ, YCbCr

Analog Video – NTSC, PAL Digital Video

Page 3: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

3.2 Color Model Light is an electromagnetic wave. Its color is

characterized by the wavelength content of the light. Laser light consists of a single wavelength: e.g., a ruby

laser produces a bright, scarlet-red beam. Most light sources produce contributions over many

wavelengths. However, humans cannot detect all light, just

contributions that fall in the “visible wavelengths". Short wavelengths produce a blue sensation, long

wavelengths produce a red one. Visible light is an electromagnetic wave in the range 400

nm to 700 nm (where nm stands for nanometer, 10−9 meters).

Page 4: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

Human Vision

The eye works like a camera, with the lens focusing an image onto the retina (upside-down and left-right reversed).

The retina consists of an array of rods and three kinds of cones. Rods respond to brightness, not sensitive to colors Cones are sensitive to colors

The three kinds of cones are most sensitive to red (R), green (G), and blue (B) light.

The sensitivity of our receptors is also a function of wave-length (Fig. 4.3 below).

Page 5: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

Spectral Sensitivity of the Eye

The eye is most sensitive to light in the middle of the visible spectrum.

The Blue receptor sensitivity is not shown to scale because it is much smaller than the curves for Red or Green – Blue is a late addition, in evolution.

Fig. 4.3 shows the overall sensitivity as a dashed line – this important curve is called the luminous-efficiency function. It is usually denoted and is formed as the sum

of the response curves for Red, Green, and Blue.

Page 6: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

The achromatic channel produced by the cones is approximately proportional to 2R+G+B/20.

Page 7: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

RGB Color Model Color model is made up of Red, Green and Blue

intensity components. RGB color model is used for CRT displays. CRT displays have three phosphors (RGB) which

produce a combination of wavelengths when excited with electrons.

The gamut of colors is all colors that can be reproduced using the three primaries.

Page 8: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

Subtractive Color: CMY Color Model

Additive color – RBG color model when two light beams impinge on a target, their

colors add; when two phosphors on a CRT screen are turned on,

their colors add.

Subtractive color model – CMY color model for ink deposited on paper, the opposite situation

holds: yellow ink subtracts blue from white illumination, but

reflects red and green; it appears yellow. Used in printing

Page 9: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

Instead of red, green, and blue primaries, we need primaries that amount to -red, -green, and -blue. I.e., we need to subtract R, or G, or B.

These subtractive color primaries are Cyan (C), Magenta (M) and Yellow (Y ) inks.

CMY model is mostly used in printing where the color on the paper absorb certain colors.

Page 10: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

Transformation from RGB to CMY

Simplest model we can invent to specify what ink density to lay down on paper, to make a certain desired RGB color

Then the inverse transform is:

Page 11: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

Fig. 4.16: color combinations that result from combining primary colors available in the two situations, additive color and subtractive color.

Page 12: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

Undercolor Removal: CMYK System

Undercolor removal calculate that part of the CMY mix that would be

black, remove it from the color proportions, and add it back as real black.

CMYK color model is made up of Cyan, Magenta, Yellow, Black

Page 13: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

Color Models in Video

Video Color Transforms Largely derive from older analog methods of coding color

for TV. Luminance is separated from color information. YIQ

• YIQ is used to transmit TV signals in North America and Japan.

• This coding also makes its way into VHS video tape coding in these countries since video tape technologies also use YIQ.

YUV• In Europe, video tape uses the PAL or SECAM codings, which

are based on TV that uses a matrix transform called YUV. YCbCr

• Finally, digital video mostly uses a matrix transform called YCbCr that is closely related to YUV

Page 14: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

YUV Color Model

Established in 1982 to build digital video standard

Luminance is separated from color information Y’ – luma,

Luminance signal Y’ = 0.299 R’ + 0.587G’ + 0.114B’

U, V – chrominance The difference between a color and a reference

white at the same luminance U = B’ – Y’, V = R’ – Y’ (4.27)

Page 15: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

YUV Color Model

For gray, R’ = G’ = B’, the luminance Y’ equals to that gray, since 0.299+0.587+0.114 = 1.0. And for a gray (“black and white") image, the chrominance (U, V ) is zero.

Human visual system is more sensitive to luminance than chrominance

YUV represent color with luminance (Y) and chrominance (U,V) Similarly YIQ, YCrCb

Advantages We can subsample the chrominance

Page 16: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.
Page 17: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

YIQ Color Model

YIQ is used in NTSC color TV broadcasting. Again, gray pixels generate zero (I,Q) chrominance signal. I and Q are a rotated version of U and V . Y ‘ in YIQ is the same as in YUV; U and V are rotated

by 330

This leads to the following matrix transform:

Page 18: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.
Page 19: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

3.3 Analog Video

An analog signal f(t) samples a time-varying image. So-called “progressive” scanning traces through a complete picture (a frame) row-wise for each time interval.

In TV, and in some monitors and multimedia standards as well, another system, called “interlaced” scanning is used: The odd-numbered lines are traced first, and then

the even-numbered lines are traced. This results in “odd” and “even” fields - two fields make up one frame.

In fact, the odd lines (starting from 1) end up at the middle of a line at the end of the odd field, and the even scan starts at a half-way point.

Page 20: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

Figure 5.1 shows the scheme used. First the solid (odd) lines are traced, P to Q, then R to S, etc., ending at T; then the even field starts at U and ends at V.

It was difficult to transmit the amount of information in a full frame quickly enough to avoid flicker.

Page 21: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

Analog Video

NTSC (National Television System Committee) Analog television system in the United States,

Canada, Japan, etc 4:3 aspect ration (ratio of picture width to its height) 525 scan lines per frame, 30 frames per second, Interlaced, each frame is divided into two fields NTSC uses YIQ color model

• 6 MHz channel• Assign a bandwidth of 4.2 MHz to Y , and only 1.6 MHz

to I and 0.6 MHz to Q • due to humans‘ insensitivity to color details.

Page 22: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.
Page 23: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

Analog Video

PAL (Phase Alternating Line) TV standard widely used in Western Europe, China,

India, and many other parts of the world 4:3 aspect ratio 625 scan lines per frame, 25 frames per second

(40msec/frame) Interlaced, each frame is divided into 2 fields PAL uses YUV color model

• 8 MHz channel• Allocates a bandwidth of 5.5 MHz to Y, and 1.8 MHz

each to U and V

Page 24: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

Table 5.2 gives a comparison of the three major analog broadcast TV systems.

SECAM is similar to PAL

Page 25: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

3.4 Digital Video

The advantages of digital representation for video are many.

For example: Video can be stored on digital devices or in memory,

ready to be processed (noise removal, cut and paste, etc.), and integrated to various multimedia applications; Direct access is possible, which makes nonlinear video

editing achievable as a simple, rather than a complex task;

Repeated recording does not degrade image quality; Ease of encryption and better tolerance to channel noise.

The usual color space is YCbCr.

Page 26: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

Chroma Subsampling

Stores color information at lower resolution than intensity information

The purpose is compression – reduce the bit rates Human visual system is more sensitive to variations

in brightness than color. Assign more bandwidth to Y than the color difference

components • Human visual system is less sensitive to the position

and motion of color than luminance• Bandwidth can be optimized by storing more luminance

detail than color detail Reduction results in almost no perceivable visual

difference

Page 27: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

Chroma Subsampling

Numbers are given stating how many pixel values, per four original pixels, are actually sent: The chroma subsampling scheme 4:4:4 indicates that no

chroma subsampling is used: each pixel's Y, Cb and Cr values are sampled, 4 for each of Y, Cb, Cr.

The scheme 4:2:2 indicates horizontal subsampling of the Cb, Cr signals by a factor of 2. That is, of four pixels horizontally labelled as 0 to 3, all four Ys are sampled, and every two Cb's and two Cr's are sampled, as (Cb0, Y0)(Cr0, Y1)(Cb2, Y2)(Cr2, Y3)(Cb4, Y4), and so on (or averaging is used).

The scheme 4:1:1 subsamples horizontally by a factor of 4. The scheme 4:2:0 subsamples in both the horizontal and

vertical dimensions by a factor of 2. Theoretically, an average chroma pixel is positioned between the rows and columns as shown Fig.5.6.

Scheme 4:2:0 along with other schemes is commonly used in JPEG and MPEG

Page 28: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.
Page 29: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

CCIR Standards for Digital Video

CCIR is the Consultative Committee for International Radio, and one of the most important standards it has produced is CCIR-601, for component digital video. This standard has since become standard ITU-R-601,

an international standard for professional video applications

• adopted by certain digital video formats including the popular DV video.

Table 5.3 shows some of the digital video specifications, all with an aspect ratio of 4:3.

Page 30: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.

CIF and QCIF

CIF stands for Common Intermediate Format specified by the CCITT. The idea of CIF is to specify a format for lower

bitrate. CIF is about the same as VHS quality. It uses a

progressive (non-interlaced) scan.

QCIF stands for “Quarter-CIF”. All the CIF/QCIF resolutions are evenly divisible by 8, and all except 88 are divisible by 16; this provides convenience for block-based video coding in H.261 and H.263, H.264

Page 31: Multimedia Basics (2) Hongli luo Fall 2010 CEIT, IPFW.