rg chromaticity

The rg chromaticity space, two dimensions of the normalized RGB space,[1] is a chromaticity space, a two-dimensional color space in which there is no intensity information.

In the RGB color space a pixel is identified by the intensity of red, green, and blue primary colors. Therefore a bright red can be represented as (R,G,B) (255,0,0), while a dark red may be (40,0,0). In the normalized rgb space or rg space, a color is represented by the proportion of red, green, and blue in the color, rather than by the intensity of each. Since these proportions must always add up to a total of 1, we are able to quote just the red and green proportions of the color, and can calculate the blue value if necessary.

Conversion between RGB and rg chromaticity

Given a color (R,G,B) where R, G, B = intensity of Red, Green and Blue, this can be converted to color (r,g,b) where  r, g, b imply the proportion of red, green and blue in the original color:[2]

 r = \frac{R}{R+G+B}

 g = \frac{G}{R+G+B}

 b = \frac{B}{R+G+B}

 r+g+b = 1

The sum of rgb will always equal one, because of this property the b dimension can be thrown away without causing any loss in information. The reverse conversion is not possible with only two dimensions, as the intensity information is lost during the conversion to rg chromaticity, e.g. (1/3, 1/3, 1/3) has equal proportions of each color, but it is not possible to determine whether this corresponds to dark gray, light gray, or white. If R, G, B, is normalized to r, g, G color space the conversion can be computed by the following:

 R = \frac{rG}{g}

 G = G

 B =\frac{(1-r-g)G}{g}

The conversion from rgG to RGB, is the same as the conversion from xyY to XYZ.[3] The conversion requires at least some information relative to the intensity of the scene. For this reason if the G is persevered then the inverse is possible.

Pixel-Based Photometric Invariance

Although rg chromaticity contains less information than RGB or HSV color spaces, it has a number of useful properties for computer vision applications. Notably, where a scene viewed by a camera is not lit evenly – for example if lit by a spotlight – then an object of a given color will change in apparent color as it moves across the scene. Where color is being used to track an object in an RGB image, this can cause problems. The lack of intensity information in rg chromaticity images removes this problem, and the apparent color remains constant. Note that in the case where different parts of the image are lit by different colored light sources, problems can still emerge.

Computer vision algorithms tend to suffer from varying imaging conditions. To make more robust computer vision algorithms it is important to use a color invariant color space. Color invariant color spaces are desensitized to disturbances in the image. One common problem in computer vision is varying light source (color and intensity) between multiple images and within a single image.[4] To properly perform image segmentation and object detection requires an increased need for images that are stable to variations in imaging conditions. Normalizing RGB color space to rgb color system performs a linear transform. Normalized rgb space eliminates the effect of varying intensities from the light source. Uniform surfaces of color with varying geometric features are affected by the angle and intensity of the light source. Where a uniform red surface with a uniform green object placed on top, should easily be segmented. Due to the shape of the 3D object shades are formed preventing uniform fields of color. Normalizing intensity out removes the shadow. A lambertian reflector under a white illumination is defined by the equation below:

f^c(x) = m^b(x)\int_{\omega} s(\lambda,x)\rho^c(\lambda)d\lambda

When the r,g,b normalized equations are substituted into the equation above the equations below are derived, which define the invariant properties of the rgb color system.

r = \frac{m^b(x)k_R}{m^b(x)(k_R+k_G+k_B)} = \frac{k_R}{k_R+k_G+k_B}

g = \frac{m^b(x)k_G}{m^b(x)(k_R+k_G+k_B)} = \frac{k_G}{k_R+k_G+k_B}

b = \frac{m^b(x)k_B}{m^b(x)(k_R+k_G+k_B)} = \frac{k_B}{k_R+k_G+k_B}

Where k_c = \int_{\omega}s(\lambda,x)\rho^c(\lambda)d\lambda and c \in {\{R,G,B\}}. The m^b(x) coefficient which denotes the relationship between the white light source and the surface reflectance. This coefficient is cancelled out, assuming a lambertian reflection and white illumination the rgb color space only depends on k_c. The normalized image is free from shadow and shading effects. The rgb color space is dependent on the color of the light source. The color space is only dependent on k_c which is made up of \rho^c(\lambda) and s(\lambda,x), \rho and s are determined by the sensor and surface of object.

Illustration

A photograph with varying illumination levels.
A visual representation of the chromaticity of the image. Each pixel has been scaled so the total red, green, and blue coordinates sum to 1. Notice the effect on the foliage and shadowed regions.
A visual representation of the average value of red, green, and blue coordinates for each pixel in the original image. This information can be combined with the rg chromaticity information to reconstruct the original image.

rg color space

Normalized rg Color Space

r, g, and b chromaticity coordinates are ratios of the one tristimulus value over the sum of all three tristimulus values. A neutral object infers equal values of red, green and blue stimulus. The lack of luminance information in rg prevents having more than 1 neutral point where all three coordinates are of equal value. The white point of the rg chromaticity diagram is defined by the point (1/3,1/3). The white point has one third red, one third green and the final third blue. On an rg chromaticity diagram the first quadrant where all values of r and r are positive forms a right triangle. With max r equals 1 unit along the x and max g equals 1 unit along the y axis. Connecting a line from the max r (1,0) to max g (0,1) from a straight line with slope of negative 1. Any sample that falls on this line has no blue. Moving along the line from max r to max g, shows a decrease in red and an increase of green in the sample, without blue changing. The further a sample moves from this line the more blue is present in the sample trying to be matched.

RGB Color specification System

The CIE 1931 RGB Color matching functions. The color matching functions are the amounts of primaries needed to match the monochromatic test primary at the wavelength shown on the horizontal scale.

RGB is a color mixture system. Once the color matching function are determined the tristimulus values can be determined easily. Since standardization is required to compare results, CIE established standards to determine color matching function.[5]

  1. The reference stimuli must be monochromatic lights R, G, B. With wavelengths \lambda_R=700.0nm, \lambda_G=546.1nm, \lambda_B=435.8nm respectively.
  2. The basic stimulus is white with equal energy spectrum. Require a ratio of 1.000:4.5907:0.0601 (RGB) to match white point.

Therefore a white with equi-energy lights of 1.000 + 4.5907 + 0.0601 = 5.6508 lm can be matched by mixing together R, G and B. Guild and Wright used 17 subjects to determine RGB color matching functions.[6] RGB color matching serve as the base for rg chromaticity. The RGB color matching functions are used to determine the trisitmulus RGB values for a spectrum. Normalizing the RGB tristimulus values converts the tristimulus into rgb. Normalized RGB tristimulus value can be plotted on an rg chromaticity diagram.

An example of color matching function below. [F_{\lambda}] is any monochromatic. Any monochromatic can be matched by adding reference stimuli R[R], G[G] and B[B]. The test light is also to bright to account for this reference stimuli is added to the target to dull the saturation. Thus R is negative. [R], [G] and [B] can be defined as a vector in a three-dimensional space. This three-dimensional space is defined as the color space. Any color [F] can be reached by matching a given amount of [R], [G] and [B].

[F_{\lambda}]+R[R] =G[G]+B[B]

[F_{\lambda}]=-R[R]+ G[G]+B[B]

The negative[R] calls for color matching functions that are negative at certain wavelengths. This is evidence of why the \overline{r} color matching function appears to have negative tristimulus values.

rg Chromaticity Diagram

rg Chromaticity Diagram

The figure to the side is a plotted rg chromaticity diagram. Noting the importance of the E which is defined as the white point where rg are equal and have a value of 1/3 . Next notice the straight line from (0,1) to (1,0), follows the expression y = -x + 1. As the x (red) increases the y (green) decreases by the same amount. Any point on the line represents the limit in rg, and can be defined by a point that has no b information and formed by some combination of r and g. Moving of the linear line towards E represents a decrease in r and g and an increase in b. In computer vision and digital imagery only use the first quadrant because a computer cannot display negative RGB values. The range of RGB is 0-255 for most displays. But when trying to form color matches using real stimuli negative values are needed according to Grassmann’s Laws to match all possible colors. This is why the rg chromaticity diagram extends in the negative r direction.

Conversion xyY color system

Avoid negative color coordinate values prompted the change from to rg to xy. Negative coordinates are used in rg space because when making a spectral sample match can be created by adding stimulus to the sample. The color match- ing functions r, g, and b are negative at certain wavelengths to allow for any monochromatic sample to be matched. This is why in the rg chromaticity diagram the spectral locus extents into the negative r direction and ever so slightly into the negative g direction. On an xy chromaticity diagram the spectral locus if formed by all positive values of x and y.

See also

References

  1. J. B. Martinkauppi and M. Pietikäinen (2005). "Facial Skin Color Modeling". In S. Z. Li and Anil K. Jain. Handbook of face recognition. Springer Science & Business. p. 117. ISBN 978-0-387-40595-7.
  2. W. T. Wintringham (1951). "Color Television and Colorimetry". In D L. MacAdam. Selected Papers on Colorimetry Fundamentals. SPIE - The International Society for Optical Engineering. p. 343. ISBN 0-8194-1296-1.
  3. Lindloom, Bruce (13 March 2009). "xyY to XYZ". http://www.brucelindbloom.com/. Retrieved 7 December 2013. External link in |website= (help)
  4. T. Gevers, A. Gijsenji, J. van de Weijer, and J. Geusebroek (2012). "Pixel-Based Photometric Invariance". In M. A. Kriss. Color in Computer Vision Fundamentals and Applications. Wiley - IS&T Series. p. 50. ISBN 978-0-470-89084-4.
  5. N. Ohto and A. R. Robertson (2005). "CIE Standard Colorimetric System". In M. A. Kriss. Colorimetry Fundamentals and Applications. Wiley - IS&T Series. p. 65. ISBN 978-0-470-09472-3.
  6. R. W. G. Hunt (2004). "The Colour Triangle". In M. A. Kriss. The Reproduction of Colour. Wiley - IS&T Series. p. 71. ISBN 0-470-02425-9.
This article is issued from Wikipedia - version of the Tuesday, April 12, 2016. The text is available under the Creative Commons Attribution/Share Alike but additional terms may apply for the media files.