Pixel

In digital imaging, a pixel, or pel,[1] (picture element[2]) is a single point in a raster image, or the smallest addressable screen element in a display device; it is the smallest unit of picture that can be represented or controlled.

Each pixel has its own address. The address of a pixel corresponds to its coordinates. Pixels are normally arranged in a two-dimensional grid, and are often represented using dots or squares. Each pixel is a sample of an original image; more samples typically provide more accurate representations of the original. The intensity of each pixel is variable. In color image systems, a color is typically represented by three or four component intensities such as red, green, and blue, or cyan, magenta, yellow, and black.

In some contexts (such as descriptions of camera sensors), the term pixel is used to refer to a single scalar element of a multi-component representation (more precisely called a photosite in the camera sensor context, although the neologism sensel is sometimes used to describe the elements of a digital camera's sensor),[3] while in others the term may refer to the entire set of such component intensities for a spatial position. In color systems that use chroma subsampling, the multi-component concept of a pixel can become difficult to apply, since the intensity measures for the different color components correspond to different spatial areas in a such a representation.

The word pixel is based on a contraction of pix ("pictures") and el (for "element"); similar formations with el  for "element" include the words voxel[4] and texel.[4]

Contents

Etymology

The word "pixel" was first published in 1965 by Frederic C. Billingsley of JPL, to describe the picture elements of video images from space probes to the Moon and Mars. However, Billingsley did not coin the term himself. Instead, he got the word "pixel" from Keith E. McFarland, at the Link Division of General Precision in Palo Alto, who did not know where the word originated. McFarland said simply it was "in use at the time" (circa 1963).[5]

The word is a combination of picture and element, via pix. The word pix appeared in Variety magazine headlines in 1932, as an abbreviation for the word pictures, in reference to movies.[6] By 1938, "pix" was being used in reference to still pictures by photojournalists.[5]

The concept of a "picture element" dates to the earliest days of television, for example as "Bildpunkt" (the German word for pixel, literally 'picture point') in the 1888 German patent of Paul Nipkow. According to various etymologies, the earliest publication of the term picture element itself was in Wireless World magazine in 1927,[7] though it had been used earlier in various U.S. patents filed as early as 1911.[8]

Some authors explain pixel as picture cell, as early as 1972.[9] In video processing, pel is often used instead of pixel.[10] For example, IBM used it in their Technical Reference for the original PC.

Words with similar etymologies

Technical

A pixel is generally thought of as the smallest single component of a digital image. However, the definition is highly context-sensitive. For example, there can be "printed pixels" in a page, or pixels carried by electronic signals, or represented by digital values, or pixels on a display device, or pixels in a digital camera (photosensor elements). This list is not exhaustive, and depending on context, there are several terms that are synonymous in particular contexts, such as pel, sample, byte, bit, dot, spot, etc. The term "pixels" can be used in the abstract, or as a unit of measure, in particular when using pixels as a measure of resolution, such as: 2400 pixels per inch, 640 pixels per line, or spaced 10 pixels apart.

The measures dots per inch (dpi) and pixels per inch (ppi) are sometimes used interchangeably, but have distinct meanings, especially for printer devices, where dpi is a measure of the printer's density of dot (e.g. ink droplet) placement.[11] For example, a high-quality photographic image may be printed with 600 ppi on a 1200 dpi inkjet printer.[12] Even higher dpi numbers, such as the 4800 dpi quoted by printer manufacturers since 2002, do not mean much in terms of achievable resolution.[13]

The more pixels used to represent an image, the closer the result can resemble the original. The number of pixels in an image is sometimes called the resolution, though resolution has a more specific definition. Pixel counts can be expressed as a single number, as in a "three-megapixel" digital camera, which has a nominal three million pixels, or as a pair of numbers, as in a "640 by 480 display", which has 640 pixels from side to side and 480 from top to bottom (as in a VGA display), and therefore has a total number of 640 × 480 = 307,200 pixels or 0.3 megapixels.

The pixels, or color samples, that form a digitized image (such as a JPEG file used on a web page) may or may not be in one-to-one correspondence with screen pixels, depending on how a computer displays an image. In computing, an image composed of pixels is known as a bitmapped image or a raster image. The word raster originates from television scanning patterns, and has been widely used to describe similar halftone printing and storage techniques.

Sampling patterns

For convenience, pixels are normally arranged in a regular two-dimensional grid. By using this arrangement, many common operations can be implemented by uniformly applying the same operation to each pixel independently. Other arrangements of pixels are possible, with some sampling patterns even changing the shape (or kernel) of each pixel across the image. For this reason, care must be taken when acquiring an image on one device and displaying it on another, or when converting image data from one pixel format to another.

For example:

Resolution of computer monitors

Computers can use pixels to display an image, often an abstract image that represents a GUI. The resolution of this image is called the display resolution and is determined by the video card of the computer. LCD monitors also use pixels to display an image, and have a native resolution. Each pixel is made up of triads, with the number of these triads determining the native resolution. On some CRT monitors, the beam sweep rate may be fixed, resulting in a fixed native resolution. Most CRT monitors do not have a fixed beam sweep rate, meaning they do not have a native resolution at all - instead they have a set of resolutions that are equally well supported. To produce the sharpest images possible on an LCD, the user must ensure the display resolution of the computer matches the native resolution of the monitor.

Bits per pixel

The number of distinct colors that can be represented by a pixel depends on the number of bits per pixel (bpp). A 1 bpp image uses 1-bit for each pixel, so each pixel can be either on or off. Each additional bit doubles the number of colors available, so a 2 bpp image can have 4 colors, and a 3 bpp image can have 8 colors:

  • 1 bpp, 21 = 2 colors (monochrome)
  • 2 bpp, 22 = 4 colors
  • 3 bpp, 23 = 8 colors
...
  • 8 bpp, 28 = 256 colors
  • 16 bpp, 216 = 65,536 colors ("Highcolor" )
  • 24 bpp, 224 ≈ 16.8 million colors ("Truecolor")

For color depths of 15 or more bits per pixel, the depth is normally the sum of the bits allocated to each of the red, green, and blue components. Highcolor, usually meaning 16 bpp, normally has five bits for red and blue, and six bits for green, as the human eye is more sensitive to errors in green than in the other two primary colors. For applications involving transparency, the 16 bits may be divided into five bits each of red, green, and blue, with one bit left for transparency. A 24-bit depth allows 8 bits per component. On some systems, 32-bit depth is available: this means that each 24-bit pixel has an extra 8 bits to describe its opacity (for purposes of combining with another image).

Subpixels

Many display and image-acquisition systems are, for various reasons, not capable of displaying or sensing the different color channels at the same site. Therefore, the pixel grid is divided into single-color regions that contribute to the displayed or sensed color when viewed at a distance. In some displays, such as LCD, LED, and plasma displays, these single-color regions are separately addressable elements, which have come to be known as subpixels. For example, LCDs typically divide each pixel horizontally into three subpixels. When the square pixel is divided into three subpixels, each subpixel is necessarily rectangular. In the display industry terminology, subpixels are often referred to as pixels, as they are the basic addressable elements in a viewpoint of hardware, and they call pixel circuits rather than subpixel circuits.

Most digital camera image sensors use single-color sensor regions, for example using the Bayer filter pattern, and in the camera industry these are known as pixels just like in the display industry, not subpixels.

For systems with subpixels, two different approaches can be taken:

This latter approach, referred to as subpixel rendering, uses knowledge of pixel geometry to manipulate the three colored subpixels separately, producing a slight increase in the apparent resolution of color displays. While CRT displays use red-green-blue-masked phosphor areas, dictated by a mesh grid called the shadow mask, it would require a difficult calibration step to be aligned with the displayed pixel raster, and so CRTs do not currently use subpixel rendering.

The concept of subpixels is related to samples.

Megapixel

A megapixel (MP or Mpx) is one million pixels, and is a term used not only for the number of pixels in an image, but to express the number of image sensor elements of digital cameras or the number of display elements of digital displays. For example, a camera with an array of 2048 × 1536 sensor elements is commonly said to have "3.1 megapixels" (2048 × 1536 = 3,145,728).

Digital cameras use photosensitive electronics, either charge-coupled device (CCD) or complementary metal–oxide–semiconductor (CMOS) image sensors, consisting of a large number of single sensor elements, each of which records a measured intensity level. In most digital cameras, the sensor array is covered with a patterned color filter mosaic having red, green, and blue regions in the Bayer filter arrangement, so that each sensor element can record the intensity of a single primary color of light. The camera interpolates the color information of neighboring sensor elements, through a process called demosaicing, to create the final image. These sensor elements are often called "pixels", even though they only record 1 channel (only red, or green, or blue) of the final color image. Thus, two of the three color channels for each sensor must be interpolated and a so-called N-megapixel camera that produces an N-megapixel image provides only one-third of the information that an image of the same size could get from a scanner. Thus, certain color contrasts may look fuzzier than others, depending on the allocation of the primary colors (green has twice as many elements as red or blue in the Bayer arrangement).

Display resolution

The display resolution of a display device (such as a digital television) is the number of distinct pixels in each dimension that can be displayed. It can be an ambiguous term especially as the displayed resolution is controlled by different factors in cathode ray tube (CRT) and flat panel or projection displays using fixed-element arrays.

See also

References

  1. ^ J.D Foley and A. Van Dam, Fundamentals of Interactive Computer Graphics, 1982
  2. ^ Rudolf F. Graf (1999). Modern Dictionary of Electronics. Oxford: Newnes. p. 569. ISBN 0-7506-43315. http://books.google.com/?id=o2I1JWPpdusC&pg=PA569. 
  3. ^ Michael Goesele (2004). New Acquisition Techniques for Real Objects and Light Sources in Computer Graphics. Books on Demand. ISBN 3833414898. http://books.google.com/?id=ZTJJ8QzNv1wC&pg=PA33&dq=sensel+sensor-element. 
  4. ^ a b Foley, James D.; Andries van Dam, John F. Hughes, Steven K. Feiner (1990). "Spatial-partitioning representations; Surface detail". Computer Graphics: Principles and Practice. The Systems Programming Series. Addison-Wesley. ISBN 0-201-12110-7. "These cells are often called voxels (volume elements), in analogy to pixels." 
  5. ^ a b Lyon, Richard F. (2006). A brief history of 'pixel'. IS&T/SPIE Symposium on Electronic Imaging.
  6. ^ "Online Etymology Dictionary". http://www.etymonline.com/index.php?search=pixel. 
  7. ^ "On language; Modem, I'm Odem", The New York Times, April 2, 1995. Accessed April 7, 2008.
  8. ^ Sinding-Larsen, Alf Transmission of Pictures of Moving Objects, US Patent 1,175,313, issued March 14, 1916.
  9. ^ Robert L. Lillestrand (1972). "Techniques for Change Detection". IEEE Trans. Computers C-21 (7). 
  10. ^ Lewis, Peter H. (February 12, 1989). The Executive Computer; Compaq Sharpens Its Video Option. The New York Times.
  11. ^ Derek Doeffinger (2005). The Magic of Digital Printing. Lark Books. p. 24. ISBN 1579906893. http://books.google.com/?id=s2hIx1amJUcC&dq=printer+dots-per-inch+pixels-per-inch. 
  12. ^ "Experiments with Pixels Per Inch (PPI) on Printed Image Sharpness". ClarkVision.com. July 3, 2005. http://www.clarkvision.com/imagedetail/printer-ppi/. 
  13. ^ Harald Johnson (2002). Mastering Digital Printing. Thomson Course Technology. p. 40. ISBN 1929685653. http://books.google.com/?id=wto19gxFyfQC&dq=inkjet+printer+4800+dpi+addressability. 
  14. ^ "Image registration of blurred satellite images". http://staff.utia.cas.cz/zitova/registration.htm. Retrieved 2008-05-09. 
  15. ^ "ScienceDirect - Pattern Recognition: Image representation by a new optimal non-uniform morphological sampling:". http://www.sciencedirect.com/science/article/B6V14-3YXB1T5-8/2/15b559f2aefbc7bdc54c55df148c2374. Retrieved 2008-05-09. 

External links