Video


Video is an electronic medium used for the recording, copying, playback, transmission, and display of moving visual images, with or without accompanying audio. Video technology was initially developed for live transmission and later expanded to include recording and storage through analog formats such as magnetic tape. Since the late 20th century, digital video has become the dominant form, enabling efficient compression, storage, editing, and distribution across broadcast television, physical media, and internet-based platforms. Advances in digital imaging, compression standards, and network infrastructure have significantly influenced media production, communication, entertainment, education, and information dissemination worldwide.
Video systems vary in display resolution, aspect ratio, refresh rate, color reproduction, and other qualities. Both analog and digital video can be carried on a variety of media, including radio, magnetic tape, optical discs, computer files, and network streaming.

Etymology

The word video comes from the Latin video, "I see," the first-person singular present indicative of videre, "to see".

History

Analog video

Video developed from facsimile systems developed in the mid-19th century. Mechanical video scanners, such as the Nipkow disk, were patented as early as 1884, but it took several decades before practical video systems could be developed. Whereas the medium of film records using a sequence of miniature photographic images visible to the naked eye, video encodes images electronically, turning them into analog or digital electronic signals for transmission and recording.
Video was originally exclusively live technology, and was first developed for mechanical television systems. These were quickly replaced by cathode-ray tube television systems. Live video cameras used an electron beam, which would scan a photoconductive plate with the desired image and produce a voltage signal proportional to the brightness in each part of the image. The signal could then be sent to televisions, where another beam would receive and display the image. Charles Ginsburg led an Ampex research team to develop one of the first practical video tape recorders. In 1951, the first of these captured live images from television cameras by writing the camera's electrical signal onto magnetic videotape. VTRs sold for around US$50,000 in 1956, and videotapes cost US$300 per one-hour reel. However, prices gradually dropped over the years, and in 1971, Sony began selling videocassette recorder decks and tapes into the consumer market.

Digital video

Digital video is capable of higher quality and, eventually, a much lower cost than its analog predecessor. After the commercial introduction of the DVD, in 1997, and later the Blu-ray Disc, in 2006, sales of videotape and recording equipment fell. Advances in computer technology allow even inexpensive personal computers and smartphones to capture, store, edit, and transmit digital video, further reducing the cost of video production and allowing programmers and broadcasters to move to tapeless production. The advent of digital broadcasting and the subsequent digital television transition are in the process of relegating analog video to the status of a legacy technology in most parts of the world. The development of high-resolution video cameras with improved dynamic range and broader color gamuts, along with the introduction of high-dynamic-range digital intermediate data formats with improved color depth, has caused digital video technology to converge with film technology. Since 2013, the use of digital cameras in Hollywood has surpassed the use of film cameras.

Characteristics

Frame rate

Frame rate—the number of still pictures per unit of time—ranges from six or eight frames per second for older mechanical cameras to 120 or more for new professional cameras. The PAL and SECAM standards specify 25 fps, while NTSC specifies 29.97 fps. Film is shot at a slower frame rate of 24 frames per second, which slightly complicates the process of transferring film to video. The minimum frame rate to achieve persistence of vision is about 16 frames per second.

Interlacing vs. progressive-scan systems

Video can be interlaced or progressive. In progressive scan systems, each refresh period updates all scan lines in each frame, in sequence. When displaying a natively progressive broadcast or recorded signal, the result is the optimum spatial resolution of both the stationary and moving parts of the image. Interlacing was invented as a way to reduce flicker in early mechanical and CRT video displays, without increasing the number of complete frames per second. Interlacing retains detail while requiring lower bandwidth compared to progressive scanning.
In interlaced video, the horizontal scan lines of each complete frame are treated as if numbered consecutively and captured as two fields: an odd field consisting of the odd-numbered lines and an even field consisting of the even-numbered lines. Analog display devices reproduce each frame, effectively doubling the frame rate as far as perceptible overall flicker is concerned. When the image capture device acquires the fields one at a time, rather than dividing up a complete frame after it is captured, the frame rate for motion is effectively doubled as well, resulting in smoother, more lifelike reproduction of rapidly moving parts of the image when viewed on an interlaced CRT display.
NTSC, PAL, and SECAM are interlaced formats. In video resolution notation, 'i' denotes interlaced scanning. For example, PAL video format is often described as 576i50, where 576 indicates the total number of horizontal scan lines, i indicates interlacing, and 50 indicates 50 fields per second.
When displaying a natively interlaced signal on a progressive scan device, the overall spatial resolution is degraded by simple line doubling—artifacts, such as flickering or comb effects in moving parts of the image, appear unless special signal processing eliminates them. A procedure known as deinterlacing can optimize the display of an interlaced video signal from an analog, DVD, or satellite source on a progressive scan device such as an LCD television, digital video projector, or plasma panel. Deinterlacing cannot, however, produce video quality that is equivalent to true progressive scan source material.

Aspect ratio

In video, an aspect ratio is the proportional relationship between the width and height of a video screen and video picture elements. All popular video formats are landscape, with a traditional television screen having an aspect ratio of 4:3, or about 1.33:1. High-definition televisions have an aspect ratio of 16:9, or about 1.78:1. The ratio of a full 35mm film frame with its sound track is 1.375:1.
Pixels on computer monitors are usually square, but pixels used in digital video often have non-square aspect ratios, such as those used in the PAL and NTSC variants of the CCIR 601 digital video standard and the corresponding anamorphic widescreen formats. The 720 by 480 pixel raster uses thin pixels on a 4:3 aspect ratio display and fat pixels on a 16:9 display.
The popularity of video on mobile phones has led to the growth of vertical video. Mary Meeker, a partner at Silicon Valley venture capital firm Kleiner Perkins Caufield & Byers, highlighted the growth of vertical-video viewing in her 2015 Internet Trends Report, noting that it had grown from 5% of viewing in 2010 to 29% in 2015. Vertical-video ads are watched in their entirety nine times more frequently than those in landscape ratios.

Color model and depth

The color model uses the video color representation and maps encoded color values to visible colors reproduced by the system. There are several such representations in common use: typically, YIQ is used in NTSC television, YUV is used in PAL television, YDbDr is used by SECAM television, and YCbCr is used for digital video.
The number of distinct colors a pixel can represent depends on the color depth expressed in the number of bits per pixel. A common way to reduce the amount of data required in digital video is by chroma subsampling. Because the human eye is less sensitive to details in color than brightness, the luminance data for all pixels is maintained, while the chrominance data is averaged for a number of pixels in a block, and the same value is used for all of them. For example, this results in a 50% reduction in chrominance data using 2-pixel blocks or 75% using 4-pixel blocks. This process does not reduce the number of possible color values that can be displayed, but it reduces the number of distinct points at which the color changes.

Quality

can be measured with formal metrics like peak signal-to-noise ratio or through subjective video quality assessment using expert observation. Many subjective video quality methods are described in the ITU-T recommendation BT.500. One of the standardized methods is the Double Stimulus Impairment Scale. In DSIS, each expert views an unimpaired reference video, followed by an impaired version of the same video. The expert then rates the impaired video using a scale ranging from "impairments are imperceptible" to "impairments are very annoying."

Compression (digital only)

delivers maximum quality, but at a very high data rate. A variety of methods are used to compress video streams, with the most effective ones using a group of pictures to reduce spatial and temporal redundancy. Broadly speaking, spatial redundancy is reduced by registering differences between parts of a single frame; this task is known as intraframe compression and is closely related to image compression. Likewise, temporal redundancy can be reduced by registering differences between frames; this task is known as interframe compression, including motion compensation and other techniques. The most common modern compression standards are MPEG-2, used for DVD, Blu-ray, and satellite television, and MPEG-4, used for AVCHD, mobile phones, and the Internet.