This article is about the storage and reproduction medium. For their content and production, seeFilm andFootage. For other uses, seeVideo (disambiguation).
A one-minute animated video showing an example of a media production process
Video is an electronic medium used for the recording, copying, playback, transmission, and display of moving visual images, with or without accompanying audio. Video technology was initially developed for live transmission and later expanded to include recording and storage through analog formats such as magnetic tape. Since the late 20th century, digital video has become the dominant form, enabling efficient compression, storage, editing, and distribution across broadcast television, physical media, and internet-based platforms. Advances in digital imaging, compression standards, and network infrastructure have significantly influenced media production, communication, entertainment, education, and information dissemination worldwide.
Video developed from facsimile systems developed in the mid-19th century. Mechanical video scanners, such as theNipkow disk, werepatented as early as 1884, but it took several decades before practical video systems could be developed. Whereas the medium offilm records using a sequence of miniature photographic images visible to the naked eye, video encodes images electronically, turning them into analog or digital electronic signals for transmission and recording.[2]
Video was originally exclusivelylive technology, and was first developed formechanical television systems. These were quickly replaced bycathode-ray tube (CRT)television systems. Live video cameras used an electron beam, which would scan a photoconductive plate with the desired image and produce a voltage signal proportional to the brightness in each part of the image. The signal could then be sent to televisions, where another beam would receive and display the image.[3]Charles Ginsburg led anAmpex research team to develop one of the first practicalvideo tape recorders (VTR). In 1951, the first of these captured live images fromtelevision cameras by writing the camera's electrical signal onto magneticvideotape. VTRs sold for around US$50,000 in 1956, and videotapes cost US$300 per one-hour reel.[4] However, prices gradually dropped over the years, and in 1971, Sony began sellingvideocassette recorder (VCR) decks and tapes into theconsumer market.[5]
Digital video is capable of higher quality and, eventually, a much lower cost than its analog predecessor. After the commercial introduction of theDVD, in 1997, and later theBlu-ray Disc, in 2006, sales of videotape and recording equipment fell. Advances incomputer technology allow even inexpensivepersonal computers andsmartphones to capture, store, edit, and transmit digital video, further reducing the cost ofvideo production and allowing programmers and broadcasters to move totapeless production. The advent ofdigital broadcasting and the subsequentdigital television transition are in the process of relegating analog video to the status of alegacy technology in most parts of the world. The development of high-resolution video cameras with improveddynamic range and broadercolor gamuts, along with the introduction of high-dynamic-rangedigital intermediate data formats with improvedcolor depth, has caused digital video technology to converge with film technology. Since 2013, the use ofdigital cameras inHollywood has surpassed the use of film cameras.[6]
Frame rate—the number of still pictures per unit of time—ranges from six or eight frames per second (frame/s orfps) for older mechanical cameras to 120 or more for new professional cameras. ThePAL andSECAM standards specify 25 fps, whileNTSC specifies 29.97 fps.[7] Film is shot at a slower frame rate of 24 frames per second, which slightly complicates the process of transferring film to video. The minimum frame rate to achievepersistence of vision (the illusion of a moving image) is about 16 frames per second.[8]
Video can beinterlaced orprogressive. In progressive scan systems, each refresh period updates all scan lines in each frame, in sequence. When displaying a natively progressive broadcast or recorded signal, the result is the optimum spatial resolution of both the stationary and moving parts of the image. Interlacing was invented as a way to reduce flicker in earlymechanical andCRT video displays, without increasing the number of completeframes per second. Interlacing retains detail while requiring lowerbandwidth compared to progressive scanning.[9][10]
In interlaced video, the horizontalscan lines of each complete frame are treated as if numbered consecutively and captured as twofields: anodd field (upper field) consisting of the odd-numbered lines and aneven field (lower field) consisting of the even-numbered lines. Analog display devices reproduce each frame, effectively doubling the frame rate as far as perceptible overall flicker is concerned. When the image capture device acquires the fields one at a time, rather than dividing up a complete frame after it is captured, the frame rate for motion is effectively doubled as well, resulting in smoother, more lifelike reproduction of rapidly moving parts of the image when viewed on an interlaced CRT display.[9][10]
NTSC, PAL, and SECAM are interlaced formats. In video resolution notation, 'i' denotes interlaced scanning. For example, PAL video format is often described as576i50, where576 indicates the total number of horizontal scan lines,i indicates interlacing, and50 indicates 50 fields (half-frames) per second.[10][11]
When displaying a natively interlaced signal on a progressive scan device, the overall spatial resolution is degraded by simpleline doubling—artifacts, such as flickering or comb effects in moving parts of the image, appear unless special signal processing eliminates them.[9][12] A procedure known asdeinterlacing can optimize the display of an interlaced video signal from an analog, DVD, or satellite source on a progressive scan device such as anLCD television, digitalvideo projector, or plasma panel. Deinterlacing cannot, however, producevideo quality that is equivalent to true progressive scan source material.[10][11][12]
In video, anaspect ratio is the proportional relationship between the width and height of a video screen and video picture elements. All popular video formats are landscape, with a traditional television screen having an aspect ratio of 4:3, or about 1.33:1. High-definition televisions have an aspect ratio of 16:9, or about 1.78:1. The ratio of a full 35mm film frame with its sound track (the "Academy ratio") is 1.375:1.[13][14]
Pixels on computer monitors are usually square, but pixels used indigital video often have non-square aspect ratios, such as those used in the PAL and NTSC variants of theCCIR 601 digital video standard and the corresponding anamorphic widescreen formats. The720 by 480 pixel raster uses thin pixels on a 4:3 aspect ratio display and fat pixels on a 16:9 display.[13][14]
The popularity of video on mobile phones has led to the growth ofvertical video.Mary Meeker, a partner at Silicon Valley venture capital firmKleiner Perkins Caufield & Byers, highlighted the growth of vertical-video viewing in her 2015 Internet Trends Report, noting that it had grown from 5% of viewing in 2010 to 29% in 2015. Vertical-video ads are watched in their entirety nine times more frequently than those in landscape ratios.[15]
Thecolor model uses the video color representation and maps encoded color values to visible colors reproduced by the system. There are several such representations in common use: typically,YIQ is used in NTSC television,YUV is used in PAL television,YDbDr is used by SECAM television, andYCbCr is used for digital video.[16][17]
The number of distinct colors a pixel can represent depends on thecolor depth expressed in the number of bits per pixel. A common way to reduce the amount of data required in digital video is bychroma subsampling (e.g., 4:4:4, 4:2:2, etc.). Because the human eye is less sensitive to details in color than brightness, the luminance data for all pixels is maintained, while the chrominance data is averaged for a number of pixels in a block, and the same value is used for all of them. For example, this results in a 50% reduction in chrominance data using 2-pixel blocks (4:2:2) or 75% using 4-pixel blocks (4:2:0). This process does not reduce the number of possible color values that can be displayed, but it reduces the number of distinct points at which the color changes.[11][16][17]
Video quality can be measured with formal metrics likepeak signal-to-noise ratio (PSNR) or throughsubjective video quality assessment using expert observation. Many subjective video quality methods are described in theITU-T recommendationBT.500. One of the standardized methods is theDouble Stimulus Impairment Scale (DSIS). In DSIS, each expert views anunimpaired reference video, followed by animpaired version of the same video. The expert then rates theimpaired video using a scale ranging from "impairments are imperceptible" to "impairments are very annoying."
Uncompressed video delivers maximum quality, but at a very highdata rate. A variety of methods are used to compress video streams, with the most effective ones using agroup of pictures (GOP) to reduce spatial and temporalredundancy. Broadly speaking, spatial redundancy is reduced by registering differences between parts of a single frame; this task is known asintraframe compression and is closely related toimage compression. Likewise, temporal redundancy can be reduced by registering differences between frames; this task is known asinterframe compression, includingmotion compensation and other techniques. The most common modern compression standards areMPEG-2, used forDVD, Blu-ray, andsatellite television, andMPEG-4, used forAVCHD, mobile phones (3GP), and the Internet.[18][19]
Two channels: a right channel for the right eye and a left channel for the left eye. Both channels may be viewed simultaneously by usinglight-polarizing filters 90 degrees off-axis from each other on two video projectors. These separately polarized channels are viewed wearing eyeglasses with matching polarization filters.
Anaglyph 3D, where one channel is overlaid with two color-coded layers. This left and right layer technique is occasionally used for network broadcasts or recent anaglyph releases of 3D movies on DVD. Simple red/cyan plastic glasses provide the means to view the images discretely to form a stereoscopic view of the content.
One channel with alternating left and right frames for the corresponding eye, usingLCD shutter glasses that synchronize to the video to alternately block the image for each eye, so the appropriate eye sees the correct frame. This method is most common in computervirtual reality applications, such as in aCave Automatic Virtual Environment, but reduces effective video framerate by a factor of two.
Many analog and digitalrecording formats are in use, anddigital video clips can also be stored on acomputer file system as files, which have their own formats. In addition to the physical format used by thedata storage device or transmission medium, the stream of ones and zeros that is sent must be in a particular digitalvideo coding format, for which a number is available.
Analog video is a video signal represented by one or moreanalog signals. Analog color video signals includeluminance (Y) andchrominance (C). When combined into one channel, as is the case among others withNTSC,PAL, andSECAM, it is calledcomposite video. Analog video may be carried in separate channels, as in two-channelS-Video (YC) and multi-channelcomponent video formats.
Analog video is used in both consumer and professionaltelevision production applications.
Video can be transmitted or transported in a variety of ways including wirelessterrestrial television as an analog or digital signal, coaxial cable in a closed-circuit system as an analog signal. Broadcast or studio cameras use a single or dual coaxial cable system usingserial digital interface (SDI). SeeList of video connectors for information about physical connectors and related signal standards.
An analog video format consists of more information than the visible content of the frame. Preceding and following the image are lines and pixels containing metadata and synchronization information. This surrounding margin is known as ablanking interval orblanking region; the horizontal and verticalfront porch and back porch are the building blocks of the blanking interval.
Early television was almost exclusively a live medium, with some programs recorded to film for historical purposes usingKinescope. The analogvideo tape recorder was commercially introduced in 1951. The following list is in rough chronological order. All formats listed were sold to and used by broadcasters, video producers, or consumers; or were important historically.[22][23]
A video codec issoftware orhardware thatcompresses anddecompressesdigital video. In the context of video compression,codec is aportmanteau ofencoder anddecoder, while a device that only compresses is typically called anencoder, and one that only decompresses is adecoder. The compressed data format usually conforms to a standardvideo coding format. The compression is typicallylossy, meaning that the compressed video lacks some information present in the original video. A consequence of this is that decompressed video has lower quality than the original, uncompressed video because there is insufficient information to accurately reconstruct the original video.[27]
^abcBovik, Alan C. (2005).Handbook of image and video processing (2nd ed.). Amsterdam: Elsevier Academic Press. pp. 14–21.ISBN978-0-08-053361-2.OCLC190789775.
^abcBrown, Blain (2013).Cinematography: Theory and Practice: Image Making for Cinematographers and Directors.Taylor & Francis. pp. 159–166.ISBN9781136047381.
^abParker, Michael (2013).Digital Video Processing for Engineers : a Foundation for Embedded Systems Design. Suhel Dhanani. Amsterdam.ISBN978-0-12-415761-3.OCLC815408915.{{cite book}}: CS1 maint: location missing publisher (link)
^Sanz, Jorge L. C. (1996).Image Technology : Advances in Image Processing, Multimedia and Machine Vision. Berlin, Heidelberg: Springer Berlin Heidelberg.ISBN978-3-642-58288-2.OCLC840292528.
^Ekmekcioglu, Erhan; Fernando, Anil; Worrall, Stewart (2013).3DTV : processing and transmission of 3D video signals. Chichester, West Sussex, United Kingdom: Wiley & Sons.ISBN978-1-118-70573-5.OCLC844775006.
^Block, Bruce A.; McNally, Phillip (2013).3D storytelling : how stereoscopic 3D works and how to use it. Burlington, MA: Taylor & Francis.ISBN978-1-136-03881-5.OCLC858027807.
^abPizzi, Skip; Jones, Graham (2014).A Broadcast Engineering Tutorial for Non-Engineers (4th ed.). Hoboken: Taylor and Francis. pp. 145–152.ISBN978-1-317-90683-4.OCLC879025861.
^Ward, Peter (2015). "Video Recording Formats".Multiskilling for television production. Alan Bermingham, Chris Wherry. New York: Focal Press.ISBN978-0-08-051230-3.OCLC958102392.
^Merskin, Debra L., ed. (2020).The Sage international encyclopedia of mass media and society. Thousand Oaks, California.ISBN978-1-4833-7551-9.OCLC1130315057.{{cite book}}: CS1 maint: location missing publisher (link)