Computer-generated imagery (CGI) is a specific-technology or application ofcomputer graphics for creating or improving images inart,printed media,simulators, videos and video games. These images are either static (i.e.still images) or dynamic (i.e. moving images). CGI both refers to 2D computer graphics and (more frequently)3D computer graphics with the purpose of designing characters,virtual worlds, or scenes andspecial effects (infilms, television programs, commercials, etc.). The application of CGI for creating/improvinganimations is calledcomputer animation (or CGI animation).
This section needs to beupdated. Please help update this article to reflect recent events or newly available information.(November 2025)
The first feature film to use CGI as well as the composition of live-action film with CGI wasVertigo,[1] which used abstract computer graphics byJohn Whitney in the opening credits of the film. The first feature film to make use of CGI with live action in the storyline of the film was the 1973 filmWestworld.[2] The first feature film to present a fully CGI character was the 1985 filmYoung Sherlock Holmes, showcasing a fully animated stained glass knight character.[3] Other early films that incorporated CGI includeDemon Seed (1977),Star Wars (1977),[2]Tron (1982),Star Trek II: The Wrath of Khan (1982),[2]Golgo 13: The Professional (1983),[4]The Last Starfighter (1984),[5]The Abyss (1989),Terminator 2: Judgement Day (1991), andJurassic Park (1993). The firstmusic video to use CGI wasWill Powers' "Adventures in Success" (1983).[6] In 1995,Pixar'sToy Story became the first fully CGI feature film, marking a historic milestone for both animation and film-making.[7]Prior to CGI being prevalent in film, virtual reality, personal computing and gaming, one of the early practical applications of CGI was for aviation and military training, namely theflight simulator. Visual systems developed in flight simulators were also an important precursor to three dimensional computer graphics and Computer Generated Imagery (CGI) systems today. Namely because the object of flight simulation was to reproduce on the ground the behavior of an aircraft in flight. Much of this reproduction had to do with believable visual synthesis that mimicked reality.[8] The Link Digital Image Generator (DIG) by the Singer Company (Singer-Link), was considered one of the world's first generation CGI systems.[9] It was a real-time, 3D capable, day/dusk/night system that was used by NASA shuttles, for F-111s, Black Hawk and the B-52. Link's Digital Image Generator had architecture to provide a visual system that realistically corresponded with the view of the pilot.[10] The basic architecture of the DIG and subsequent improvements contained a scene manager followed by geometric processor, video processor and into the display with the end goal of a visual system that processed realistic texture, shading, translucency capabilities, and free of aliasing.[11]
Combined with the need to pair virtual synthesis with military level training requirements, CGI technologies applied in flight simulation were often years ahead of what would have been available in commercial computing or even in high budget film. Early CGI systems could depict only objects consisting of planar polygons. Advances in algorithms and electronics in flight simulator visual systems and CGI in the 1970s and 1980s influenced many technologies still used in modern CGI adding the ability to superimpose texture over the surfaces as well as transition imagery from one level of detail to the next one in a smooth manner.[12]
The evolution of CGI led to the emergence ofvirtual cinematography in the 1990s, where the vision of thesimulatedcamera is not constrained by the laws of physics. Availability of CGI software and increased computer speeds have allowed individual artists and small companies to produce professional-grade films, games, and fine art from their home computers.
Not only do animated images form part of computer-generated imagery; natural looking landscapes (such asfractal landscapes) are also generated via computeralgorithms. A simple way to generate fractal surfaces is to use an extension of thetriangular mesh method, relying on the construction of some special case of ade Rham curve, e.g.,midpoint displacement.[13] For instance, the algorithm may start with a large triangle, then recursively zoom in by dividing it into four smallerSierpinski triangles, then interpolate the height of each point from its nearest neighbors.[13] The creation of aBrownian surface may be achieved not only by adding noise as new nodes are created but by adding additional noise at multiple levels of the mesh.[13] Thus atopographical map with varying levels of height can be created using relatively straightforward fractal algorithms. Some typical, easy-to-program fractals used in CGI are theplasma fractal and the more dramaticfault fractal.[14]
Many specific techniques have been researched and developed to produce highly focused computer-generated effects — e.g., the use of specific models to represent the chemical weathering of stones to model erosion and produce an "aged appearance" for a given stone-based surface.[15]
A computer-generated image featuring a house at sunset, made inBlender
Modern architects use services from computer graphic firms to create 3-dimensional models for both customers and builders. These computer generated models can be more accurate than traditional drawings.Architectural animation (which provides animated movies of buildings, rather than interactive images) can also be used to see the possible relationship a building will have in relation to the environment and its surrounding buildings. The processing of architectural spaces without the use of paper and pencil tools is now a widely accepted practice with a number of computer-assisted architectural design systems.[16]
Architectural modeling tools allow an architect to visualize a space and perform "walk-throughs" in an interactive manner, thus providing "interactive environments" both at the urban and building levels.[17] Specific applications in architecture not only include the specification of building structures (such as walls and windows) and walk-throughs but the effects of light and how sunlight will affect a specific design at different times of the day.[18][19]
Architectural modeling tools have now become increasingly internet-based. However, the quality of internet-based systems still lags behind sophisticated in-house modeling systems.[20]
In some applications, computer-generated images are used to "reverse engineer" historical buildings. For instance, a computer-generated reconstruction of the monastery atGeorgenthal in Germany was derived from the ruins of the monastery, yet provides the viewer with a "look and feel" of what the building would have looked like in its day.[21]
Computer generated models used inskeletal animation are not always anatomically correct. However, organizations such as theScientific Computing and Imaging Institute have developed anatomically correct computer-based models. Computer generated anatomical models can be used both for instructional and operational purposes. To date, a large body of artist-producedmedical images continue to be used by medical students, such as images byFrank H. Netter, e.g.Cardiac images. However, a number of online anatomical models are becoming available.
A single patientX-ray is not a computer generated image, even if digitized. However, in applications which involveCT scans a three-dimensional model is automatically produced from many single-slice x-rays, producing "computer generated image". Applications involvingmagnetic resonance imaging also bring together a number of "snapshots" (in this case via magnetic pulses) to produce a composite, internal image.
In modern medical applications, patient-specific models are constructed in 'computer assisted surgery'. For instance, in totalknee replacement, the construction of a detailed patient-specific model can be used to carefully plan the surgery.[22] These three-dimensional models are usually extracted from multipleCT scans of the appropriate parts of the patient's own anatomy. Such models can also be used for planningaortic valve implantations, one of the common procedures for treatingheart disease. Given that the shape, diameter, and position of thecoronary openings can vary greatly from patient to patient, the extraction (fromCT scans) of a model that closely resembles a patient's valve anatomy can be highly beneficial in planning the procedure.[23]
To date, making the clothing of a digital character automatically fold in a natural way remains a challenge for many animators.[25]
In addition to their use in film, advertising and other modes of public display, computer generated images of clothing are now routinely used by top fashion design firms.[26]
The challenge in renderinghuman skin images involves three levels of realism:
Photo realism in resembling realskin at the static level
Physical realism in resembling its movements
Function realism in resembling its response to actions.[27]
When animating a texture like hair or fur for a computer generated model, individual base hairs are first created and later duplicated to demonstrate volume.[28] The initial hairs are often different lengths and colors, to each cover several different sections of a model. This technique was notably used inPixar'sMonsters Inc (2001) for the character Sulley, who had approximately 1,000 initial hairs generated that were later duplicated 2,800 times.[28] The quantity of duplications can range from thousands to millions, depending on the level of detail sought after.
Interactive visualization is the rendering of data that may vary dynamically and allowing a user to view the data from multiple perspectives. The applications areas may vary significantly, ranging from the visualization of the flow patterns influid dynamics to specificcomputer aided design applications.[29] The data rendered may correspond to specific visual scenes that change as the user interacts with the system — e.g. simulators, such asflight simulators, make extensive use of CGI techniques for representing the world.[30]
At the abstract level, an interactive visualization process involves a "data pipeline" in which the raw data is managed and filtered to a form that makes it suitable for rendering. This is often called the"visualization data". The visualization data is then mapped to a "visualization representation" that can be fed to a rendering system. This is usually called a"renderable representation". This representation is then rendered as a displayable image.[30] As the user interacts with the system (e.g. by using joystick controls to change their position within the virtual world) the raw data is fed through the pipeline to create a new rendered image, often making real-time computational efficiency a key consideration in such applications.[30][31]
While computer-generated images of landscapes may be static,computer animation only applies to dynamic images that resemble a movie. However, in general, the term computer animation refers to dynamic images that do not allow user interaction, and the termvirtual world is used for the interactive animated environments.
Computer animation is essentially a digital successor to the art ofstop motion animation of 3D models and frame-by-frame animation of 2D illustrations. Computer generated animations are more controllable than other more physically based processes, such as constructingminiatures for effects shots or hiringextras for crowd scenes, and because it allows the creation of images that would not be feasible using any other technology. It can also allow a single graphic artist to produce such content without the use of actors, expensive set pieces, or props.
To create the illusion of movement, an image is displayed on thecomputer screen and repeatedly replaced by a new image which is similar to the previous image, but advanced slightly in the time domain (usually at a rate of 24 or 30 frames/second). This technique is identical to how the illusion of movement is achieved withtelevision andmotion pictures.
An image conditioned on the promptan astronaut riding a horse, byHiroshige, generated byStable Diffusion 3.5, a large-scale text-to-image model first released in 2022
Text-to-image models gradually began to be developed in the mid-2010s during the beginnings of theAI boom, as a result of advances indeep neural networks. In 2022, the output of state-of-the-art text-to-image models—such as OpenAI'sDALL-E 2,Google Brain'sImagen, Stability AI'sStable Diffusion,Midjourney, and Runway's Gen-4—began to be considered to approach the quality ofreal photographs and human-drawnart.
Text-to-image models are generallylatent diffusion models, which perform the diffusion process in a compressedlatent space rather than directly in pixel space. An autoencoder (often avariational autoencoder (VAE)) is used to convert between pixel space and this latent representation. These systems typically use a pretrained language or vision–language model to convert the input prompt into a text embedding, and a diffusion-based generative image model that produces images conditioned on that embedding. The most effective models have generally been trained on massive amounts of image and text datascraped from the web.[32]
A virtual world is anagent-based andsimulated environment allowing users to interact with artificially animated characters (e.gsoftware agent) or with other physical users, through the use ofavatars. Virtual worlds are intended for itsusers to inhabit and interact, and the term today has become largely synonymous with interactive 3D virtual environments, where the users take the form ofavatars visible to others graphically.[33] These avatars are usually depicted as textual, two-dimensional, orthree-dimensional graphical representations, although other forms are possible[34] (auditory[35] and touch sensations for example). Some, but not all, virtual worlds allow for multiple users.
Computer-generated imagery has been used in courtrooms, primarily since the early 2000s. However, some experts have argued that it is prejudicial. They are used to help judges or the jury to better visualize the sequence of events, evidence or hypothesis.[36] However, a 1997 study showed that people are poor intuitive physicists and easily influenced by computer generated images.[37] Thus it is important that jurors and other legal decision-makers be made aware that such exhibits are merely a representation of one potential sequence of events.
Weather visualizations were the first application of CGI in television. One of the first companies to offer computer systems for generating weather graphics wasColorGraphics Weather Systems in 1979 with the "LiveLine", based around anApple II computer, with later models from ColorGraphics usingCromemco computers fitted with theirDazzler video graphics card.
It has now become common in weather casting to display full motion video of images captured in real-time from multiple cameras and other imaging devices. Coupled with 3D graphics symbols and mapped to a common virtual geospatial model, these animated visualizations constitute the first true application of CGI to TV.
CGI has become common in sports telecasting. Sports and entertainment venues are provided with see-through and overlay content through tracked camera feeds for enhanced viewing by the audience. Examples include the yellow "first down" line seen in television broadcasts ofAmerican football games showing the line the offensive team must cross to receive a first down. CGI is also used in association with football and other sporting events to show commercial advertisements overlaid onto the view of the playing area. Sections ofrugby fields andcricket pitches also display sponsored images. Swimming telecasts often add a line across the lanes to indicate the position of the current record holder as a race proceeds to allow viewers to compare the current race to the best performance. Other examples include hockey puck tracking and annotations of racing car performance[38] and snooker ball trajectories.[39][40] Sometimes CGI on TV with correct alignment to the real world has been referred to asaugmented reality.
Computer-generated imagery is often used in conjunction withmotion capture to better cover the faults that come with CGI and animation.Computer-generated imagery is limited in its practical application by how realistic it can look. Unrealistic, or badly managed computer-generated imagery can result in theuncanny valley effect.[41] This effect refers to the human ability to recognize things that look eerily like humans, but are slightly off. Such ability is a fault with normal computer-generated imagery which, due to the complex anatomy of the human body, can often fail to replicate it perfectly. Artists can use motion capture to get footage of a human performing an action and then replicate it perfectly with computer-generated imagery so that it looks normal.
In many instances, motion capture is needed to accurately mimic an actor's full body movements while slightly changing their appearance with de-aging.De-aging is a visual effect used to alter the appearance of an actor, often through facial scanning technologies,motion capture, and photo references. It is commonly used for flashback scenes and cameos to have an actor appear younger.Marvel'sX-Men: The Last Stand was the first film to publicly incorporate de-aging, which was used on actorsPatrick Stewart andIan Mckellen for flashback scenes featuring their characters at a younger age.[42] The visual effects were done by the companyLola VFX, and used photos taken of the actors at a younger age as references to later smooth out the wrinkles on their face with use of CGI. Overtime, de-aging technologies have advanced, with films such asHere (2024), portraying actors at younger ages through the use of digital AI techniques, scanning millions of facial features and incorporating a number of them onto actors' faces to alter their appearance.[43]
The lack of anatomically correct digital models contributes to the necessity of motion capture as it is used with computer-generated imagery. Because computer-generated imagery reflects only the outside, or skin, of the object being rendered, it fails to capture the infinitesimally small interactions between interlocking muscle groups used infine motor skills like speaking. The constant motion of the face as it makes sounds with shaped lips and tongue movement, along with the facial expressions that go along with speaking are difficult to replicate by hand.[44] Motion capture can catch the underlying movement of facial muscles and better replicate the visual that goes along with the audio.
^Rolfe, J. M. and Staples, K.J. (May 27, 1988).Flight Simulation (Cambridge Aerospace Series, Series Number 1). New York: Cambridge University Press.ISBN978-0-521-35751-7.{{cite book}}: CS1 maint: multiple names: authors list (link)
^Suminski, Leonard and, Hulin, Paul (September 26, 1980). "Computer Generated Imagery Current Technology".U.S. Army Project Manager for Training Devices:14–18.{{cite journal}}: CS1 maint: multiple names: authors list (link)
^Cook, A.D. (2009). A case study of the manifestations and significance of social presence in a multi-user virtual environment. MEd Thesis. Availableonline
^Kassin, S. M. (1997). "Computer-animated Display and the Jury: Facilitative and Prejudicial Effects".Law and Human Behavior.40 (3):269–281.doi:10.1023/a:1024838715221.S2CID145311101.[2]
Begault, Durand R. (1994).3-D Sound for Virtual Reality and Multimedia. AP Professional.ISBN978-0-1208-4735-8.
Biocca, Frank; Levy, Mark R. (1995).Communication in the Age of Virtual Reality. Lawrence Erlbaum Associates.ISBN978-0-8058-1549-8.
Peitgen, Heinz-Otto; Jürgens, Hartmut; Saupe, Dietmar (2004).Chaos and Fractals: New Frontiers of Science. Springer Science & Business Media.ISBN978-0-387-20229-7.