CROSS-REFERENCE TO RELATED APPLICATIONSThis application is a division of U.S. patent application Ser. No. 08/923,091, filed Sep. 4, 1997, which claims priority based on U.S. Provisional Patent Application Serial No. 60/025,604, filed Sep. 9, 1996, entitled “Apparatus For Video Access And Control Over Computer Network,” and based on U.S. Provisional Application Serial No. 60/033,485, filed Dec. 20, 1996, entitled “Apparatus For Video Access And Control Over Computer Network, Including Image Correction.” Both provisional applications are incorporated herein by reference in their entirety.[0001]
INCORPORATION BY REFERENCEAdditionally, the following patents, patent applications and publications are incorporated herein by reference in their entirety:[0002]
U.S. Pat. No. 5,559,549, issued Sep. 24, 1996 to Hendricks et al.,[0003]
U.S. Pat. No. 5,600,573, issued Feb. 4, 1997 to Hendricks et al.,[0004]
U.S. Pat. No. 5,185,667, issued Feb. 9, 1993 to Zimmerman,[0005]
U.S. Pat. No. 5,313,306, issued May 17, 1994 to Kuban et al.,[0006]
U.S. Pat. No. 5,359,363, issued Oct. 25, 1994 to Kuban et al.,[0007]
U.S. Pat. No. 5,384,588, issued Jan. 24, 1995 to Martin et al.,[0008]
U.S. Pat. No. 5,489,940, issued Feb. 6, 1996 to Richardson et al.,[0009]
U.S. Pat. No. 6,201,536, issued Mar. 13, 2001 to Hendricks et al.,[0010]
PCT Publication No. WO 96/07269, published Mar. 7, 1996 by Jambhekar et al.,[0011]
PCT Publication No. WO 96/08105, published Mar. 14, 1996 by Labun,[0012]
PCT Publication No. WO 96/18262, published Jun. 13, 1996 by Richardson et al.,[0013]
PCT Publication No. WO 96/21173, published Jul. 11, 1996 by Harris et al., and[0014]
PCT Publication No. WO 96/21205, published Jul. 11, 1996 by Harris et al..[0015]
BACKGROUND OF THE INVENTIONThis invention relates to the distribution of audiovisual signals through communications networks such as computer networks and servers. The invention has particular use with respect to global networks such as the Internet and “World Wide Web”. The invention also relates to education. Particularly, the invention provides an alternative to in-person classroom instruction.[0016]
1. Field of the Invention[0017]
The present invention relates to the fields of education, audiovisual systems, communications systems and computer networks.[0018]
Individuals from around the world exchange ideas and information with each other in order to learn more about other people, cultures, and the environment in which we live. Video and audio signals are commonly transmitted over broadcast communications media to provide viewers with news and entertainment. Computer networks are used for the remote exchange of data and other information. Broadly speaking, these systems are attempts to communicate useful knowledge between geographically separate individuals and institutions. The invention generally relates to improvements in the transmission of information between remote locations.[0019]
2. Description of Related Art[0020]
There is a constant desire to improve education and knowledge at all levels. It is thought that true human progress can only be achieved if people's understanding of each other is improved and if people's understanding of nature and the environment is improved. Traditionally, education and knowledge have been obtained ill schools from classroom instruction and from the reading of books.[0021]
The disadvantage of current classroom instructional systems is that students must be physically present in the classroom to participate in the educational process. Therefore, students who are geographically displaced from the location of the classroom often cannot attend class instruction as often or as timely as students who are nearby to the classroom.[0022]
The disadvantage of textbooks is that they are often not kept current with recent events or technological changes. Textbooks are usually only updated on a yearly or less frequent basis, while important changes may occur monthly or more frequently. Also, to save funds, schools may not purchase new textbooks even though the textbooks have been updated. Therefore, the new knowledge, although available, is not communicated to students.[0023]
Recently, audiovisual presentations have begun to be used in the field of education. These systems may provide playback of a recording of a lecturer who provides a presentation on an educational topic. For example, students may learn about math from watching a videotape or television broadcast of a math professor's lecture. Education can also occur on a more informal basis. For example, specialty channels in the United States such as the Discovery Channel® and The Learning Channel® (headquartered in Bethesda, Md., U.S.A.) broadcast educational programming which both entertains and educates a diverse viewership.[0024]
The disadvantage of these audiovisual systems is that they are not interactive. Students are unable to ask questions, and the lecturer is unable to tailor the presentation of material to the specific needs of the current student audience. Consequently, the needs of the students are not met.[0025]
Cable and broadcast television are commonly known media that supply information to large numbers of viewers equipped with receivers known as “television sets.” By receiving a broadcast, cablecast or satellite signal, users are able to view scenes from remote locations and observe newsworthy events that occur far from the user's location. However, conventional television is a one-way media in which users cannot communicate with each other or the broadcaster.[0026]
Recently, the advent of the “Internet,” and “World Wide Web,” in conjunction with the proliferation of personal computers, has allowed people to exchange information and ideas on a global and inexpensive basis. Generally speaking, the Internet is a large computer network which connects “host” computers. Users with a computer, modern and telephone line commonly call via telephone to connect with a “host.” The “host,” being in communication with other hosts (connected to other users) is able to transfer information between users. The Internet is used, for example, to transfer, data files, still images, sounds and messages between virtually any two points in the world with telephone access.[0027]
The use of the Internet has increased dramatically since 1981, when approximately 300 host computers were linked together. It has been estimated that in 1989, the number of linked host computers was fewer than 90,000; but by 1993, over a million host computers were connected. Currently over 9.4 million host computers are linked (not including the personal computers people use to access these hosts via modems) and as many as 40 million people around the world may have access to the Internet medium. This number is expected to grow to 200 million by the year 1999.[0028]
Users on the Internet are able to transfer text, graphics, and still pictures between 30 remote locations. Other types of information that can be transmitted include files containing prerecorded sequences of images. To view these images, users download a large data file, and after running appropriate software, see a sequence of images on the computer screen. These images are not provided in real time, and are not viewable while the user is accessing the Internet.[0029]
Therefore, even though the Internet is a two-way communication medium, it is not currently being utilized to provide video information and audiovisual presentations. This is a disadvantage, in that a large number of people have been accustomed to television audiovisual presentations, and prefer an audio-video presentation to a textual or graphical presentation.[0030]
What is needed is a medium of communication that is interactive and which carries audio, video, text, and graphics.[0031]
What is needed is an educational system that is user friendly and entertaining.[0032]
What is needed is to improve the Internet such that users can access many audiovisual programs.[0033]
What is needed is to provide users with live video from remote sites.[0034]
What is needed is a remote video system with increased realism and accuracy, such that users feel as though they were actually present at the remote location.[0035]
SUMMARY OF THE INVENTIONIn accordance with the present invention, video is collected at a remote site. (The term “video”, as used herein, includes stereophonic or monophonic audio signals that may accompany a video signal. Additionally, “video” is used broadly herein to include still images, groups of related still images, animation, graphics, pictures, or other visual data.) The remote video information may be obtained from a video cassette, CD ROMs, television channels, one or more video cameras, or other well known sources. If video cameras are used, they may be connected to a computer so that they are remotely controllable, or they may be oriented such that a perception of control can be created for users. The video may relate to remote sites of interest, such as a pyramid in Egypt, or the images may relate to an educational lecture being conducted at a remote site.[0036]
The collected video is transferred to a web site, either in compressed or uncompressed form. The video may be physically transported or may be transmitted through a communications medium to the web site.[0037]
The web site contains a storage media that may store some or all of the video. Additionally, the web site passes camera control commands, if applicable, to the remotely controlled cameras or may simulate the remote control of a camera. The main function of the web site is to pass video to a plurality of users, through a communication media Such as the Internet, in response to user selections. The video passed to the plurality of users may be live video being fed to the web site, or may be stored video. A number of video servers are used to output the video to the users through the communications media, such as the Internet. The video may be tailored by the web site for the particular user's hardware, including data communication equipment, or memory size, etc . . . , i.e., the data rate matches the highest speed which the user's equipment can handle.[0038]
Users receive and display the video sent from the web site. Many simultaneous video pictures may be received. Of course, the quality and frame rate of the video is dependent on the user's communications hardware. Users with high-speed modems or cable modems receive higher quality video. The users are able to send commands and/or queries to the web site. The commands and queries are forwarded to remote locations to control remote cameras or query remotely located instructors. Alternatively, the commands cause the web site to change from among many video signals with different camera angles or locations (or to transmit a different portion of a wide angle image), causing the user to have a perception of remote camera control. The user's commands may also cause a different portion of a received wide-angle image to be displayed, giving the user a perception of camera control.[0039]
In addition to video, the web site provides information, such as graphics and text, which is related to the video. This information may be automatically supplied, or provided upon user request. Therefore, the user is provided with a comprehensive set of information concerning remote sites, enabling the user to be quickly educated about the remote site of interest.[0040]
BRIEF DESCRIPTION OF THE DRAWINGSFIG. 1 is a block diagram of an embodiment of the invention where remote video is provided to a web server by videocassette and by ordinary television.[0041]
FIG. 2 is a block diagram of an embodiment of the invention where remote video is provided by remotely located cameras and a communication network carries the video to the web server.[0042]
FIGS. 3A and 3B are a block diagrams of an embodiment of the invention using the embodiments of FIGS. 1 and 2 with remotely controllable cameras.[0043]
FIG. 4 shows remote cameras positioned around a building for perceived camera control.[0044]
FIGS. 5A, 5B,[0045]5C, and5D show video images from specific cameras shown in FIG. 4.
FIG. 6 shows remote cameras deployed to follow a parade route.[0046]
FIGS. 7A and 7B show remotely controlled cameras at a remote location.[0047]
FIGS. 8A and 8B show a single remote camera at a remote location, where the camera has a 180 degree spherical (or other wide angle) lens.[0048]
FIGS. 9A and 9B are block diagrams of a server platform.[0049]
FIG. 10 is a block diagram of communications paths from the server site to remote users.[0050]
FIG. 11 shows a home page in accordance with an embodiment of the invention.[0051]
FIG. 12 shows a “society” page in accordance with another embodiment of the invention.[0052]
FIG. 13 shows a “map” page of remote camera locations throughout the world.[0053]
FIG. 14 shows a “watch” page containing live video feeds from five remote cameras.[0054]
FIG. 15 shows a page directed to determining the user's data rate.[0055]
FIG. 16 shows a page of an interactive lecture.[0056]
FIGS. 17 and 18 show pages of an embodiment of the invention that combines live video, pre-stored video, graphics, and interactive questions.[0057]
FIG. 19 shows a flow diagram of a method of automatically monitoring and panning an area using perceived camera control.[0058]
FIG. 20 is an exemplary screen display of the present invention, showing video and also showing video data.[0059]
FIG. 21 is a diagram showing the interaction between a computer network embodiment of the present invention and a cable television system.[0060]
DETAILED DESCRIPTION OF THE DRAWINGSAs stated previously, the present invention is related to obtaining video from remote sites and interactively presenting that video to users. The video is obtained at a remote site, communicated to a web site (where it may be stored), and forwarded to users.[0061]
1. Obtaining Video from Remote Sites, Communicating the Video to a Web Site, and Streaming the Video to Users.[0062]
FIG. 1 shows a preferred embodiment of the invention where remote video sources are videocassette and television programs. FIG. I shows[0063]remote sites102,remote cameras104,videocassette106,compression devices108,114,digital storage device110 andweb site112. As shown in FIG. 1, avideo camera104 is used to film activity atremote site102. As discussed below, numerous video cameras at a single remote site may be used to obtain different views and audio (preferably stereophonic) of the remote site from different angles and orientations. Also, numerous remote sites, each with its own video camera, may used as shown at102′,102″ and104′ and104″. The video cameras film events at the remote sites, and record the events onvideocassette106 or other suitable media.
The recorded information is then transported to a[0064]web site112, or to a site in communication withweb site112. As shown in FIG. 1, the recorded information from video-tape106 is then compressed incompression unit108 and stored indigital storage media110. Many compression algorithms may be used, such as MPEG-1, MPEG-2 and Wavelet. Compression systems currently available from The Duck Corp, Xing Technology Corp., Indeo, Digital Video Arts, Ltd., VDOnet Corp. and Intel Corp., may be used with the system. The digital storage media may be any known storage device, such as a hard disk, CD ROM, digital video disc (DVD), digital tape, video file server or other media.
The stored and compressed audio/video is then provided on a number of streamed audio-[0065]video outputs116 from theweb site112. This enables many users to access the stored video and audio, and allows for one user to receive numerous audio-video signals, i.e., split the display into numerous “camera” feeds.
In addition to providing streamed audio and video from videocassette, the[0066]web site112 may provide audio and video from television channels. The television signals are received by a conventional television receiver (not shown), and digitally compressed by thecompression unit114 and fed through theweb site112 to the streamed output. It is not normally necessary to store the television programs in a digital storage unit (such as the storage unit110), since the audio and video is constantly incoming and changing. However, certain segments of broadcast television may be stored in a storage device (not shown) for recall by a user.
FIG. 2 shows another embodiment of the invention where similar reference numerals indicate items that correspond to the items shown in FIG. 1. The system of FIG. 2 uses remote cameras and a communication network to provide remote video to the web site. FIG. 2 shows[0067]remote sites102,video cameras104,compression unit118,data communication network120,web site130,digital storage unit132, and streamedvideo116.
As shown in FIG. 2,[0068]remote sites102 are filmed by cameras104 (as in FIG. 1). However, in this embodiment, the outputs of thecameras104 pass through acompression unit118. The compressed audio and video is communicated overdata communication network120 toweb site130. Thedata communication network120 may be any network currently known to one of ordinary skill in the art, such as land-leased lines, satellite, fiber optic cable, microwave link or any other suitable network.
Other suitable networks may be cellular networks or paging networks. In a paging network,[0069]cameras104 may be connected to a paging device and/or digital storage media or paging transmitter for communication of the video (including still images) to theweb site130. The following publications, hereby incorporated by reference, disclose relevant systems: PCT Publication No. WO 96/07269, published Mar. 7, 1996 by Jambhekar et al.; PCT Publication No. WO 96/21173, published Jul. 11, 1996 by Harris et al.; PCT Publication No. WO 96/21205, published Jul. 11, 1996 by Harris et al.
The[0070]web site130 in this example is adapted to receive information from thedata communication network120. The web site may pass the video fromcameras104 to users at streamed video outputs116. In alternative embodiments, the web site may contain a decompressor to decompress the video prior to streaming it to users, or change the compression scheme of the video to one that is compatible with the connected user. Alternatively, the video may be compressed at the streamed video output and users who connect to theweb site130 may run decompression software. Theweb site130 may store the audio and video received overdata communication network120 indigital storage unit132 before providing it to the streamed outputs116. Alternatively, the audio and video may be directly passed to the streamed outputs116.
FIG. 3A shows another embodiment of the invention that combines the embodiments of FIGS. 1 and 2 and adds remote camera control. FIG. 3A shows[0071]remote sites102,cameras104,computer134,video path122,129,control path124,126,128,compressors108,114,118,136data communication network120,web site140, digital storage means132, and streamedvideo116. As with FIGS. 1 and 2,remote sites102 are filmed bycamera104. As with FIG. 1, theweb site140 is able to receive video-tape106, compress the audio and video incompression unit108, and store the compressed audio andvideo110. Audio and video from television stations may also be compressed bycompression unit114 and stored or passed as streamedvideo116, as in FIG. 1.
Likewise, the[0072]cameras104 may be connected to compression unit118 (as in FIG. 2) and communicate compressed audio and video toweb site140 viadata communication network120. Thus the functions performed by the embodiments shown in FIGS. 1 and 2 may be combined in a variety of manners at asingle web site140.
FIGS. 3A and 3B add the additional feature of camera control to the previously described embodiments. As shown in FIG. 3A, a[0073]computer134 is connected toremote camera104. The computer is able to control a mechanical or electrical device on thecamera104, to alter the camera's orientation (including position and/or angle). Audio and video from thecamera104 passes to thecomputer134. The video may be processed and stored in the computer. Preferably, as shown in FIG. 3B, the computer is connected to multipleremote cameras104′ and104″ so that multiple users may each control a camera. Thecomputer134 may either contain a compressor or be connected to anexternal compression unit136. The video fromcameras104′ and104″ is compressed and provided todata communications network120. This compressed video is subsequently received byweb site140. Theremote cameras104′,104″ (FIG. 3B) may be controlled by control signals passed fromcomputer134 onpath124. The control signals are received bycomputer134 from thedata communications network120 over thecamera control path126. Theweb site140 provides the control information to thedata communications network120 overpath128. Theweb site140 of this example is adapted to pass control signals tocameras104 and to store video images in a digital storage means132. The web site provides a number of streamedvideo outputs116 as in the other examples.
This embodiment allows remote users to control the angle or orientation of[0074]cameras104′,104″. Users are connected to theweb site140 and receive the streamedvideo116 from thecameras104′,104″. If the users wish to move thecamera104′,104″ to the right, they may enter a user command (such as “pan right”) at their terminal. The command is received by theweb site140, and formatted, if necessary. The command is outputted to thedata communication network120 as a control signal through thecamera control path128. Theremote computer134 receives the camera control signals from thecommunication network120 overcamera control path126. Theremote computer134 may be adapted to control multiple cameras atmultiple locations102, or multiple cameras at thesame location102.
The[0075]computer134 is connected to theremote camera104 by acamera control path124. This path allows control commands from the computer to travel to thecameras104′,104″ and control thecameras104′,104″. Thecameras104′,104″ may have computer-controlled swivel motors (not shown) for panning left and right, may have a computer-controlled pivot motor (not shown) for panning up and down, and may have a computer-controlled motor (not shown) for moving a zoom lens. These motors are known to the artisan and are currently available. A plurality of cameras may be provided at a single site to allow multiple users to have camera control at the same time.
This system of obtaining and/or storing video at a web site is extremely flexible. The system allows for perceived camera control by multiple cameras, actual camera control of one or more cameras, perceived camera control via a wide-angle lens on a single camera, and for the generation of comprehensive interactive programs.[0076]
2. Perceived Camera Control with Multiple Cameras.[0077]
In one alternative embodiment, shown more clearly in FIGS.[0078]4-6, users are given the perception of camera control. To achieve this, a plurality of fixedcameras104,150,152,153,154,156,158,160,162 (FIG. 4) are disposed around aremote site102. In accordance with this embodiment, it appears to users that they are controlling the angle or position of a camera when in actuality they are merely being transferred to the video output of a different camera. FIGS.4-6 show this concept in greater detail.
As shown in FIG. 4, a[0079]building146 is being prepared for demolition. Disposed around thebuilding146 arecameras104,150,152,153,154,156,158,160,162, connected to acomputer135. Thecomputer135 is connected to a communication network120 (not shown). The video fromcameras104,150,152,153,154,156,158,160,162 is digitized and preferably compressed prior to communication overnetwork120, either by compressors connected to the cameras (not shown) or by a compressor connected to the computer135 (not shown). The cameras may be digital cameras or analog cameras connected to an analog-to-digital converter.
The cameras specifically identified around the periphery are[0080]cameras150,152,153,154,156,158,160, and162. For reference, the building contains the letter “A” and the letter “B” on two sides as shown at144 and148 in FIGS. 4 and 5. A number ofadditional cameras104 are disposed about the periphery of the building in a circular pattern. The pattern and number of cameras are not critical, but will control how the user perceives movement of the “camera”.
Referring to FIG. 4, a[0081]video camera150 faces side A, avideo camera152 is between sides A and B, avideo camera153 faces side B and avideo camera154 is between side B and the side opposite side A. Thevideo cameras156,158,160 and162 are disposed closer to the building, as shown. All the video cameras contain audio pickups (preferably stereo). Additionally, all the video cameras are connected to acomputer135 which outputs compressed audiovisual signals to thecommunication network120 and consequently to the web site. The system shown in FIG. 4 may be implemented by the systems shown in either FIG. 2 or FIG. 3. Any number of users in communication with theweb site130,140 may receive the audio and video from these cameras.
FIG. 5A shows a[0082]typical screen view150 of the video presented to remote users who are connected to the web site of the present invention. As shown, the user is observing live video fromcamera150, which provides a view of the building on side A. A “toolbar” ofcommands151 is presented to the user, including a pan left command “←”, a pan right command “→”, a pan up command “↑” and a pan down command “↓”. An “autopan” command is used in conjunction with another command (such as pan right). The “autopan” command is used to automatically move the picture position in the direction previously entered. For example, if “autopan” is entered after “pan right,” then the picture will keep panning right until another key is pressed or a default key (such as the ESCape key) is pressed. The speed of the “autopan” function is controlled by the “speed” command, which is used in conjunction with the “+” and “−” commands. Additionally, the “+” and “−” commands, when used alone, control a “zoom-in” and “zoom-out” function, respectively. The “toolbar” commands are selected via a user input device, which may be a keyboard, mouse, trackball, remote control, etc.
When any user wishes to switch from the view of the camera[0083]150 (FIG. 5A) and pan to the right, the user initiates a pan right command “→”, which is transmitted to theweb site130,140 (FIGS. 2 and 3). The web site receives the command, and in response, causes the video from the camera positioned to the right of thecamera150, in this case the video camera152 (FIG. 4) to be transmitted to the user. The user then observes the picture appearing in FIG. 5B, which appears to be a view to the right from the previous position (camera150). If the user continues to pan right, he is presented with the FIG. 5C view, received from thecamera153. The user may continue to pan right all away around the building in this manner.
Additionally the user has special functions available, such as “autopan” and “zoom.” For example, “autopan” in conjunction with “pan right” would cause the view of the building to rotate, at a speed dictated by the “speed” function and the “+” and “−” keys. Using the “+” and “−” keys alone causes the view to change to a closer camera (“+”) or a camera further away (“−”). As shown in FIG. 4, the[0084]cameras156,158,160 and162 are disposed closer to the building thancameras150,152,153 and154. A “magnified” image, obtained from thecamera156, is shown in FIG. 5D. If no cameras are disposed closer or further away, digital image processing may be used to digitally increase or reduce the size of the image. The software that controls these functions may be disposed either at the web server or on the user's computer.
Thus, users may obtain different views of the[0085]building146 as if they were remotely controlling the positioning of a single remote camera. The users may observe the demolition of the building from many exciting perspectives. This “perceived” camera control is advantageous because it allows any number of users to “control” a camera. A single camera that is remotely controllable is only controllable by a single user. Thus, the present invention is suitable for large audiences. The realism of this perceived control is directly dependent upon the number of cameras and their distances from the viewed object.
Therefore, when the[0086]building146 is demolished, any number of users may pan around the building in real time as if they were actually present at the site. When the building is demolished, the video cameras pick up, preferably in stereo, the sounds of the demolition. Users who have loudspeakers connected to their computer may experience the demolition almost as if they were present.
FIG. 6 shows a deployment of a number of[0087]cameras104 which are arranged in a linear fashion around a point of interest, each camera connected tocomputer135 as in FIG. 4. As with FIGS.4-5, this embodiment uses “perceived” camera control that may be achieved by the systems shown in FIGS.2 or3. In this example, the remote location and point of interest is a parade, such as a New Year's Day Parade. With the camera deployment shown, a user may traverse the length of the parade without actually being present. Users may view whichever part of the parade they are interested in, for as long as they desire, without worry that they have missed an interesting band or float. In this example, the camera deployment merely follows the parade route. Parents who have children in a band or float may search for the child and follow the child throughout the parade route, rather than having to monitor every moment of the parade on television in the hopes that the child will pass the reviewing camera when the parents are watching. The parents merely “move” from different cameras along the parade route as their children progress in the parade.
3. Actual Camera Control of Single/Multiple Cameras.[0088]
FIGS. 7A and 7B show another embodiment, where a number of[0089]cameras160,162,164,166, are provided. These cameras are in direct communication with and are controlled bycomputer170. Although it is possible to form a ring of cameras to perform “perceived” camera control (as in FIGS.4-6), the embodiment shown uses fourcameras160,162,164,166 that contain motors105 (FIG. 7B) for controlling the camera's positioning. The motors are controlled bycomputer170. Either asingle computer170 or a number ofcomputers170 may be used. The remote location and point of interest shown in FIGS. 7A and 7B are, for example, a watering hole or desert oasis. Users who access theweb site140 are able to observe live video of wildlife behavior at the watering hole. Thecameras160,162,164,166 are disposed at an island in the middle of the watering hole. Thetoolbar151 of FIG. 5 is also used in this embodiment and enables users to choose camera control commands to spin the cameras around or perform other camera functions, such as zoom. Users are therefore able to receive different views and angles, and observe the entire watering hole.
FIG. 7B shows the control and video paths of the FIG. 7A system combined with system shown in FIGS. 3A and 3B. The video from[0090]cameras160,162,164,166 is communicated tocomputer170, in compressed or uncompressed form onpath122. Thecomputer170 communicates the video tocommunications network120 for reception by the web site140 (FIGS. 3A, 3B). Preferably the video is digitized and compressed by either thecameras160,162,164,166, thecomputer170, or an external analog-to-digital converter (not shown) and compressor136 (FIGS. 3A, 3B) prior to transfer to thecommunications network120.
Camera control commands are received by the[0091]computer170 oncontrol line126, as shown in FIGS. 3A, 3B and7B. The commands are formatted, if necessary, bycomputer170 and transferred to controlunits105 attached tocameras160,162,164,166. Thecontrol units105 are connected to spin, zoom, or otherwise control the cameras as directed by the user.
Communications links[0092]124 and122 may be wired, wireless, digital or analog, andcomputer170 may be located nearby or remote from thesite102.
The system of FIGS. 7A and 7B are unlike the embodiments shown in FIGS.[0093]4-6, because each user is assigned a remote camera in the FIG. 7A, 7B embodiment. Since each user must be assigned their own controllable camera, users will have to contend for available cameras. The number of controllable cameras may range from a single camera to any number, and is preferably statistically determined to correlate to the average number of users who access theweb server140 at any given time or at peak times. The number of cameras may be reduced by using known systems that utilize queuing, reservations, and time limits.
4. Perceived Camera Control Using a Single Camera and a Wide-angle Lens.[0094]
FIGS. 8A and 8B show another embodiment, using only a single camera, where an unlimited number of users may view any portion of the[0095]remote site102. This embodiment uses aspherical lens182 in optical communication with thecamera180. Theremote site102 shown in FIG. 8 is a remote watering hole or oasis as in FIGS. 7A and 7B.
As shown in FIG. 8A, a[0096]camera180 has a spherical (or other wide angle)lens182, which provides a180□ spherical (or other wide-angle) view. This view, which is communicated to acomputer184, contains distortion. Thecomputer184 communicates and compresses the distorted video back to theweb site130 or140 that stores and may process the image. Rather than using thecomputer184, a simple transmitter may be used to convey the entire spherical video to theweb site130,140 (FIGS. 2 and 3). By using appropriate image processing software, the web site removes the barrel distortion and stores data relating to the entire spherical view. Users may then access different portions of the180□ sphere. In this embodiment, thetoolbar151 of FIG. 5 is also used. By using thetoolbar151, users may move across the spherical view and obtain the “perception” of camera control. This embodiment is advantageous in that it can provide the perception of camera control to any number of users simultaneously using only one remote camera.
FIG. 8B shows alternative embodiments of the system shown in FIG. 8A. As shown in FIG. 8B, the spherical (or other wide angle)[0097]lens182 is used withvideo camera180″, which conveys video information tocomputer184.Computer184 communicates the video overcommunications network120 to theweb site130. Theweb site130 may store or process the received video, and make the video available to users atuser terminals302,304,306,308,310 by communicating the video overcommunication network125.Communication network125 is explained in more depth below with respect to FIG. 10.
Because wide-angle lenses generate distortion, processing is conducted on the distorted image to remove the distortion from a segment of the image. This processing may be performed at the[0098]computer184, or theweb site130, but is preferably performed at theuser terminals302,304,306,308,310.
Thus, the[0099]web site130 has available wide-angle video for sending to users. Users display and view only a segment of the wide-angle video at a time. Then, by using toolbar151 (FIG. 5), the user may select adjacent segments of the video for view. When a user selects an adjacent segment of the video for display, the adjacent segment is processed to remove distortion and then displayed. Displaying the adjacent segment gives the appearance that the camera was physically “moved” to the adjacent side of the original segment.
One system for electronically removing the distortion from a segment of an image obtained from a fish-eye lens is disclosed in U.S. Pat. No. 5,185,667, issued Feb. 9, 1993 to Zimmerman, incorporated herein by reference. Zimmerman's apparatus uses the following hardware for processing a captured and digitized image: a microcomputer connected to a remote control, computer control, X-Map and Y-Map; an input image buffer connected to the X-Map and Y-Map with an output connected to an image filter and an output image buffer. This hardware, for example, or any other suitable hardware, may be placed at the[0100]computer184, or theweb site130, but is preferably located at theuser terminals302,304,306,308,310.
As a preferred alternative, the specialized hardware is removed and the hardware functionality is implemented in software at the[0101]computer184 orweb site130, but preferably the software is loaded into theuser terminal302,304,306,308,310. Thus, in accordance with the present invention a spherical (or other wide-angle) image is supplied to the user's terminal, which executes appropriate software (which may be a “plug-in” for a browser application program) for displaying a segment of the image (or video) without distortion. Additionally, the distorted spherical image (or video) may be saved to a storage medium, either at the user's terminal or at the web site, for future loading and viewing.
FIG. 8B also shows how to remove the lens distortion without special processing. As shown in FIG. 8B, a spherical (or other wide angle)[0102]lens182 is in optical communication with avideo camera180′. However, anonlinear imaging sensor186 is placed between thespherical lens182 and thevideo camera180′. The imaging sensor is designed to provide a distorted output which cancels out the distortion of thespherical lens182, and thus an undistorted wide-angle image is provided tovideo camera180′. Alternatively,imaging sensor186 may itself provide a digital output, making it unnecessary to use acamera180′. In this case, theimaging sensor186 would be directly connected tocomputer184.
Examples of[0103]imaging sensors186 are disclosed in U.S. Pat. No. 5,489,940, issued on Feb. 6, 1996 to Richardson et al., and in PCT publication WO 96/12862, published Jun. 13, 1996 to Richardson et al., each incorporated herein by reference. Other suitable imaging sensors may be used with the present invention.
The image obtained by the[0104]imaging sensor186 may be undistorted and not require further processing. A segment of the image may then be selected for display by simply passing the image data to a display device. If the imaging sensor is imperfect, further processing may occur to correct for defects in the sensor. Additionally, further processing for “zoom” and “unzoom” functions may occur. This further processing may take place at theweb site130 or at the user'sterminal302,304,306,308,310.
The embodiments of FIGS. 5 through 8 may be used in conjunction with either live audio and video or prerecorded video data (with audio) (shown in FIGS.[0105]1-3). For example, if nothing interesting is happening at the watering hole, a connected user may access a stored audio and video clip of a lion attack which occurred the day before. If “perceived” camera control is utilized, the stored audio and video preferably includes all camera angles (or a wide-angle view), such that the ability to pan and zoom is preserved.
5. Web Site Configuration.[0106]
FIGS. 9A and 9B show a more detailed view of the web site, listed as web site[0107]140 (FIG. 3), but which may also correspond to web sites112 (FIG. 1) and130 (FIG. 2). Theweb site140 is connected to adata communication network120, theInternet242, anddirect connections244. The web site containstransmission equipment210, receiveequipment220,220,′ twocompression units108,114, aweb server200, arouter230, andcommunication equipment240. Theweb server200 itself contains adigital matrix switch250, a plurality ofdigital video servers252,252′,252″,252,′″ a firewallaccess control unit254, adatabase server256, an audio andvideo storage unit258, adata storage unit260, anadministrative unit262, adigital matrix switch264, acamera control unit268 and a digitalvideo matrix switch270.
The[0108]web site140 is connected to thedata communication network120 bytransmission equipment210 and receiveequipment220. As shown,multiple receivers220,220′ may be used. Also, as shown, the receivers may have more than one video output. Audio and video signals may also be input to theweb server200 by videocassette (or other suitable recorded media) or simply by feeding in television programming. As with FIGS. 1 and 3, these signals are preferably compressed bycompression units108,114. On the opposite side, theweb server200 is connected to remote users by arouter230 andcommunication equipment240, which in turn are connected to theInternet242 or directly connected244 to users. Thecommunications equipment240 outputs the video streams116 through a number of input/output ports.
As previously stated, the[0109]web server200 contains adigital matrix switch250, a plurality ofdigital video servers252,252′,252″,252,′″ a firewallaccess control unit254, adatabase server256, an audio andvideo storage unit258, adata storage unit260, anadministrative unit262, adigital matrix switch264, acamera control unit268 and avideo matrix switch270.
The[0110]digital matrix switch250 receives all incoming compressed video signals from thereceivers220,220′ and thecompressor units108,114. Thematrix switch250 also receives compressed video data fromdatabase server256. Under control of theadministrative unit262, thedigital matrix switch250 outputs the input compressed video signals todigital video servers252,252′,252″,252′″. In this manner, any input signal can be transferred to any video server as directed by the admin unit. Also, stored programming from thedatabase server256 is routed to thedigital matrix switch250 to be switched as if it were incoming live video. The outputs of thedigital matrix switch250 also connect to thedatabase server256, so that anything at the inputs, such as incoming live audio and video, can be stored in thedatabase server256.
The compressed input video is passed into various[0111]digital video servers252,252′,252″,252′″ for formatting. Users who connect toweb server200 preferably run their own decompression software so that the no decompression need occur at theweb server200. As an alternative, the digital video servers may decompress the input video.
The audio and video from the[0112]video servers252 are passed through a second digital (video)matrix switch270. Since switching has already occurred at thedigital matrix switch250, the secondvideo matrix switch270 is not required, but is desired for maximum flexibility. It is also optimal where the number of users exceeds the number of video inputs, as one input may be channeled to numerous connected users.
In a preferred embodiment, the[0113]matrix switch270 may contain a processor that joins different frames of video and audio such that each output contains frames for multiple video pictures (including audio). This enables users to receive split screen images of video and select an audio track for playback (see FIG. 14, discussed below). The split-screen images may be formed by using known methods, which may differ depending on the type of compression used. For example, digital images may be decompressed, combined with other decompressed images, and then re-compressed; or the images may-be decompressed and converted to analog, combined, and then converted to digital and compressed for transmission.
The signals switched by the[0114]video matrix switch270 are preferably digital. This is because the communicatedvideo streams116 are preferably digital. It is preferred to process all the signals in the web server in the digital domain to improve simplicity and maintain maximum flexibility.
The various streams of video output from the[0115]video matrix switch270 are passed to the firewallaccess control unit254 for output to therouter230 and thecommunication equipment240.
Using this system, any user may receive any signal present at any input, including stored signals within audio and[0116]video database258 ordata storage unit260. Additionally, any compressed digital signal present at the input todigital matrix switch250 may be stored in the audio andvideo storage unit258 ordata storage unit260. This is advantageous in the perceived camera control embodiment (FIGS.4-8) where theweb server200 must output a different video picture to the user upon user request. When the user request is received by theweb server200, theadministrative unit262 directs the matrix switches250 and270 to output the correct video stream to the user. If the user is requesting stored video, the administrative unit directs thedatabase server256 to provide the video todigital matrix switch250. If graphics or textual data are required, theadministrative unit262 directs thedatabase server256 to output the text or graphics todigital matrix switch264.
Although shown as one functional box, the[0117]database server256 may be implemented by using several servers and/or multiport servers. The audio andvideo storage unit258 anddata storage unit260 may be implemented by using many storage media of different types, such as optical storage devices (i.e., CD-ROM), magnetic disks, magnetic tape, or memory circuits (i.e., RAM/ROM). The number of units depends on the amount of stored data, the number of users, and the desired output speed. Thedatabase server256 may be one or multiple units. The audio andvideo storage unit258 stores (preferably compressed) audio and video presentations, including all relevant camera angles. Thevideo servers252 may also be implemented as one or more servers and/or multiport servers.
The[0118]data storage unit260 is used to store information relating to audiovisual displays. This information relates to the menu structure and screen displays communicated to connected users. The stored information may also relate to specifically to the audio and video that is currently being displayed and heard. For example, in the demolition embodiment of FIG. 5, a user may click on a “more info” icon, to obtain information on demolition. Such information, which could include statistics on dynamite, for example, would be stored as text or graphics indata storage unit260. The “more info” command would be transmitted to thecommunications equipment240, pass through therouter230, and thefirewall access control254 toadministrative unit262. Theadministrative unit262 then directs thedatabase server256 to recall the relevant information, such as statistics on dynamite, fromdata storage device260 and pass the information todigital matrix switch264. The recalled information is then passed to the firewallaccess control unit254, therouter230, and thecommunication equipment240 for transmission to the proper subscriber. The data may be combined with audio and video in the firewallaccess control unit254, or be a separate transmission.
In the perceived camera control embodiment, the[0119]communication equipment240 forwards the user's command (such as “pan right”) to therouter230, which detects the command and forwards it to the firewallaccess control unit254, which passes it to theadministrative unit262. Theadministrative unit262 controls the video being fed to each connected user. Theadministrative unit262 also responds to user commands by instructing either thematrix switch250 or thematrix switch270 to pass a different audiovisual signal from another source (i.e., camera, for example, the camera to the right of the present camera) to the connected user. If the user is receiving a stored image fromdatabase258, the administrative unit instructs thedatabase server256 to recall the appropriate video signal.
In the actual camera control embodiment (shown in FIGS. 3 and 7), commands from the user (such as “pan right”) are received by the[0120]communication equipment240 and forwarded to therouter230. The commands enter theweb server200 via the firewallaccess control unit254, and are passed to theadministrative unit262. The commands may be stored in theadministrative unit262 or passed to thedatabase server256. Either way, the commands pass through thecamera control unit268 that formats the commands as necessary for remote camera control. The formatted commands are passed to thetransmission unit210. Thetransmission unit210 provides the commands todata communication network120 for reception at remote cameras and CPU134 (FIG. 3).
In the spherical (or other wide angle) lens embodiment (shown in FIGS. 8A and 8B), where the remote camera uses a[0121]spherical lens182, theadministrative unit262 determines which segment or quadrant of the audiovisual image is to be supplied to the user in response to the user's command. In this embodiment, the spherical image is stored indatabase258 prior to being output todigital matrix switch250. The image is split into a number of sections, which when combined form the entire 180° sphere. By using suitable image processing software, the distortion is removed or minimized in each segment. Theadministrative unit262, in response to a user command, determines which segment of the sphere should be sent to the user. The administrative unit then directs thedatabase server256 to retrieve and output the correct segment to thedigital matrix switch250. By controlling thedigital matrix switch250 andvideo matrix switch270, theadministrative unit262 is able to ensure that the user receives the correct segment of the spherical image.
However, as previously stated, in one preferred embodiment the entire spherical (or other wide angle) video is communicated to the user, and the distortion removed by software at the user's terminal. This minimizes the complexity of the processing necessary at the[0122]web site140, and allows the user to store the entire spherical (or other wide angle) video.
Preferably, the[0123]communication equipment240 is designed to automatically determine the maximum data rate at which information can be transmitted to the connected users. The data rate depends on the type of connection the web sites has with the user, and the type of equipment the user is operating. In one embodiment, the communications equipment uses the maximum data rate possible as sensed from the user's communications. Alternatively, users may enter their data rates when prompted by a menu screen, as shown in FIG. 15 and described below. The data rates are then stored incommunications equipment240. Thecommunications equipment240 may also compress the video streams prior to transmission using any known compression algorithm. Additionally, the communications equipment may remove video frames, preferably prior to compression, such that the resulting data rate is reduced to be compatible with the user.
FIG. 9B is identical to FIG. 9A, but contains an input interface[0124]225 and anoutput interface235. The input interface225 is used to obtain digital video from other sources, such as a paging system, cellular system, cable television system, etc . . . The output interface connects the web site to other communications systems such as paging systems, cellular systems, or cable television systems. In the case where the input interface connects to an analog system, it contains suitable analog to digital converters (not shown). Also, where the output interface connects to an analog system, it contains suitable digital to analog converters (not shown).
For example, the input interface[0125]225 may obtain images or video from a paging system, and the output interface225 may be connected to a paging system to broadcast video or images to a selective call receiver. In this regard, the following publications are incorporated by reference, each of which relates video/images to selective call receivers: PCT Publication No. WO 96/07269, published Mar. 7, 1996 by Jambhekar et al., PCT Publication No. WO 96/21173, published Jul. 11, 1996 by Harris et al., and PCT Publication No. WO 96/21205, published Jul. 11, 1996 by Harris et al.
6. Communication to the User Terminals.[0126]
FIG. 10 shows how the users are connected to the web site, and shows an example of a communications network[0127]125 (FIG. 8B) in detail. The connections shown in FIG. 10 apply to the web sites of the previous figures, including the web site112 (FIG. 1),130 (FIG. 2) and140 (FIGS. 3 and 9). FIG. 10 shows aserver platform200, theInternet242, twodirect connection244, two traditional Internet hosts272,274, two cable Internet hosts276,278, a satellite-basedInternet host280, atelephone dialup282, anISDN channel284, acable plant286,288, a satellite system290 and a plurality ofconnected user terminals302,304,306,308,310.
In operation, the[0128]web site112,130,140 may communicate over theInternet242 to a number of different systems. These systems include atraditional Internet host272,274 and a cableheadend Internet host276. Thetraditional Internet host272,274 may be connected via atelephone line282 or anISDN channel284 to a plurality ofremote user terminals302,304, respectively. Thecable Internet host276 may be connected via acable plant286 to aremote user306.
Alternatively, the web site is connected via a[0129]direct connection244 to a cableheadend Internet host278 or satellite-basedInternet host280. The cableheadend Internet host278 communicates to acable plant288 and aremote user terminal308. The satellite-basedInternet host280 communicates via a satellite290 to auser terminal310. Thesedirect connections244 enable a higher data rate and use a high-speed cable modem.
It is advantageous that the communications equipment[0130]240 (FIG. 9) enables communications with any type of user terminal no matter what the data rate or system. Of course, user terminals with higher data rates will receive higher quality audio and video images.
7. Exemplary Screen Displays and Features.[0131]
FIGS.[0132]11-16 show examples of display pages that are shown at the remote user's terminal. The pages and menus are stores in data storage unit260 (FIG. 9) as graphical and/or textual information.
FIG. 11 shows an example of a home page, using advantages of the present invention. The[0133]home page400 contains a number ofadvertisements402,numerous web links404, asociety link406, options forviewing television programming408, a plurality of rapidaccess entry options409 including a “World Watch Live”option410, and options forclubs412.
The[0134]advertisements402 are useful for the page provider to generate revenue. As described previously, the system is designed such that television programming can be supplied over the Internet. Users may view television programming by selecting the homepage television option408. TheMagazines404 are used to provide information concerning specific topics to the user. Users may join a society, having additional membership benefits, through the “society”selection406. The “World Watch Live”feature410, part of the rapidaccess entry options409, is selected when users wish to watch live video from remote sites. The clubs shown in theclub option412 are selected by users who wish to obtain information related to common areas of interest.
FIG. 12 shows a[0135]society menu406, selected from the FIG. 11 home menu page. As shown in FIG. 12 there are options for “World Watch Live”420, there is anadvertisement402,subscription information424, andnumerous club options422. This screen and all the functions selected in response to the displayed options may be provided on a subscription or temporarily free basis.
FIG. 13 shows one example of a “World Watch Live”[0136]menu440. This menu is used to select remote locations from which to observe live or prerecorded video. in this example, a map of the world is presented with sites that are available to select for observing live video. The screen indicates sites that are active442 or underconstruction444. This menu also contains twoadvertisements402.
The “World Watch Live” embodiment allows connected users to visit virtually anyplace in the world to learn more about its culture, geography, or environment. Coupled with perceived or actual camera control and associated pre-stored video, textual and graphical information, a powerful and inexpensive learning tool is realized. This is more closely shown in FIG. 14.[0137]
FIG. 14 shows a menu[0138]450 that corresponds to the Egyptian site in FIG. 13. This screen concerns “Giza, Egypt”, and contains live video from five cameras. As shown in the screen, there is camera one452, cameras two through five454, a “Map”option456, an “About This Site”option458, an “About Egypt”option460, an “Upcoming Events”option462 and a “Remote Control”option464. Camera one452 is the default for the main viewing camera. The user may select video image sizes and the number of images to be displayed, limited by the equipment the user is operating. Video from cameras two through five are supplied along with that from camera one to provide alternative sites and viewpoints about the topic of the screen (i.e., Egypt).
The “Map”[0139]option456 brings the user back to the world map (FIG. 13) to select additional sites. The “About This Site”option458 brings up text, graphics or additional video concerning the site of Giza, Egypt. For example, a professor appears and talks about the origin of the Sphinx (shown by camera1). The embodiment shown in FIG. 16 and described below (interactive lecture) may be combined with the “About This Site” option. Additionally, other video may be displayed in response to selection of “About This Site”. Such video may be a documentary of the Sphinx or discussion about the technology that historians estimate was used to construct the Sphinx.
The “About Egypt”[0140]option460 brings up graphics, text or additional video concerning Egypt. For example, a map of Egypt with population densities may be shown. The option for “Upcoming Events”462 brings graphics, text or video concerning new events in Egypt. For example, text and newspaper articles concerning the construction of new irrigation canals is displayed. “Remote Control”option464 brings up a command menu (such as the “tool bar”151 of FIGS.5A-D) that allows the user to change camera angles or positioning in any of the cameras capable of that effect. The menu would apply to actual or perceived camera control. For example, the user could pan around the Sphinx (camera1, shown at452) to observe it from the front, each side, and back.
Thus, this single screen relating to Egypt provides a wealth of information at a single Internet address (or web site). It is unnecessary for a user to “link” to other locations on the Internet. Audiovisual presentations are displayed, which give the user insight into the people and culture of Egypt. Text, graphics, and additional stored video are available to further educate the user. Camera control (actual or perceived) gives the user the feeling of walking around different locations in Egypt.[0141]
FIG. 15 shows a[0142]screen470 that asks users about their equipment in order to determine the appropriate data rate for communications. Preferably the screen is not needed and the data rate is determined bycommunication equipment240 automatically. Note that anadvertisement402 is also shown on this screen.
FIG. 16 shows an interactive lecture embodiment of the present invention. As shown in FIG. 16,[0143]live video500 of an astronomy professor's lecture is transmitted to connected users. The users are able to ask the professor questions510 and receiveanswers512. Thelive video500,questions510, and answers512 are shown to all connected users. Preferably, the users enter questions via keyboard or microphone. However, if suitable data rates are available, the user may ask a question via video. Thus a split screen video showing both the person asking the question and the lecturer may be presented to all users simultaneously. The answers are preferably given by the lecturer, who may observe the question on a remote display. Alternatively, the answers may be supplied by the web site as text, graphics, or pre-stored video. The answer may pass through a closed captioning device, be encoded, and displayed on the screen in ananswer box512.
Referring to FIG. 9A, questions are sent to the[0144]web site140 as part of the normal user terminal communication. Theweb site140 receives the question at thecommunications equipment240 and forwards the question throughrouter230 and the firewall/access control unit254 to theadministrative unit262. Theadministrative unit262 determines whether the question can be answered by playing stored video or showing stored text or graphics. If so, theadministrative unit262 directs thedatabase server256 to recall the appropriate information. The information is then output through the matrix switches250,270 or264, under control of the administrative unit, as appropriate. The ability of the administrative unit to answer questions depends upon the complexity of its software. Simple, pre-stored answers to frequently asked or standard questions may be provided in a basic system. More advanced systems may utilize an interpreter to analyze the question before providing an answer. For example, frequently asked questions in the astronomy field may be “what is a star?” or “how was the galaxy formed?” In response to these questions, which may even be provided on a menu or list, the administrative unit recalls pre-stored answers in either video, text, or graphics.
If a question cannot be answered by the administrative unit, or is sent directly to the remote lecturer, the question proceeds to the remote lecturer in a similar fashion as the camera control signal (FIG. 3) discussed previously. However, in the interactive lecture embodiment, the camera control unit[0145]268 (FIG. 9) is replaced with a question format unit (not shown) that reformats the question under control of theadministrative unit262.Transmitter210 then transmits a question signal to the location of the remote lecture via thedata communication network120 and thecommunication paths126,128. The lecturer has a display that shows questions received over the data communication network.
In an alternative embodiment, the lecturer or a number of assistants may select from among many pre-stored answers in response to a question. In this embodiment, the remote lecturer has a computer and monitor (not shown) that displays the questions and the available pre-stored answers. The lecturer or assistants then match answers with the questions. The pre-stored answers are preferably forwarded to the individual who asked the associated question. In order for others to learn from the questions, the questions and answers may be provided to all connected users.[0146]
FIGS.[0147]17-18 shows an embodiment of the invention using a combination of live video, stored video, stored graphics, camera control and interactive questioning. Thelive video550 ofcamera1 shown in FIG. 17 relates to a geological site, i.e., the geyser, “Old Faithful”. Since the site is located on a National Park, the display screen has been customized to allow for the selection “About National Parks”604. When this is selected, the user's command is communicated to theweb server112,130,140 for analysis by theadministrative unit262. TheAdministrative unit262 determines that pre-stored video and graphics are required, and instructs thedatabase server256 to output the correct information: video to thematrix switch250, and graphics to thematrix switch264. The matrix switches250,270, and264, under control of theadministrative unit262, forward the video and graphics to the user through thecommunication equipment240.
FIG. 18 shows the result at the user terminal. The communicated pre-stored video[0148]560 of a Park Ranger appears on the screen. The Park Ranger discusses the topic of National Parks. The discussion occurs in conjunction with a graphical display of the locations of all National Parks, shown at the screen location570.
The user may select other options, such as “[0149]Map600” to return to the map of all remote sites, “About This Site”602 to learn more about the site currently viewed, “More About National Parks”614 for even more information about National Parks, “Upcoming Events”606 for a schedule of upcoming events, “Remote Control”608 for remote (either actual or perceived) control of the camera (i.e., camera1), “Ask Questions”610 for asking questions (as in FIG. 16) to an on-line Park Ranger, and “Other Topics”612, for a list of other topics and/or options.
Therefore, the present invention provides an easy and fun way to learn, by combining live video, pre-stored video, graphics and text with interactive questioning and actual or perceived camera control.[0150]
8. Surveillance Systems.[0151]
The present invention may be used in a surveillance or tracking system. For example, a researcher may place a video camera in the center of a watering hole, preferably connected to a video recorder for storing many hours of activity at the watering hole. Preferably multiple cameras or a wide-angle lens are used such that virtual camera control (as described previously) may be performed on the video. Such a surveillance system has many advantages.[0152]
First, the system allows for automatic scanning of the surveyed area, without the need for moving any cameras. Additionally, multiple segments of the area under surveillance may be viewed at the same time in a split-screen image. All that needs to be done is the removal of distortion in multiple segments of the video (if using a wide-angle lens). The disclosure of U.S. Pat. No. 5,359,363, issued Oct. 25, 1994 to Kuban et al., incorporated herein by reference, discloses one example usable with the present surveillance system.[0153]
Second, automatic monitoring and/or tracking may be performed. Often, researchers and photographers wait through long periods of inactivity before a desired event occurs. For example, a photographer may wait for hours for a lion or other wildlife to approach the photographer's position. The present invention may be used to automatically monitor a remote region for activity. In this case, a processor may monitor the multiple cameras or the digital wide-angle video for pixel changes indicating the desired event. For example, an approaching lion in an otherwise inactive desert environment will cause a moving pattern to form on a camera's output or in the wideangle image. A processor may detect the pattern and alert a wildlife researcher that an event is occurring.[0154]
Further, the processor may automatically and continually display the relevant camera output, or the segment of the wide-angle image containing the lion, thereby tracking the lion. Thus, the present invention may employ tracking techniques, known in the prior art, to the obtained digital image.[0155]
In the monitoring and tracking embodiment of the present invention, it may be desirable to remove the distortion from the wide-angle image prior to performing the processing to determine whether an event is occurring. The type of event being monitored and nature of the object being tracked controls whether monitoring and/or tracking may be performed on the distorted or undistorted image. One of ordinary skill in the art will choose the system best suited for the particular monitored event or tracked object.[0156]
FIG. 19 shows a flow diagram of a monitoring and tracking system using the present invention. The software necessary to perform the monitoring/tracking functions may be located at the web site or at the user's terminal. The image/video signal to be processed for monitoring and/or tracking may be a live video feed or be played back from stored video. Thus, a wildlife scientist may leave multiple video cameras running overnight (or a single video camera with a wide-angle lens) and when the video-tape is played back, the segments/cameras containing activity are displayed.[0157]
Referring to FIG. 19, an “input frame of reference”[0158]routine700 is executed. This routine is optional, and is used to establish a frame of reference direction, such as north. The frame of reference may determine the first segment of a wide-angle image to view, or the first camera to view. Next, a “reset segment counter” routine710 is executed. This sets the segment or camera to be first displayed.
Each segment or camera is viewed only for a limited time, prior to viewing the next segment or camera. Thus, a “reset timer” routine[0159]715 is executed to reset the interval when segments or cameras are switched.
Next, the “obtain image” routine[0160]720 is executed. This routine obtains the wideangle image (live or prerecorded), or images from all the cameras (in the multiple camera perceived control embodiment of FIGS. 4 and 5). The obtained image from a wideangle lens may be processed to remove the distortion or not, depending on what is being monitored.
The obtained image is processed to determine active areas (cameras or segments). Active areas are areas where the processor determines that activity is taking place, either by changes in the pixels at those locations, by using other known image/video processing techniques, or by using external sensors. The processing is performed as known in the art and is not described further herein. The processing occurs during the “process for activity”[0161]routine730. This routine uses the frame of reference to determine which segment(s), relative to the normal (i.e., north) is/are active.
If activity is present, the “display active segments” routine[0162]750 displays the active segments or cameras on a display. Distortion from the relevant segments is removed in the wide-angle lens embodiment. If more than one segment is active, a split screen display may show the each segment simultaneously. The each split screen display may make reference to the frame of reference that was previously entered duringroutine700. The “reset timer” routine710 is then executed so that the last segment under view is returned when activity is no longer present.
If activity is not present, the “display current segment”[0163]routine760 is executed. This routine displays the current segment or camera until the timer expires, at which point the next segment or camera is displayed. The display may make reference to the frame of reference that was previously entered duringroutine700.
After displaying the current segment or camera, the “time limit exceeded”[0164]routine770 is executed. If the time limit has not been exceeded, a branch to the “obtain image” routine720 occurs and processing continues until the time limit is exceeded, or until activity occurs. In an “autopan” embodiment (FIG. 5) the time limit value may be increased by pressing the “−” button in conjunction with the “speed” button (FIG. 5), for a slower autopan, and the time limit may be decreased by pressing the “+” button in conjunction with the “speed” button (FIG. 5) for a faster autopan.
If the time limit is exceeded, the, the segment (or camera) counter is incremented by the “increment segment counter”[0165]routine780. If the counter is greater than the maximum number of cameras or segments, the “counter>max” routine790 branches to the “reset segment counter” routine710, to restart the automatic panning. If the counter is not greater than allowed, a branch occurs to the “reset timer” routine715 so that the next segment or camera may be displayed, and processing for activity continues.
Thus, the flow chart of FIG. 19 allows for automatic panning and for automatic tracking. If the “process for activity” routine[0166]730, the “activity?”test740, and the “display active segments” routine750 were removed, the “autopan” function described previously and shown with respect to FIG. 5 would be achieved. In this case, “display current segment” routine760 would follow “obtain image”routine740.
Monitoring and automatic panning may be combined. When combined, all active segments or cameras are automatically panned for a brief timeframe. Thus, if a lion and zebra were both moving towards the camera from opposite direction, each would be displayed for a brief timeframe before switching to a display of the other. This is an alternative to the split screen display previously described.[0167]
9. Display of Video Data.[0168]
In certain embodiments of the present invention, the user may select or be provided data concerning the video currently displayed. For example, superimposed on the video may be the date and time the video was recorded, a name of the image location, remaining time for the video, or data pertaining to the segment (or camera source) of the video which is currently being viewed.[0169]
This segment/camera data may be a compass heading (such as north) or angle from a reference (such as 40 degrees), or coordinate information (such as X/Y, X/Y/Z, R/θ, X/R/θ etc . . . ) relating to the location of the center of the segment/video currently displayed in relation to the wide angle image or other cameras. A graphical representation of the lens (or layout of the cameras) may show which segment of the wide-angle image (or camera) is being displayed. In order to display the image segment, a frame of reference may be adopted, especially for a spherical lens. The frame of reference would be either generated by a processor at the web site or user's terminal, or is entered by a user or operator. For example, the user may select which direction is “north” or position the axis of a coordinate system if a coordinate display is to be used for a particular lens.[0170]
Additionally, the image's magnification and its density/colors may also be shown on the display, such as “magnification=10×, picture density=200×200 pixels, 64 colors.”[0171]
The display of image data may be used in all embodiments of the present invention, and are preferably updated when the displayed image changes.[0172]
FIG. 20 shows an[0173]exemplary display800 showing acoral reef805 where users have virtual camera control via multiple underwater cameras. On thescreen807, thedate810 is displayed along with thetime820. The location is shown at830 and the remaining time of the program at840. The magnification is shown at850 and the density and colors at860. Thesegment camera field870 shows that the user is viewing camera no.3. This segment/camera data may be shown graphically, as depicted at880.Field880 is a top view of thecoral reef805 and the layout of the cameras, in thiscase cameras1 through10. The square around camera no.3 indicates that this camera is the source of the picture on thedisplay800. The frame of reference (north) is indicated at890 for the graphical segment data and895 for the video data.
10. Storing Video and Interactive Presentations.[0174]
The images, video, and image data may also be stored at the user's terminal (or receiving apparatus). Preferably, the wide angle distorted image is stored, along with the image data, if present. Storage of the image and image data enables the user to retrieve the image and view a segment at a later date. Optionally, the entire interactive presentation may be stored at the user's terminal (including associated graphics, text, video, data, or other information), although all the pertinent files and data would have to be received by the user.[0175]
The disclosure of PCT Publication No. WO 96/08105, published Mar. 14, 1996 by Labun, incorporated herein by reference is related to storing images and may be used with the present invention.[0176]
The video or image may be stored in either its distorted or undistorted state. Storing the video or image in its undistorted state has the advantage in that tall and/or wide pictures may be stored in their most viewable state, and in that editing may be performed on the images more easily if they are retrieved with the distortion removed.[0177]
11. Broadcast Television and Cable Television.[0178]
The perceived camera control of the present invention may also be used in the field of broadcast television or the field of cable television. Rather than supply the wideangle images (FIGS. 8A and 8B) to terminals via the Internet, a transmitter may broadcast the images to television receivers. The television receivers are equipped with decoders to decode the wide-angle image as, for example only, disclosed in U.S. Pat. No. 5,384,588, issued Jan. 24, 1995 to Martin et al., incorporated herein by reference. The broadcast television transmitter (not shown) may be connected to remote cameras[0179]104 (FIGS.1-3), output interface235 (FIG. 9B), Internet hosts272,274,276,278,280 (FIG. 10),communications media120,125 (FIG. 8B), or even a user'sterminal302,304,306,308,310 (FIG. 10).
In the field of cable television, a separate decoder or a cable set top converter box contains the appropriate decoding circuitry. A cable television transmitter is connected to remote cameras[0180]104 (FIGS.1-3), output interface235 (FIG. 9B), Internet hosts272,274,276,278,280 (FIG. 10),communications media120,125 (FIG. 8B), or even a user'sterminal302,304,306,308,310 (FIG. 10).
U.S Pat. No. 5,559,549, issued Sep. 24, 1996 to Hendricks et al., incorporated herein by reference, discloses a cable television system using an[0181]operation center1000,network controller1020, concatenated cable system (unnumbered), and settop terminals1030. The cable television system is preferably, digital, and may easily interact with the present invention.
FIG. 21 shows the interaction between an embodiment of the[0182]present invention900 and, for example, thegeneral system910 of the Hendricks et al. '549 patent. Digital signals from the present invention, relating to ordinary video, stored video, wide-angle video, video from multiple cameras, information of any type and interactive presentations may be provided to various elements of the Hendricks et al. '549patent910. It is understood that such digital signals may be supplied to corresponding elements of traditional analog and digital cable television systems that accept digital signals at an input (i.e., stand-alone or using a digital to analog converter).
Specifically,[0183]digital video920 fromremote camera104 and remote wide-angledigital video930, processed/compresseddigital video940 fromcomputer184,video950 fromcommunication network120, streamedvideo960 fromweb site140,video970 fromcommunications network125, andvideo980 from the user terminals (i.e.,302) may be communicated to the digital cable television system of the '549 Hendricks et al patent. These video signals may be received by either theoperations center1000,satellite1010,cable headend1020, or settop terminals1030 of the '549 Hendricks et al patent.
Likewise, the[0184]operations center1000,satellite1010,cable headend1020, and settop terminals1030 may communicate digital signals to the Internet structure of the present invention. Specifically, these communicated signals may be received by theremote computer184, data communication network120 (including web site130),data communication network125, and user terminals (i.e.,302).
U.S. Pat. No. 5,600,573 to Hendricks et al, incorporated herein by reference, discloses an operations center with a file server. This operations center may substitute for the[0185]operations center1000 shown in FIG. 21.
U.S. Pat. No. 6,201,536, issued Mar. 13, 2001, incorporated herein by reference, discloses a network manager for a cable headend. This network manager may be included in the[0186]cable headend1020 shown in FIG. 21.
Thus, the present invention is capable of fully integrating with cable television systems able to transmit and receive digitally. The present invention breaks down the barrier between television networks and computer networks, allowing for a single integrated programming system.[0187]
It will be appreciated by the artisan of ordinary skill that other aspects of the patent applications, patents and publications incorporated herein by reference may be applied to the present invention. As such, the patent applications, patents and publications are incorporated herein in their entirety. The terms and descriptions used herein are set forth by way of illustration only and are not meant as limitations. Those skilled in the art will recognize that numerous variations are possible within the spirit and scope of the invention as defined in the following claims.[0188]