Movatterモバイル変換


[0]ホーム

URL:


GB2552150A - Augmented reality system and method - Google Patents

Augmented reality system and method
Download PDF

Info

Publication number
GB2552150A
GB2552150AGB1611911.7AGB201611911AGB2552150AGB 2552150 AGB2552150 AGB 2552150AGB 201611911 AGB201611911 AGB 201611911AGB 2552150 AGB2552150 AGB 2552150A
Authority
GB
United Kingdom
Prior art keywords
video
image
mode
captured
display
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
GB1611911.7A
Other versions
GB201611911D0 (en
Inventor
Camp Philip
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sony Interactive Entertainment Inc
Original Assignee
Sony Interactive Entertainment Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sony Interactive Entertainment IncfiledCriticalSony Interactive Entertainment Inc
Priority to GB1611911.7ApriorityCriticalpatent/GB2552150A/en
Publication of GB201611911D0publicationCriticalpatent/GB201611911D0/en
Priority to PCT/GB2017/051180prioritypatent/WO2018007779A1/en
Publication of GB2552150ApublicationCriticalpatent/GB2552150A/en
Withdrawnlegal-statusCriticalCurrent

Links

Classifications

Landscapes

Abstract

An augmented reality system having a video camera to capture video images of a real-world scene at a first frame rate, a display operable in a real-time mode and a slow-motion mode and a sensor to detect the orientation of the display. The real-time mode may display live footage, and the slow-motion mode displays the real-world images at a second frame rate slower than the first frame rate. The slow motion images are cropped and are less than the full captured images and the displayed slow motion image is responsive to the detected orientation of the display. In the slow motion mode, live footage may be stored in a buffer. The application states that: In principle it is not possible to provide the ability to look around in the slow-motion mode because the slowed down video can no longer track the users movements in real time; therefore the perceived advantages of the system are that the cropped images displayed in the slow motion mode allow a degree of head movement whilst in the slow motion mode within the limits of the cameras field of view at the corresponding time when the footage was captured.

Description

(71) Applicant(s):
Sony Interactive Entertainment Inc.
1-7-1 Konan, MinatoKu 108-8270, Tokyo, Japan (72) Inventor(s):
Philip Camp (74) Agent and/or Address for Service:
D Young & Co LLP
120 Holborn, LONDON, EC1N 2DY, United Kingdom (51) INT CL:
H04N 5/262 (2006.01) G02B 27/01 (2006.01) H04N 5/783 (2006.01) H04N 7/18 (2006.01) (56) Documents Cited:
US 9360671 B1 US 9143693 B1
US 20080276178 A1 (58) Field of Search:
INT CL G02B, G06F, H04N
Other: EPODOC, INTERNET, TXTE & WPI
Title of the Invention: Augmented reality system and method
Abstract Title: Augmented reality system for creating a slow-motion effect
An augmented reality system having a video camera to capture video images of a real-world scene at a first frame rate, a display operable in a real-time mode and a slow-motion mode and a sensor to detect the orientation of the display. The real-time mode may display live footage, and the slow-motion mode displays the realworld images at a second frame rate slower than the first frame rate. The slow motion images are cropped and are less than the full captured images and the displayed slow motion image is responsive to the detected orientation of the display. In the slow motion mode, live footage may be stored in a buffer. The application states that: “In principle it is not possible to provide the ability to look around in the slow-motion mode because the slowed down video can no longer track the user’s movements in real time”; therefore the perceived advantages of the system are that the cropped images displayed in the slow motion mode allow a degree of head movement whilst in the slow motion mode within the limits of the camera’s field of view at the corresponding time when the footage was captured.
Figure GB2552150A_D0001
Figure 3 oo
CM ’Μ- co ©
©
Figure GB2552150A_D0002
m co r-
Figure GB2552150A_D0003
Figure GB2552150A_D0004
<=i <
CD
CO
Figure 1
Figure GB2552150A_D0005
2/3
Figure GB2552150A_D0006
280
283
Figure 2
3/3
Figure GB2552150A_D0007
S310
S320
S330
S340
Figure 3
Application No. GB1611911.7
RTM
Date :21 December 2016
Intellectual
Property
Office
The following terms are registered trade marks and should be read as such wherever they occur in this document:
HDMI
FreeBSD
Intellectual Property Office is an operating name of the Patent Office www.gov.uk/ipo
AUGMENTED REALITY SYSTEM AND METHOD
The present invention relates to an augmented reality system and method.
Current augmented reality (AR) systems comprise a means for a user to observe the real world, coupled with a means to augment that view with computer graphics.
Typical devices comprise transparent glasses onto which graphics can be projected, creating a semi-transparent augmentation overlay, or similar glasses with light emitting elements embedded therein, such as a sparse array of organic light emitting diodes (OLEDs), whose comparative brightness when active tend to obscure the scene behind them and so present an augmentation that appears more solid.
Another device that can provide augmented reality is a suitably adapted virtual reality (VR) device. Virtual reality devices typically only display computer graphics and hence to not comprise transparent gasses or a window onto the real world; however, if also equipped with a forward facing video camera, the video image captured by this camera can be displayed to provide a similar AR experience to the other devices. In this case, the video image can be modified directly to include an augmentation layer of computer graphics of any effective degree of transparency.
These devices have the potential to bring information and entertainment into the user’s everyday experience of the real world. However, there are still limits to how this experience can be augmented by such devices.
The present invention seeks to address or mitigate this problem.
In a first aspect of the present invention, an augmented reality system is provided in accordance with claim 1.
In another aspect, a method of video processing for an augmented reality system is provided in accordance with claim 3.
Further respective aspects and features of the invention are defined in the appended claims.
Embodiments of the present invention will now be described by way of example with reference to the accompanying drawings, in which:
Figure 1 is a schematic diagram of an entertainment system in accordance with an embodiment of the present invention.
Figure 2 is a schematic diagram of a head mounted display in accordance with an embodiment of the present invention.
Figure 3 is a flow chart of a method of video processing in accordance with an embodiment of the present invention.
An augmented reality system and method are disclosed. In the following description, a number of specific details are presented in order to provide a thorough understanding of the embodiments of the present invention. It will be apparent, however, to a person skilled in the art that these specific details need not be employed to practice the present invention. Conversely, specific details known to the person skilled in the art are omitted for the purposes of clarity where appropriate.
An example arrangement of an entertainment system comprising a videogame console and a head mounted display unit (HMD) is illustrated in Figure 1.
Figure 1 schematically illustrates the overall system architecture of a Sony® PlayStation 4® entertainment device. A system unit 10 is provided, with various peripheral devices connectable to the system unit.
The system unit 10 comprises an accelerated processing unit (APU) 20 being a single chip that in turn comprises a central processing unit (CPU) 20A and a graphics processing unit (GPU) 20B. The APU 20 has access to a random access memory (RAM) unit 22. The APU 20 communicates with a bus 40, optionally via an I/O bridge 24, which may be a discreet component or part of the APU 20. Connected to the bus 40 are data storage components such as a hard disk drive 37, and a Blu-ray ® drive 36 operable to access data on compatible optical discs 36A. Additionally the RAM unit 22 may communicate with the bus 40. Optionally also connected to the bus 40 is an auxiliary processor 38. The auxiliary processor 38 may be provided to run or support the operating system.
The system unit 10 communicates with peripheral devices as appropriate via an audio/visual input port 31, an Ethernet ® port 32, a Bluetooth ® wireless link 33, a Wi-Fi ® wireless link 34, or one or more universal serial bus (USB) ports 35. Audio and video may be output via an AV output 39, such as an HDMI port.
The peripheral devices may include a monoscopic or stereoscopic video camera 41 such as the PlayStation Eye ®; wand-style videogame controllers 42 such as the PlayStation Move ® and conventional handheld videogame controllers 43 such as the Dual Shock 4 ®; portable entertainment devices 44 such as the PlayStation Portable ® and PlayStation Vita ®; a keyboard 45 and/or a mouse 46; a media controller 47, for example in the form of a remote control; and a headset 48. Other peripheral devices may similarly be considered such as a printer, or a 3D printer (not shown).
The GPU 20B, optionally in conjunction with the CPU 20A, generates video images and audio for output via the AV output 39. Optionally the audio may be generated in conjunction with or instead by an audio processor (not shown). The video and optionally the audio may be presented to a television 51. Where supported by the television, the video may be stereoscopic. The audio may be presented to a home cinema system 52 in one of a number of formats such as stereo, 5.1 surround sound or 7.1 surround sound. Video and audio may likewise be presented to a head mounted display unit 53 worn by a user 60.
In operation, the entertainment device defaults to an operating system such as a variant of FreeBSD 9.0. The operating system may run on the CPU 20A, the auxiliary processor 38, or a mixture of the two. The operating system provides the user with a graphical user interface such as the PlayStation Dynamic Menu. The menu allows the user to access operating system features and to select games and optionally other content.
Referring now to Figure 2, the user 60 is wearing an exemplary HMD 53 on their head. In this example, the HMD comprises a frame 240 formed of a rear strap and a top strap, and a display portion 250 that comprises a separate respective display for each of the user's eyes.
In this example, the HMD of Figure 2 completely (or at least substantially completely) obscures the user's view of the surrounding environment. All that the user can see is the pair of images displayed within the HMD. The HMD has associated headphone audio transducers or earpieces 260 which fit into the user's left and right ears 270. The earpieces 260 replay an audio signal provided from an external source, which may be the same as the video signal source which provides the video signal for display to the user's eyes.
The combination of the fact that the user can see only what is displayed by the HMD and, subject to the limitations of the noise blocking or active cancellation properties of the earpieces and associated electronics, can hear only what is provided via the earpieces, mean that this HMD may be considered as a so-called “full immersion” HMD.
As noted above, in embodiments of the present invention, however, in a first mode the user is provided with a real-time view of the real, ‘outside’ world. This can be provided by projecting a view of the outside (captured using a camera, for example a camera mounted on the HMD) via the HMD’s displays, and by providing a microphone to generate an input sound signal (for transmission to the earpieces) dependent upon the ambient sound.
A front-facing camera 222 may capture images to the front of the HMD, in use. A Bluetooth® antenna 224 may provide communication facilities or may simply be arranged as a directional antenna to allow a detection of the direction of a nearby Bluetooth transmitter.
In operation, a video signal is provided for display by the HMD. This could be provided by an external video signal source 280 such as the PlayStation 4 or a data processing apparatus (such as a personal computer), in which case the signals could be transmitted to the HMD by a wired or a wireless connection 282 or 39. For example, video signals from the front-facing camera may be sent to the PlayStation 4 for processing, and then are relayed back to the HMD for display. Examples of suitable wireless connections include Bluetooth® connection 33. Audio signals for the earpieces 260 can be carried by the same connection. Similarly, any control signals passed from the HMD to the video (audio) signal source, such as position or orientation signals obtained from one or more accelerometers or gyroscopes in the HMD (not shown) may be carried by the same connection. Furthermore, a power supply 283 (including one or more batteries and/or being connectable to a mains power outlet) may be linked by a cable 284 to the HMD. Note that the power supply 283 and the video signal source 280 may be separate units or may be embodied as the same physical unit.
If one or more cables are used, the physical position at which the cable 282 and/or 284 enters or joins the HMD is not particularly important from a technical point of view. Aesthetically, and to avoid the cable(s) brushing the user’s face in operation, it would normally be the case that the cable(s) would enter or join the HMD at the side or back of the HMD (relative to the orientation of the user’s head when worn in normal operation). Accordingly, the position of the cables 282, 284 relative to the HMD in Figure 2 should be treated merely as a schematic representation.
Accordingly, the arrangement of Figure 2 provides an example of a head-mounted display unit (HMD) comprising a frame to be mounted onto an observer’s head, the frame defining one or two eye display positions which, in use, are positioned in front of a respective eye of the observer and a display element mounted with respect to each of the eye display positions, the display element providing a virtual image of a video display of a video signal from a video signal source to that eye of the observer.
Figure 2 shows just one example of an HMD. Other formats are possible: for example an HMD could use a frame more similar to that associated with conventional eyeglasses, namely a substantially horizontal leg extending back from the display portion to the top rear of the user's ear, possibly curling down behind the ear.
The above system provides an illustrative example of an augmented reality system.
More generally, in an embodiment of the present invention, the augmented reality system comprises a video input adapted to receive video images of a real-world scene captured by a video camera at a first frame rate; a display operable to display to a user video images originating from the video camera at a first resolution; a sensor operable to detect an orientation of the display; and an image processor operable to process video images captured by the video camera.
These components may be integral to or inserted/attached to a head mounted unit (HMD) worn by a user, or (for example in the case of the image processor) may be remote to the head mounted unit but in communication with it. It will be appreciated that the video camera may be a stereoscopic video camera providing left-eye and right-eye video images; consequently subsequent references to a video camera herein may be assumed to encompass such a stereoscopic video camera unless otherwise stated, and similarly subsequent references to an image or video image can be assumed to encompass stereoscopic images unless otherwise stated.
Typically the video camera is integral to the head-mounted unit so that the captured video images track the head movements of the user. In this case, the video input is internal to the system and refers to the link between the camera and the processor. Alternatively however, the video camera may be attachable to the head mounted unit to achieve a similar effect but may optionally be reusable for other purposes. Hence for example the camera may be part of a mobile phone that can be inserted into the head mounted unit, or may be a webcam or sports video camera designed to be mountable on headgear. In this case the video input will typically be a wired or wireless interface with the camera.
The display may comprise two separate display units for the left and right eyes, or may comprise a single display unit logically split into left and right eye display regions. Again the display may be integral to the head mounted unit, or may be attachable (for example in the form of a display on a mobile phone that is inserted into the head mounted unit).
The sensor may comprise one or more from a list consisting of an accelerometer, gyroscope or optical tracking means, selected to allow detection of the orientation of the head mounted unit (and by extensions, the display within the head mounted unit) as the user moves their head around. Optionally lateral movement horizontally, vertically and forwards/backwards may also be detected in addition to rotation. The sensor may be integral to the head mounted unit, or may be part of a detachable device such as a mobile phone inserted into the head mounted unit.
The image processor may comprise one or more central processing units or one or more graphics processing units, or any combination of these either as separate components or a single component (for example, the APU 20 of the PlayStation 4). At least one processor of the image processor may be integral to the head mounted unit, or may be attachable, for example in the form of one or more processors of a mobile phone attached to the head mounted display; similarly at least one processor of the image processor may be remote from the head mounted display, for example in a videogame console operating in conjunction with the head mounted display. Hence the image processor may comprise on or more processors located as applicable integrally or detachably within the head mounted unit, or remotely in an associated computer.
In an embodiment of the present invention, the augmented reality system is operable in two modes. In the first mode, the video camera is operable to capture video images of a realworld scene at a first frame rate (for example 60fps, but equally it may be 50, 30, 25 or any other suitable rate as the first frame rate), and the display is operable to display to a user video images originating from the video camera at a first resolution, and typically at a frame rate corresponding to the first frame rate (for example if the first frame rate is 60fps, then showing each captured image at 60fps, or showing every other captured image at 30 fps, which is the effective capture rate for alternate frames). As such, the user is presented with a real-time view from the video camera feed. The view may correspond to the whole of the captured video image, or may correspond to only a portion of the captured video image, for example where there is a difference in aspect ratio between the resolutions of the camera and the display. Optionally, the presented video images are augmented with computer graphics to generate an augmented image.
However, the augmented reality system is also operable in a second mode. In this second mode, the captured video is sent to a buffer, and the buffered video images originating from the video camera are displayed to the user at a second, different, slower rate than the capture rate.
This creates a slow-motion effect for the real-world scene perceived by the user, optionally also augmented with computer graphics as in the first mode. For example, if the first frame rate was 60fps, then the second frame rate for showing each captured image may be 30fps, representing a 50% slowdown. As noted above, this could be achieved by changing the frame rate for each frame from 60fps to 30fps, of by switching from showing alternate frames captured at 60fps at 30fps to showing each frame at 30fps (which may reduce processing and display overheads during the majority of the time when the device is in the first mode).
In the event that the desired slow-down would result in an effective second frame rate that is below an acceptable threshold (for example, a 75% slowdown resulting in an unacceptable 15fps output), then optionally the image processor is operable to perform inter-image interpolation to generate one or more processed video images for presentation in between processed video images corresponding to respective captured video images. In this way, the rate of images directly derived from the capture video can still be reduced by 75% whilst increasing the rate of images presented to the user to an acceptable rate. For example, for a capture rate of 60fps, and a slowdown of 75% to a slow motion frame rate of 15fps, if one image was interpolated between each original image, then an acceptably smooth presentation frame rate of 30fps can be achieved whilst still maintaining the 75% slow down.
Alternatively or in addition, the video camera may be operable to capture video images at a third, higher frame rate than the first frame rate. For example, increasing the capture frame rate to 120fps would enable a 75% slow down to 30 fps without interpolation.
Hence more generally, the image processor is operable to present to the display successive processed video images corresponding to respective captured video images at a second, lower frame rate than the first frame rate, thereby creating a slow-motion effect.
As was noted previously, a characteristic feature of virtual reality and augmented reality is the ability to look around a scene by moving your head to change the orientation of the head mounted unit. In the first mode of operation, this is straightforward to achieve because the camera is generating a live feed from which images can be obtained for presentation to the user in real time.
However, in slow motion, the live feed is buffered and presented to the user at a reduced rate; this means that movements of the user’s head will not be reflected in the next image from the camera that is being supplied to the user, since that image was obtained earlier than the head movement, and supplied to the buffer. Hence in principle it is not possible to provide the ability to look around in the second mode of operation because the slowed down video can no longer track the user’s movements in real time.
However, in an embodiment of the present invention, in the second mode the image processor is operable to process a captured video image so that it corresponds to a portion of the full captured video image that is less than the full captured video image.
Hence for example the image processor may only present a 75% x 75% central portion of the image, compared to either the original image or the portion of the image presented in the first mode if this is already cropped for reasons of aspect ratio or the like.
Consequently, the position of the portion of the captured video image being processed to generate the current processed video image for display is responsive to the detected orientation of the display.
In other words, when the user subsequently moves their head, the selected portion pans within the uncropped/original/larger image in a direction corresponding with the direction of head movement, to give the illusion of being able to look around the view.
It will be appreciated that this illusion will break once the user’s head movements correspond to a position where the selected portion would extend beyond the limits of the original captured image, but within these limits, the illusion of time-dilated AR can be maintained.
It will be appreciated however that the smaller the size of the selected portion relative to the original captured image, the greater the relative amount of head movement that can be accommodated before the illusion is broken. However, there is also a corresponding trade-off between image portion size and the effective resolution of the presented image if this portion is blown-up to occupy the full display for the user.
Accordingly, in an embodiment of the present invention, in the second mode the image processor performs intra-image interpolation to re-scale a portion of a captured video image to the first resolution to generate a processed video image.
It will similarly be appreciated that a sudden switch from the first mode, with a first field of view and a real time presentation, to a second mode with a significantly smaller field of view and a slow motion presentation could be disorientating for a user.
Accordingly, in an embodiment of the present invention the image processor is operable to present a transition phase in which the portion of the full captured video image changes over successive presented images from a first size portion to a second, smaller size portion. Hence for example it may transition from 100% to 75%, or some other size (66%, 50%, 33% or any other suitable size). Typically this transition may take between 1/10 and 1 second, but may be faster or slower.
In the above example, it was assumed that the selected portion was initially central and/or that the portion at the end of a transition phase was similarly central. However, in an embodiment of the present invention, the position of the portion (or the portion at the end of a transition phase) may be selected responsive to a predetermined feature of either the captured video image being processed, or a computer generated augmentation generated for superposition over the presented processed image.
Examples of such features in a capture image or image sequence include a face (or a specific recognised face), a fiduciary marker that may be a target for an augmentation, an object being tracked by an AR application, an element of the image that appears to be moving, or moving in a different direction to the overall image motion, or moving at a speed within the image above a predetermined threshold, or an element of the image having a brightness or colour that is either above or (separately) below a threshold, or a threshold difference for a mean of the image. Examples of such features in a computer generated augmentation include the appearance of a new virtual object (for example a ball, or an explosion / gun shot hit or the like); or a change of state for a virtual object (for example the ball crossing a goal line, or a virtual character pulling a taunting expression).
It will be appreciated that not every instance of one of these features may trigger a transition between first and second mode.
Alternatively to the above, the user may be presented with only a portion of the original captured image in both first and second modes, so that the switch between modes is not accompanied by a change in the field of view. As a non-limiting example, the images presented to the user in both modes may comprise the central 75% x 75% portion of the captured image, before any panning of that portion in the second mode.
In other words, in this embodiment then by default the camera has a wider field of view than the images presented to the user in either mode.
This may be the case for example, if the video images are received from a remote camera that is used by plural AR users, for example as spectators at a sports match; this enables each user to look around independently of the others, since no actual motion of the camera is required. Optionally the camera may have a much higher resolution than the augmented reality system display; for example, the camera may be a so-called 4K UHD camera, whilst the display is a 1080 or 720 HD display. This enables presentation of good quality images whilst still permitting the user to look around, at least to some extent. In a variant of this embodiment, the orientation of the user’s display is transmitted to a broadcast server, so that only the relevant portion of the 4K captured image is transmitted to the respective user, thereby saving bandwidth. Optionally a small buffer region of additional image may be sent so that the selected portion can be slightly adjusted into such a region to account for any incremental motion of the user / display during the frame transmission time.
Given such an arrangement, the transition from first to second mode may be triggered remotely by the broadcaster, and the portion may also be selected by the broadcaster, for example to let the user see if a ball was out in a tennis match, or if a foul was committed in a game of football. It will be appreciated therefore that such a system could be used to present highlights to all viewers, and also by individual viewers to focus on features of personal interest during any suitably adapted broadcast, whether in sports, news, the arts or any appropriate field.
Separately to the above, the rate of slow-down may also transition from the first frame rate to the second frame rate over a finite period, which would typically correspond to that of the portion transition.
In any of the above cases, the transition between fields of view or effective rates of time may also be accompanied by a characteristic sound effect that the user can learn to associate with the transition from first to second modes, such as a dropping audio tone evocative of a recording slowing down.
Alternatively or in addition, the augmented reality system may comprise a microphone (for example integral to the head mounted unit, or attachable thereto, for example as part of a mobile phone that can be attached to the head mounted unit, or remote, for example associated with a remote camera in a broadcast scenario).
Then, in the first mode, external sounds are captured by the microphone and presented to the user in real time, but in the second mode, the sounds are processed to slow them down by an extent corresponding to the change in rate between the first video capture rate and the second video capture rate.
This can provide an additional level of realism to the user’s experience. Optionally stereo balance or a directional mix of the recorded sounds can be adjusted responsive to the user’s head motions (and by extension, the orientation of the display).
The above apparatus and techniques enable an augmented reality system to provide a slowmotion or time dilation view of the external, real world, in which the user is still able to look around such that their viewpoint changes in real time, at least to some extent, within this slow-motion representation.
However, by decoupling the user from a real-world video feed (particularly in the case of a feed from a camera associated with their own head mounted unit, there is a risk that the user will no longer know what is currently happening in the outside world.
Accordingly, in an embodiment of the present invention, the second mode can be terminated in the event of one or more circumstances, including but not limited to:
detection of a predetermined event within a currently captured video image not yet presented for display, such as detection of movement within the video towards the user (for example an object appearing to get larger over successive frames);
detection of a nearby object in the environment of the user, for example by use of proximity detectors on the head mounted display, or analysis of an image from a camera monitoring the user;
detection of a sound in the environment of the user that exceeds a threshold volume, and/or has a predetermined property, such as a harmonic/formant structure indicative of a person’s voice; and a termination signal from a broadcaster that had initiated the second mode.
In addition to these externally driven means of terminating the second mode, in embodiments of the invention the user themselves may terminate the second mode, for example by pressing a predetermined button on a controller, or performing a predetermined motion or sequence of motions, such as shaking their head as if to say ‘no’.
Similarly, the system may have one or more termination criteria. One such criteria may be that the buffer used to hold captured video frames for processing has reached capacity. Another may be that the system has been in the second mode for a predetermined period of time. It will be appreciated that the predetermined period of time may be shorter than the time taken to reach buffer capacity.
If the purpose of the time limit is to preserve the illusion of being able to look around within the time dilated view, then the predetermined period may also vary depending on the comparative size of the portion of the image used in the second mode, since the user is more likely to move outside the available area of the original image in a short time if the portion of the image is comparatively large. Hence the predetermined period of time for a 75%x75% portion may be shorter than for a 40%x40% portion.
It will be appreciated that the user will not necessarily be able to hear or respond to current events during use of the head mounted device in the second mode. Accordingly, the head mounted display may comprise an externally detectable indicator that is operable to be activated for the duration of the second mode. This tells anyone intending to interact with the user that the user is in the second mode and may not be responsive to normal audio or visible cues.
Referring now to Figure 3, a method of video processing for an augmented reality system according to an embodiment of the present invention comprises:
In a first step s310, capturing video images of a real-world scene at a first frame rate;
In a second step s320, displaying in a first mode to a user video images originating from the video camera at said first frame rate and at a first resolution;
In a third step 330, detecting an orientation of a display; and
In a fourth step 340, processing video images captured by the video camera, and wherein in a second mode, the processing step comprises the substeps of:
in a first sub-step 342, processing a captured video image to generate a processed video image, the processed video image corresponding to a portion of the full captured video image that is less than the full captured video image; and in a second sub-step 344, displaying successive processed video images corresponding to respective captured video images at a second, lower frame rate than the first frame rate, thereby creating a slow-motion effect; and wherein the position of the portion of the captured video image being processed to generate the current processed video image for display is responsive to the detected orientation of the display.
It will be apparent to a person skilled in the art that variations in the above method corresponding to operation of the various embodiments of the apparatus as described and claimed herein are considered within the scope of the present invention, including but not limited to:
- in the second mode, performing intra-image interpolation to re-scale a portion of a captured video image to the first resolution to generate a processed video image;
in the second mode, capturing video images at a third, higher frame rate than the first frame rate;
- in the second mode, performing inter-image interpolation to generate one or more processed video images for presentation in between processed video images corresponding to respective captured video images;
- presenting a transition phase in which the portion of the full captured video image changes over successive presented images from a first size portion to a second, smaller size portion;
- presenting a transition phase in which the position of the portion selected from the full captured video corresponds to a predetermined feature of one or more selected from the list consisting of the captured video image being processed; and a computer generated augmentation generated for superposition over the presented processed image;
ending the second mode of operation in the event of one or more selected from the list consisting of: a predetermined time elapsing; detection of a predetermined motion or sequence of motions; activation of a predetermined control by a user; detection of a predetermined event within a currently captured video image not yet presented for display; detection of a nearby object in the environment; and a memory reaching a predetermined occupancy;
- activating an externally detectable indicator for the duration of the second mode;
- generating a sound effect specifically associated with a transition from the first mode to the second mode;
- in the first mode, capturing external sounds with a the microphone and presenting them to the user; and in the second mode, processing the sounds to slow them down by an extent corresponding to the change in rate between the first video capture rate and the second video capture rate; and
- in the first mode, processing a captured video image to generate a processed video image, the processed video image corresponding to a portion of the full captured video image that is less than the full captured video image.
It will be appreciated that the above methods may be carried out on conventional hardware suitably adapted as applicable by software instruction or by the inclusion or substitution of dedicated hardware.
Thus the required adaptation to existing parts of a conventional equivalent device may be implemented in the form of a computer program product comprising processor implementable instructions stored on a non-transitory machine-readable medium such as a floppy disk, optical disk, hard disk, PROM, RAM, flash memory or any combination of these or other storage media, or realised in hardware as an ASIC (application specific integrated circuit) or an FPGA (field programmable gate array) or other configurable circuit suitable to use in adapting the conventional equivalent device. Separately, such a computer program may be transmitted via data signals on a network such as an Ethernet, a wireless network, the Internet, or any combination of these or other networks.

Claims (15)

1. An augmented reality system, comprising a video input adapted to receive video images of a real-world scene captured by a video camera at a first frame rate;
a display operable in a first mode to display to a user video images originating from the video camera at a first resolution;
a sensor operable to detect an orientation of the display; and an image processor operable to process video images captured by the video camera;
and wherein the augmented reality system is operable in a second mode, in which the image processor is operable to process a captured video image to generate a processed video image, the processed video image corresponding to a portion of the full captured video image that is less than the full captured video image;
the image processor is operable to present to the display successive processed video images corresponding to respective captured video images at a second, lower frame rate than the first frame rate, thereby creating a slow-motion effect; and wherein the position of the portion of the captured video image being processed to generate the current processed video image for display is responsive to the detected orientation of the display.
2. The augmented reality system of claim 1, in which in the second mode, the image processor performs intra-image interpolation to re-scale a portion of a captured video image to the first resolution to generate a processed video image.
3. The augmented reality system of claim 1 or claim 2, in which in the second mode, the video camera is operable to capture video images at a third, higher frame rate than the first frame rate.
4. The augmented reality system of any one of the preceding claims, in which in the second mode, the image processor is operable to perform inter-image interpolation to generate one or more processed video images for presentation in between processed video images corresponding to respective captured video images.
5. The augmented reality system of any proceeding claim, in which the image processor presents a transition phase in which the portion of the full captured video image changes over successive presented images from a first size portion to a second, smaller size portion.
6. The augmented reality system of any proceeding claim, in which the image processor presents a transition phase in which the position of the portion selected from the full captured video corresponds to a predetermined feature of one or more selected from the list consisting of:
i. the captured video image being processed; and ii. a computer generated augmentation generated for superposition over the presented processed image.
7. The augmented reality system of any preceding claim, in which the image processor is operable to end the second mode of operation in the event of one or more selected from the list consisting of:
i. a predetermined time elapsing;
ii. detection of a predetermined motion or sequence of motions;
iii. activation of a predetermined control by a user;
iv. detection of a predetermined event within a currently captured video image not yet presented for display;
v. detection of a nearby object in the environment; and vi. a memory reaching a predetermined occupancy.
8. The augmented reality system of any preceding claim, comprising:
an externally detectable indicator that is operable to be activated for the duration of the second mode.
9. The augmented reality system of any preceding claim, comprising an audio output; and in which the augmented reality system is operable to generate a sound effect specifically associated with a transition from the first mode to the second mode.
10. The augmented reality system of any preceding claim, comprising a microphone; and an audio output; and in which in the first mode, external sounds are captured by the microphone and presented to the user;
and in the second mode, the sounds are processed to slow them down by an extent corresponding to the change in rate between the first video capture rate and the second video capture rate.
11. The augmented reality system of any proceeding claim, in which the image processor is operable in the first mode to process a captured video image to generate a processed video image, the processed video image corresponding to a portion of the full captured video image that is less than the full captured video image.
12. A method of video processing for an augmented reality system, comprising the steps of receiving captured video images of a real-world scene at a first frame rate; displaying in a first mode to a user video images originating from the video camera at said first frame rate and at a first resolution;
detecting an orientation of a display; and processing video images captured by the video camera; and wherein in a second mode, the processing step comprises processing a captured video image to generate a processed video image, the processed video image corresponding to a portion of the full captured video image that is less than the full captured video image;
the processing step comprises displaying successive processed video images corresponding to respective captured video images at a second, lower frame rate than the first frame rate, thereby creating a slow-motion effect; and wherein the position of the portion of the captured video image being processed to generate the current processed video image for display is responsive to the detected orientation of the display.
13. The method of video processing of claim 12, in which in the second mode, the video camera is operable to capture video images at a third, higher frame rate than the first frame rate.
14. The method of video processing of any one of claims 12 or 13, in which in the second mode, the image processor is operable to perform inter-image interpolation to generate one or more processed video images for presentation in between processed video images corresponding to respective captured video images.
15. A computer readable medium having computer executable instructions adapted to cause a computer system to perform the method of any one of claims 12-14.
Intellectual
Property
Office
Application No: GB1611911.7 Examiner: Mr Gareth James
GB1611911.7A2016-07-082016-07-08Augmented reality system and methodWithdrawnGB2552150A (en)

Priority Applications (2)

Application NumberPriority DateFiling DateTitle
GB1611911.7AGB2552150A (en)2016-07-082016-07-08Augmented reality system and method
PCT/GB2017/051180WO2018007779A1 (en)2016-07-082017-04-27Augmented reality system and method

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
GB1611911.7AGB2552150A (en)2016-07-082016-07-08Augmented reality system and method

Publications (2)

Publication NumberPublication Date
GB201611911D0 GB201611911D0 (en)2016-08-24
GB2552150Atrue GB2552150A (en)2018-01-17

Family

ID=56891009

Family Applications (1)

Application NumberTitlePriority DateFiling Date
GB1611911.7AWithdrawnGB2552150A (en)2016-07-082016-07-08Augmented reality system and method

Country Status (2)

CountryLink
GB (1)GB2552150A (en)
WO (1)WO2018007779A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
WO2022245690A1 (en)*2021-05-192022-11-24Snap Inc.Extended field-of-view capture of augmented reality experiences

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN109558243A (en)*2018-11-302019-04-02Oppo广东移动通信有限公司Virtual data processing method and device, storage medium and terminal
CN112288877B (en)*2020-10-282024-12-13北京字节跳动网络技术有限公司 Video playback method, device, electronic device and storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20080276178A1 (en)*2007-05-042008-11-06Apple Inc.Adjusting media display in a personal display system based on perspective
US9143693B1 (en)*2014-06-102015-09-22Google Inc.Systems and methods for push-button slow motion
US9360671B1 (en)*2014-06-092016-06-07Google Inc.Systems and methods for image zoom

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
GB9911737D0 (en)*1999-05-211999-07-21Philips Electronics NvAudio signal time scale modification
EP2502410B1 (en)*2009-11-192019-05-01eSight CorporationA method for augmenting sight
US9197864B1 (en)*2012-01-062015-11-24Google Inc.Zoom and image capture based on features of interest

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20080276178A1 (en)*2007-05-042008-11-06Apple Inc.Adjusting media display in a personal display system based on perspective
US9360671B1 (en)*2014-06-092016-06-07Google Inc.Systems and methods for image zoom
US9143693B1 (en)*2014-06-102015-09-22Google Inc.Systems and methods for push-button slow motion

Cited By (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
WO2022245690A1 (en)*2021-05-192022-11-24Snap Inc.Extended field-of-view capture of augmented reality experiences
US11982808B2 (en)2021-05-192024-05-14Snap Inc.Extended field-of-view capture of augmented reality experiences

Also Published As

Publication numberPublication date
WO2018007779A1 (en)2018-01-11
GB201611911D0 (en)2016-08-24

Similar Documents

PublicationPublication DateTitle
US11079999B2 (en)Display screen front panel of HMD for viewing by users viewing the HMD player
US10497175B2 (en)Augmented reality virtual monitor
JP6845111B2 (en) Information processing device and image display method
US10543430B2 (en)Expanded field of view re-rendering for VR spectating
US9992400B2 (en)Real-time changes to a spherical field of view
JP5967343B2 (en) Display system and method for optimizing display based on active tracking
US9579574B2 (en)Image capture method and apparatus
JP2020518321A (en) Second screen virtual window to VR environment
EP4270155B1 (en)Virtual content
JP2017518663A (en) 3D viewing
GB2501929A (en)Augmented reality game apparatus
US20120092327A1 (en)Overlaying graphical assets onto viewing plane of 3d glasses per metadata accompanying 3d image
JP6616023B2 (en) Audio output device, head mounted display, audio output method and program
US11590415B2 (en)Head mounted display and method
JP2019536339A (en) Method and apparatus for synchronizing video content
US12340015B2 (en)Information processing system, information processing method, and program
JPWO2016199731A1 (en) Head mounted display, display control method and program
CN106792096A (en)A kind of augmented reality method and its system based on barrage
US11187895B2 (en)Content generation apparatus and method
CN115623149A (en)Video recording and playback system and method
WO2018007779A1 (en)Augmented reality system and method
WO2021161894A1 (en)Information processing system, information processing method, and program
JP2016509486A (en) Method and system for generating and encoding video game screen images for transmission over a network
JP2024031113A (en)Information processing device and image generation method
US10834382B2 (en)Information processing apparatus, information processing method, and program

Legal Events

DateCodeTitleDescription
WAPApplication withdrawn, taken to be withdrawn or refused ** after publication under section 16(1)

[8]ページ先頭

©2009-2025 Movatter.jp