Movatterモバイル変換


[0]ホーム

URL:


US9294754B2 - High dynamic range and depth of field depth camera - Google Patents

High dynamic range and depth of field depth camera
Download PDF

Info

Publication number
US9294754B2
US9294754B2US13/757,301US201313757301AUS9294754B2US 9294754 B2US9294754 B2US 9294754B2US 201313757301 AUS201313757301 AUS 201313757301AUS 9294754 B2US9294754 B2US 9294754B2
Authority
US
United States
Prior art keywords
frequency
optical power
interest
peak optical
depth
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related, expires
Application number
US13/757,301
Other versions
US20130201288A1 (en
Inventor
Bryed Billerbeck
An-chun Tien
Lucas MORALES
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Lumentum Operations LLC
Original Assignee
Lumentum Operations LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Lumentum Operations LLCfiledCriticalLumentum Operations LLC
Priority to US13/757,301priorityCriticalpatent/US9294754B2/en
Assigned to JDS UNIPHASE CORPORATIONreassignmentJDS UNIPHASE CORPORATIONASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS).Assignors: TIEN, AN-CHUN, MORALES, LUCAS, BILLERBECK, BRYED
Publication of US20130201288A1publicationCriticalpatent/US20130201288A1/en
Assigned to LUMENTUM OPERATIONS LLCreassignmentLUMENTUM OPERATIONS LLCASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS).Assignors: JDS UNIPHASE CORPORATION
Assigned to LUMENTUM OPERATIONS LLCreassignmentLUMENTUM OPERATIONS LLCCORRECTIVE ASSIGNMENT TO CORRECT INCORRECT PATENTS 7,868,247 AND 6,476,312 ON PAGE A-A33 PREVIOUSLY RECORDED ON REEL 036420 FRAME 0340. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT.Assignors: JDS UNIPHASE CORPORATION
Assigned to LUMENTUM OPERATIONS LLCreassignmentLUMENTUM OPERATIONS LLCCORRECTIVE ASSIGNMENT TO CORRECT PATENTS 7,868,247 AND 6,476,312 LISTED ON PAGE A-A33 PREVIOUSLY RECORDED ON REEL 036420 FRAME 0340. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT.Assignors: JDS UNIPHASE CORPORATION
Publication of US9294754B2publicationCriticalpatent/US9294754B2/en
Application grantedgrantedCritical
Assigned to DEUTSCHE BANK AG NEW YORK BRANCH, AS COLLATERAL AGENTreassignmentDEUTSCHE BANK AG NEW YORK BRANCH, AS COLLATERAL AGENTPATENT SECURITY AGREEMENTAssignors: LUMENTUM OPERATIONS LLC, OCLARO FIBER OPTICS, INC., OCLARO, INC.
Assigned to LUMENTUM OPERATIONS LLC, OCLARO, INC., OCLARO FIBER OPTICS, INC.reassignmentLUMENTUM OPERATIONS LLCRELEASE BY SECURED PARTY (SEE DOCUMENT FOR DETAILS).Assignors: DEUTSCHE AG NEW YORK BRANCH
Expired - Fee Relatedlegal-statusCriticalCurrent
Adjusted expirationlegal-statusCritical

Links

Images

Classifications

Definitions

Landscapes

Abstract

In order to maximize the dynamic range and depth of field for a depth camera used in a time of flight system, the light source is modulated at a plurality of different frequencies, a plurality of different peak optical powers, a plurality of integration subperiods, a plurality of lens foci, aperture and zoom settings during each camera frame time. The different sets of settings effectively create subrange volumes of interest within a larger aggregate volume of interest, each having their own frequency, peak optical power, lens aperture, lens zoom and lens focus products consistent with the distance, object reflectivity, object motion, field of view, etc. requirements of various ranging applications.

Description

CROSS-REFERENCE TO RELATED APPLICATIONS
The present invention claims priority from U.S. Patent Application Ser. No. 61/594,745 filed Feb. 3, 2012, which is incorporated herein by reference.
TECHNICAL FIELD
The present invention relates to a high dynamic range and depth of field depth camera, and in particular to a depth camera, which has a plurality of frequency, peak optical power and integration period settings within each frame time for use in a time of flight based camera system.
BACKGROUND OF THE INVENTION
In traditional photography, depth of field (DOF) is how much of your image is in focus. Shallow depth of field refers to when only things that are very close to the plane of the subject you focus on are in focus. Objects that are behind or in front of your subject will appear out of focus. Dynamic range refers to the difference between your highlights and shadows in an image. On average, the human eye can see dynamic range on the order of 1,000,000:1. What this means is that we can see details in both very bright and very dark areas of a scene at the same time. A digital sensor has a dynamic range on the order of 1,000:1, which means that if your subject is very bright compared to your background, when you expose the image correctly for your subject, you background will appear to be very dark. Conversely, if you exposed the image for the background, your subject may appear “blown out” or very bright white and overexposed.
As with traditional photography cameras, depth of field, i.e. image sharpness consistency within the camera Field of View (FOV) and range of interest, and image dynamic range, i.e. image contrast range within the camera FOV and range of interest, requirements can be a significant challenge to a depth camera. In the case of a depth camera, objects close to the camera's light source and or of high reflectivity can cause over saturation of sensor pixels while objects further from the camera's light source and or of low reflectivity can be difficult for the sensor/camera to detect at all.
A “Time-of-Flight” based depth camera comprises a depth image sensor, which is typically made using a standard CMOS fabrication process, and an IR light source for measuring distance, which is proportional to the length of time the IR light takes to travel from and return to the camera. A depth camera system generates depth images and transmits them to the host processor over a suitable, e.g. USB, interface.
The camera hardware includes a light source module, an IR light detecting, e.g. CMOS, image sensor, and an ambient light-color sensing, e.g. CMOS, image sensor. A 3D imager produces phase measurements that are processed either on sensor or in a remote coprocessor to produce actual range data. Such a camera can be used in “Z-only” mode for applications, which require the use of range data only. The camera could also be used in “RGB+Z”, i.e. full 3D depth and 2 dimensional colors, modes for applications which utilize both traditional color as well as depth images. Depth and color processing can be done in the camera or with a pass-through mode in which unprocessed data can be passed to the host for processing.
In a depth only camera, the sensor and light source will be synchronized in time. In RGB and depth cameras, the light source and the two sensors will be synchronized in time, such that both sensors start their frames cycles with a known and locked timing relationship, e.g. at the same time, with each other and the light source. Also, the frame start time of each sensor can be adjusted with respect to the data stream to the host to provide a system-level synchronization capability. Data from each sensor and audio can be transmitted to host devices on separate streams over various interfaces, such as a USB2.0 isochronous link, which may include a tagging capability to insert timestamps into each frame of each sensor. Similarly, the data streams could be integrated before transmission and de-integrated by the host.
The camera enables developers to create many new kinds of applications, e.g. gesture control of host devices, interactive games, etc., requiring both depth and color video.
In a typical Time of Flight (ToF) based camera, the camera is designed to synchronously modulate a light source at a fixed Peak Optical Power (POP) level with a sensor's active integration period, i.e. frame time. In some implementations, camera operating frequencies are varied in fairly narrow ranges, e.g. +/−20 MHz, to be able to detect distance aliasing artifacts caused by frequency wrap around or reflections of objects beyond the camera's working range created by frequency range multiples. In ToF based cameras, frequency is proportional to distance, e.g. an object at 6 m, which is 1 m outside a 5 m range of interest, can be falsely detected as an object of low reflectivity at 3 m, a distance within the range of interest.
An object of the present invention is to overcome the shortcomings of the prior art by providing a depth camera for a ToF system that divides the overall range of interest and Field of View (FoV) into Volume of Interest (VOI) sub-ranges with different frequency, peak optical power and integration period pairs or triplets for each VOI sub-range.
SUMMARY OF THE INVENTION
Accordingly, the present invention relates to a depth camera for a time of flight device for presenting three dimensional data to a host comprising:
a light source for launching a beam of light at a range of interest for a field of view, the light source having an adjustable frequency and an adjustable peak optical power;
a detector array for receiving and detecting portions of the beam of light reflected off of objects within the range of interest and field of view; and
a controller for adjusting the light source frequency and peak optical power within a single frame time period in accordance with a plurality of frequency and peak optical power pairs to obtain a plurality of three dimensional data measurements for each frame,
whereby the range of interest and field of view is divided into a plurality of volumes of interest, each volume of interest having a different, frequency, peak optical power, and minimum and maximum distance from the light source, to increase dynamic range and depth of field within each volume of interest.
Another aspect of the present invention relates to a method of operating a depth camera comprising:
a) launching light from a light source at a range of interest over a field of view; and
b) receiving and detecting portions of the beam of light reflected off of various objects within the range of interest and the field of view;
wherein step a) includes:
launching a first beam of light at a first frequency and a first peak optical power for a first integration period within a single frame time period to increase depth of field and dynamic range within a first volume of interest to generate a first three dimensional data measurement;
launching a second beam of light at a second frequency greater than the first frequency and a second peak optical power less than the first peak optical power for a second integration period within the same single frame time period to increase depth of field and dynamic range within a second volume of interest to generate a second three dimensional data measurement; and
launching a third beam of light at a third frequency greater than the second frequency and a third peak optical power less than the second peak optical power for a third integration period within the same single frame time period to increase depth of field and dynamic range within a third volume of interest to generate a third three dimensional data measurement.
Yet another aspect of the present invention includes a time of flight based depth camera for presenting three dimensional data to a host device comprising:
a light source for launching a beam of light at a range of interest for a field of view, the light source having variable integration time periods and peak optical power;
a detector array for receiving and detecting portions of the beam of light reflected off of objects within the range of interest and field of view; and
a controller for adjusting the light source integration time period and peak optical power within a single frame time period in accordance with a plurality of integration time period and peak optical power pairs to obtain a plurality of three dimensional data measurements,
whereby the range of interest and field of view is divided into a plurality of volumes of interest, each volume of interest having a different, integration time period, peak optical power, and minimum and maximum distance from the light source, to increase dynamic range and depth of field within each volume of interest.
BRIEF DESCRIPTION OF THE DRAWINGS
The invention will be described in greater detail with reference to the accompanying drawings which represent preferred embodiments thereof, wherein:
FIG. 1 is an isometric view of a depth camera in accordance with the present invention;
FIG. 2 is a schematic representation of the depth camera ofFIG. 1 with predefined volume of interest sub-ranges in accordance with the present invention;
FIG. 3 is a plot of volume of interest sub-range triplet information in accordance with the present invention; and
FIG. 4 is a schematic representation of and an ASIC of the depth camera of the present invention.
DETAILED DESCRIPTION
With reference toFIG. 1, adepth camera1, in accordance with the present invention, for a Time of Flight (ToF) system, e.g. for a gesture recognition device, includes alight source2, e.g. an LED, Laser or other light emitting device with associated drive and control circuitry, for launching a beam of light into an overall Range of Interest (RoI) and Field of View (FoV), a receiving optic with an IR filter, and a depthsensor detector array3 for receiving and detecting portions of the beam of light reflected off of various moving objects within the RoI and FoV. Thedepth camera1 also includes apower port4 for connection to a remote or host power supply, and adata port6 for connection to a host processor. In the illustrated application, a color (RGB)camera7 is also provided, along withvariable speed fans8 for cooling the system.
When considering dynamic range and depth of field, higher frequencies and lower powers work well for short range applications (0.5 mm to 3 m), low frequencies and high powers work well for long range applications (3 m to 5 m), and something in between is better for intermediate range applications (2 m to 4 m). However, thecamera1, of the present invention, provides a working range of between 0.5 mm and 5 m or more.
With reference toFIG. 2, the RoI and FoV, for thecamera1 according to the present invention, is divided into a plurality of sub-range volumes of interest (VoI),16,17 and18, each with their own operating frequency and synchronized with a different light source peak optical power (POP) during a integration subperiod within the camera's frame time to create frequency, POP and integration subperiod pairs or triplets more finely tuned for the plurality of sub-ranges within the camera application's range of interest. The summing or aggregation of these sub-range intra-frame time frequency, POP and integration subperiod pairs or triplets creates a higher dynamic range and sharper depth of field image within the given overall RoI. Three VoI's are illustrated; however, two, four or more VoI's are possible and within the scope of this invention.
An aggregation process is used to stitch together intra frame depth image fragments generated by thedepth camera1 using optical power, frequency and integration period pairs or triplets into a single or multiple frame composite image(s). Such aggregation could be done by, but not limited to, the simple or weighted averaging of individual pixel values generated by intra frame or multiple frame pairs, triplets, quadruplets, etc. Similarly mean values, mode values, etc. could be used.
The determination of appropriate frequency, power values and ideally integration subperiods can be made via estimation, calculation or empirically for static or dynamic scenes, statically, e.g. during an initial set up, or dynamically, e.g. during use. An example of a simple static estimation would be to break the FOV Volume Of Interest (VOI) into a plurality of sections, e.g. two to five or more VoI sub-ranges, in the depth direction and assume that the power required for successive sections would increase moving away from the camera. The appropriate number of VoI sub-ranges is dependent on the application's overall range and depth of field requirements, and typical object reflectivity's in the range, i.e. longer distances and greater target object reflectivity ranges could mean more sections. The VoI sub-ranges can have equal ranges, and volumes, e.g. 1.5 m to 3 m, preferably 2 m deep, or they can have different ranges and volumes, e.g. thefirst volume16 is smaller than thesecond volume17, which is smaller than thethird volume18. Typically these sub-ranges or volumes will grow with distance from thecamera1.
The VoI sub-ranges can be discreet volumes, e.g. VoI16 has an R1minto R1maxof from 0.5 m to 1.5 m,VoI17 has a R2minto R2maxof from 1.5 m to 3 m, andVoI18 has a R3minto R3maxof from and 3 m to 5 m. Alternatively, the VoI sub-ranges can be a series of overlapping volumes, e.g. VoI16 has an R1minto R1maxof from 0.5 m to 2 m,VoI17 has a R2minto R2maxof from 1 m to 3 m, andVoI18 has a R3minto R3maxof from and 2 m to 5 m.
Next, with reference toFIG. 3, a frequency fnis selected, e.g. a suitable unambiguous, frequency or the highest (non-aliasing) prime frequency thecamera1 is capable of, within a selected low range of frequencies (10 to 40 MHz) for the farthest VoI sub-range, e.g. VoI18.
For example: at 5 m R3max, the non-aliasing max frequency is 20 MHz, but 19 MHz, which is the highest prime frequency close to but not higher than 20 MHz, may be the better choice for thefarthest VoI subrange18, so that erroneous readings do not occur due to the closer subranges in which the frequency might be set to 40 MHz or 60 MHz, thereby creating interfering harmonics. Accordingly, 37 MHz or 59 MHz would likely be better choices for themiddle subrange VoI17, and 97 MHz, etc. for thenearest subrange VoI16. Ideally, the least common multiple of the three frequencies f1, f2, fn, selected are outside the working range of the camera.
The POP Pnfor the farthest VoI sub-range, e.g. VoI18, is then determined based on the frequency fnand the POP necessary to resolve the required minimum object size at its lowest targeted reflectivity. As the range VoI moves closer to the camera, the frequency and POP values are modified to optimize performance within the closer VoI sub-ranges, e.g. VoI sub-ranges16 and17, e.g. lower power, higher frequency in successive ranges getting closer to camera. The frequency values between the VoI's16 to18 are at much wider ranges, e.g. greater than +/−30 MHz, 50 MHz and up at least 100 MHz, than prior art systems.
For example: for thefarthest VoI sub-range18 the frequency fnselected might be 10 to 40 MHz, preferably 20 to 25 MHz to avoid distance aliasing in the range, and the POP Pnmight be determined to be 5 W or more, preferably 1 W or more, most preferably greater than 500 mW, but ideally as low as possible to reduce power requirements. The middle VoI sub-range17 can have a frequency f2higher than thefarthest VoI sub-range18, e.g. 30 to 70 MHz, preferably 40 to 50 MHz to avoid distance aliasing within this midrange, with a POP P2less than thefarthest VoI sub-range18, e.g. less than half of the farthest sub-range or 250 mW to 4 W, preferably 250 mW to 1 W, and most preferably 250 mW to 500 mW, but ideally as low as possible. Theclosest VoI sub-range16 can have a frequency f1higher than the other two or more VoI sub-ranges, e.g. 60 MHz or more, preferably 75 MHz or more, but typically as high as possible below the distance aliasing frequency, with a POP P1less than the other ranges, e.g. 400 mW or less, preferably less than 250 mW, and most preferably 50 mW or less, but ideally as low as possible.
In a simple embodiment, the intra-frame integration period I is divided equally amongst the VoI sub-ranges; however, the intra-frame integration periods I1, I2, . . . Incan be divided unevenly amongst the frequency/POP pairs to more optimally allocate power within or distribute power to the various VoI sub-ranges. Intra-frame integration time, frequency and or power might be reduced to avoid saturation of more reflective objects with in the range's VOI. For example: an integration period between say 5% and 40%, preferably between 25% and 40%, of the total frame period would be suitable time periods, depending on system requirement. Typically, the larger VoI sub-ranges and those with longer ranges require more time, e.g. a larger percentage of the integration period.
If it is determined that the frequency, POP and integration subperiod pairs or triplets require less than a full frame period to achieve desired image quality, the light source can be turned off to conserve power and reduce the generation of unwanted heat in the system, as illustrated inFIG. 3.
The aggregation of object depth values across these more optimized sub-range intra-frame time frequency, POP and integration time triplets R1, R2, . . . Rnenables the creation of a higher dynamic range, sharper depth of field and more accurate depth image of a given range of interest. As the VoI sub-ranges move closer to thecamera1, the pair or triplet values can be modified to optimize performance within the second sub-range's VoI, e.g. lower power, higher frequency in successive ranges getting closer to camera.
In a more complex case, e.g. a dynamic scene where objects of varying brightness are moving within the scene and various VoI sub-ranges, historical frame data can be used to make predictive modifications dynamically during camera operation. The object motion is tracked in three dimensions as well as its brightness. As an example, if in a few frames of data a bright or highly reflective object is found within a VoI sub-range's moving closer to thecamera1, the triplet R of that sub-range VoI or zone could be modified, as necessary. For example: the light source power P could be reduced and/or the frequency f could be increased and/or the integration period I could be reduced, or any similar weighted combinations of these variables could be changed, using the components and processes described in the below explanation of the 3D camera block diagram, to avoid a saturation issue. Similarly, a darker or less reflective object moving away from thecamera1 could trigger modification to a range's VoI triplet to keep the object from disappearing while still in the range.
In yet another embodiment, the intra-frame triplet variations could be used in conjunction with an autofocus/zoom lens9, with or without an equivalent light source zoom capability. In the case of the autofocus/zoom lens9, a sub-range VoI triplet could become a quadruplet. In addition to modifying one or more of intra-frame frequency, integration time, and optical power, the FOV/VOI could be increased or decreased as an application required by the widening or narrowing the FOV/VOI of one of the sub-range VoI's16 to18. As an example, the sub-range VoI's closer to thecamera1 typically require larger FOVs than sub-range VoI's further from thecamera1. Narrowing or widening the camera's FOV has nearly the same effect as increasing or decreasing POP for a sub-range VoI, the amount of which depends on whether or not the light source is synchronously zoomed with thelens9 to cover the same FOV. As objects get closer to thecamera1, they get larger, taller and wider and would benefit from a larger FOV. So, rather than simply reducing the POP, frequency or integration period to avoid saturation, the lens and the light source can be zoomed out, which would effectively reduce the per pixel POP and similarly avoid potential object saturation issues. Alternatively, as objects move further from thecamera1, they get smaller, shorter and narrower which leaves room to reduce the cameras FOV. So, rather than simply increasing the POP, frequency or integration period to avoid object disappearance, thelens9 andlight source2 can be zoomed in to effectively increase per pixel POP. This could significantly reduce the maximum POP and electrical power requirement of thecamera1 throughout its total range.
In yet another embodiment, the intra-frame triplet, or intra-frame triplet pluszoom lens9 intra-frame quadruplet variations. could be used in conjunction withlens9 with variable aperture capability, with or without an equivalent lens or light source zoom capability. In the case of thevariable aperture lens9, a sub-range VoI triplet could become a quadruplet or quintuplet if used with the lens' zoom capability. In addition to modifying one or more of intra-frame frequency, integration subperiod, and optical power, the FOV/VOI could be increased or decreased as an application required by the widening or narrowing the lens aperture setting of one of the sub-range VoI's16 to18. As an example, a camera sub-range VoI may require more or less light than the previous or next camera sub-range VoI. Narrowing or widening the camera's aperture has nearly the same effect as increasing or decreasing POP for a sub-range VoI, the amount of which depends on whether or not thelight source2 is synchronously zoomed and or focused with thelens9 to cover the same FOV.
Thelight source2 comprises a laser required to actively illuminate the specified optical field of view within the camera working range with modulated light. Thelight source2 provides the specified wavelength of monochromatic light for the active illumination. A typicallight source2 comprises a laser, highspeed driver circuitry10 and a diffuser for uniform light distribution within the FOV. Thelaser driver circuit10 is controlled with signals coming from acoprocessor21. Thelight source2 is modulated during the integration time of thesensor3 and is tightly coupled with sensor operation. Care needs to be taken to control the timing and waveform of signals going to thelight source2 and within thelight source2 to produce proper illumination over the operating temperature.
Light source frequency is a critical system performance variable and must be selected carefully based on camera operating range with higher frequencies being better for near range applications and lower frequencies being better for long range applications. Along with multi-frequency implementation, aliasing artifacts can be reduced by proper frequency selection for an application.
Depending on sensor architecture and output, thecoprocessor21 is provided that can translate phase data to depth data, performs depth calibration, depth data corrections, RGB color processing, compression sensor control, RGB & Z data synchronization, tagging and registration. For low functioning sensors (RGB or Z), some or all of these above processes could be handled by thecamera coprocessor21 or raw data could be passed through to the host for processing. For higher functioning sensors, fewer of these processes will need to be handled by thecoprocessor21 within thecamera1. The choice ofcoprocessor21 is very much dependent on the target host processing capability. Less intelligent host devices require more in camera processing while more intelligent host devices can be less dependent on in camera processing.
With reference toFIG. 4, in the present invention, theabove coprocessor21 and the following 3D imaging related functions could be implemented in gate level logic and integrated into a single application specificintegrated circuit ASIC21 or implemented in host side software or hardware or any combination in between. In a conventional time of flight based 3D input device, phase data coming from a depth sensor, raw RGB data coming from a traditional color sensor, and audio coming from a microphone or microphone array must be further processed prior to presentation to various host devices, such as PCs, TVs, mobile devices, etc. for display.
Conventional discrete component implementations require analog to digital converters (ADC) for each of the sensors and microphones, an image correction logic, compression logic and compression memory. Most of these resources can be shared when consolidated into theASIC21 resulting in a significant size and cost savings. Similarly, pure logic devices are typically fabricated on smaller fabrication process geometries resulting in additional size and cost reductions when the RGB ISP, RGB & Z corrections and control functions are moved off sensor into theASIC21.
Depending on the RGB and Depth sensor formats, the raw data streams can exceed the bandwidth of the host input port, typically USB 2.0 and must therefore be processed and or compressed prior to transport.
The below described functions of theASIC21 can be performed faster when implemented in hardware logic as opposed to software running on a host's processor. If the host's processor is required to perform the below mentioned ASIC functions as well as application processing the result can be increased application latency, which can be distracting to a user. Similarly, keeping data flow on chip during processing can increase processing speed and further decrease application latency.
An embedded microprocessor/controller22 controls the flow of data and command instructions within theASIC21, and turns on or off specific ASIC logic block level functions or functions within ASIC logic blocks. Themicroprocessor22 also controls data and command instruction flow of sensors, i.e. thedepth sensor3 andRGB sensor7, and thelight source2, in particular from the sensor and lightsource control module23. Themicroprocessor22 also supports device level user interfacing, e.g. buttons, switches,display26, etc., and supports firmware based processing functions and programmability, e.g. exposure control, gain control, light source frequency and duty cycle control,depth sensor3 toRGB7 andlight source2 synchronization, etc.
I2C 27 is a command/instruction bus allowing command and response flow within theASIC21.I2C 27 is the command, control and feedback mechanism; however, other standard communication, control and feedback standards or proprietary methods could be used. Any 3D processor block could be integrated into a single chip or parts in any combination of discrete chips. The system power supply (not shown) could be a standalone power supply, integrated into the 3D processor or received from ahost28 via an integrated power and data bus such as USB, Firewire, etc.
Analog to digital converter (ADC)29 converts analog data to digital data for further on chip and host processing.
One or more Input/Output Port(s)31 supports the input of raw or preprocessed data from theRGB sensor7, thedepth sensor3, anoptional microphone32, thelight source2 or other data; as well as command, clock or other general purpose input to theASIC21 from its peripherals and from theASIC21 to its peripherals. The input/output port31 also provides processed data output to thedisplay26.
Flash Memory33 stores or buffers data from the various input devices, e.g. depth andRGB sensors3 and7, for the various on chip processing modules, e.g.depth processing module34 andcolor processing module36.
RAM37 stores command and control instruction firmware for execution by the microprocessor/controller22.
A Clock/PLL38 provides an internal or externally synchronized clock reference for chip function timing and control.
TheUSB Phy6, which could be PCI, MIPI or any other standard or proprietary Phy provides a physical interface from theASIC21 to thehost device28.
AnAudio Codec41 converts raw or preprocessed audio data frommics32 to an industry standard or proprietary data format for use by thehost28.
TheColor Processing module36 performs white balancing/color correction, color demosaicing, color space conversion, and other raw or preprocessed RGB data stream related processing functions to the data from theRGB sensor7.
TheDepth Processing module34 performs phase to depth data conversion, etc. on raw or preprocessed depth data form thedepth sensor3.
ACorrections module42 performs lens, Gama, dark level compensation, sensor defect, scaling, horizontal/vertical flip/rotation, filtering, flicker, dealiasing, binning etc.
ACompression module43 compresses raw or preprocess RGB, Depth, Audio or other data from the correspondingmodules7,3,32 to industry standard, e.g. JPEG, MJPEG, H.264, Dolby AC3, etc. or proprietary formats for use byhost device28.
AData Bus44 transfers data between the functional blocks,e.g. modules3,7,32,memory33,37 andports31 of theASIC21 under direction of the microprocessor/controller22.
Synchronization, Tagging & Mergingmodule44 tags, synchronizes and multiplexes packetized RGB, Depth and Audio raw or processed data streams from the correspondingsensors7,3 and32 for transfer and presentation to thehost28 via theUSB Phy6.
Calibration andRegistration module46 uses algorithms and internal coefficients created from camera measurements made during the calibration phase of camera testing to correlate camera depth values with actual object distance values. Registration is the process of correlating the x & y locations of pixels between the RGB and Depth sensors.
Themicroprocessor22 first applies settings for one of the predetermined frequency, power and integration period triplet stored in theRAM37 or other suitable non-transitory memory. Preferably, themicroprocessor22 then assesses the image quality, such as depth accuracy, and image sharpness, e.g. by determining at least one of: the number of blurred pixels, the number of saturated pixels, and the number of dark pixels, etc., within the field of view and volume of interest sub-range, and comparing the determined number to a predetermined threshold value, based on algorithms stored in non-volatile memory or hard coded in the processor or other chip. The micro-processor22 then, using other algorithms stored in memory or hard coded in the processor or other chip, calculates appropriate new triplets and triplet sequences for best or improved dynamic range. The micro-processor22 then adjusts the clock timing (frequency),light source2 power level,lens9 zoom and focus settings corresponding to each sequential triplet, as necessary. Image quality based feedback loops may or may not be used to validate or monitorsensor3,lens9 orlight source2 actual settings or performance. The resulting triplets could be sequenced in any order and the system could be implemented with a fixed focus lens with appropriate depth of field for the field of view and volume of interest. The micro-processor22 then merges the data from each triplet into or calculates a composite frame(s) for transfer to thehost device28.
Such processing and control could also be implemented on and performed by thesensor3 or the raw data could be transferred to thehost28 for processing or in any distributed processing device combination. Similarly, such processing could be performed before, after or during phase to depth calculation. i.e. on phase, depth/range or intermediate data; and before, after or during other (calibration, registration, corrections, compression and audio, RGB & Depth data synchronization, tagging and merging, etc.) data processing activities.

Claims (24)

We claim:
1. A depth camera for a time of flight device for presenting three dimensional data to a host comprising:
a light source for launching a beam of light at a range of interest for a field of view, the light source having an adjustable frequency and an adjustable peak optical power;
a detector array for receiving and detecting portions of the beam of light reflected off of objects within the range of interest and field of view; and
a controller for adjusting the light source frequency and peak optical power within a single frame time period in accordance with a plurality of frequency and peak optical power pairs to obtain a plurality of three dimensional data measurements for each frame time period,
whereby the range of interest and field of view is divided into a plurality of volumes of interest, each volume of interest having a different, frequency, peak optical power, and minimum and maximum distance from the light source, to increase dynamic range and depth of field within each volume of interest.
2. The depth camera according toclaim 1, wherein initial light source frequency and peak optical power settings for each volume of interest are predetermined and stored in non-volatile memory accessible by the controller.
3. The depth camera according toclaim 2, wherein the controller adjusts at least one of the light source frequency and the peak optical power based on feedback from the detector array.
4. The depth camera according toclaim 1, wherein the controller also adjusts a light source integration period in accordance with predetermined or calculated integration periods for each frequency and peak optical power pair.
5. The depth camera according toclaim 4, further comprising a lens for focusing the beam of light; wherein the controller is also capable of adjusting a focus of the lens along with each frequency, peak optical power and integration period.
6. The depth camera according toclaim 5, further comprising an adjustable aperture for the lens; wherein the controller is also capable of adjusting the adjustable aperture along with each frequency, peak optical power, integration period and lens focus.
7. The depth camera according toclaim 1, further comprising an adjustable lens for focusing the beam of light; wherein the controller is also capable of adjusting a focus of the lens along with each frequency and peak optical power.
8. The depth camera according toclaim 1, further comprising a lens and an adjustable aperture for the lens; wherein the controller is also capable of adjusting the adjustable aperture along with each frequency and peak optical power pair.
9. The depth camera according toclaim 1, wherein the controller is also capable of aggregating the plurality of three dimensional data measurements.
10. The depth camera according toclaim 9, further comprising an ASIC incorporating the controller and other control and processing functions of the camera to minimize latency between the depth camera and the host.
11. The depth camera according toclaim 1, wherein the controller is also capable of dynamically adjusting any one or more of the frequencies, and the peak optical powers depending on a reflectivity of an object.
12. The depth camera according toclaim 1, wherein a first frequency is between 10 and 40 MHz, and a first peak optical power is greater than 1 W.
13. The depth camera according toclaim 1, wherein a second frequency is between 40 and 50 MHz, and a second peak optical power is between 250 mW and 1 W.
14. The depth camera according toclaim 1, wherein a third frequency is greater than 75 MHz, and a third peak optical power is less than 250 mW.
15. A method of operating a depth camera comprising:
a) launching light from a light source at a range of interest over a field of view; and
b) receiving and detecting portions of the light reflected off of various objects within the range of interest and the field of view;
wherein step a) includes:
launching a first beam of light at a first frequency and a first peak optical power for a first integration period within a single frame time period to increase depth of field and dynamic range within a first volume of interest to generate a first three dimensional data measurement,
launching a second beam of light at a second frequency greater than the first frequency and a second peak optical power less than the first peak optical power for a second integration period within the same single frame time period to increase depth of field and dynamic range within a second volume of interest to generate a second three dimensional data measurement; and
launching a third beam of light at a third frequency greater than the second frequency and a third peak optical power less than the second peak optical power for a third integration period within the same single frame time period to increase depth of field and dynamic range within a third volume of interest to generate a third three dimensional data measurement.
16. The method according toclaim 15, wherein step b) including aggregating the first, second and third three dimensional data measurements.
17. The method according toclaim 15, further comprising adjusting a focus of a lens associated with the light source for each of the first, second and third volumes of interest.
18. The method according toclaim 15, further comprising adjusting an aperture of a lens associated with the light source for each of the first, second and third volumes of interest.
19. The method according toclaim 15, further comprising dynamically adjusting any one or more of the first, second and third frequencies, and the first, second and third peak optical powers depending on a reflectivity of one of the objects.
20. The method according toclaim 15, wherein the first, second and third integration periods are different.
21. The method according toclaim 15, wherein the first frequency is between 20 and 25 MHz, and the first peak optical power is greater than 500 mW.
22. The method according toclaim 15, wherein the second frequency is between 40 and 50 MHz, and the second peak optical power is between 250 and 500 mW.
23. The method according toclaim 15, wherein the third frequency is greater than 60 MHz, and the third peak optical power is less than 250 mW.
24. A time of flight based depth camera for presenting three dimensional data to a host device comprising:
a light source for launching a beam of light at a range of interest for a field of view, the light source having variable integration time periods and peak optical power;
a detector array for receiving and detecting portions of the beam of light reflected off of objects within the range of interest and field of view; and
a controller for adjusting the light source integration time period and peak optical power within a single frame time period in accordance with a plurality of integration time period and peak optical power pairs to obtain a plurality of three dimensional data measurements,
whereby the range of interest and field of view is divided into a plurality of volumes of interest, each volume of interest having a different, integration time period, peak optical power, and minimum and maximum distance from the light source, to increase dynamic range and depth of field within each volume of interest.
US13/757,3012012-02-032013-02-01High dynamic range and depth of field depth cameraExpired - Fee RelatedUS9294754B2 (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
US13/757,301US9294754B2 (en)2012-02-032013-02-01High dynamic range and depth of field depth camera

Applications Claiming Priority (2)

Application NumberPriority DateFiling DateTitle
US201261594745P2012-02-032012-02-03
US13/757,301US9294754B2 (en)2012-02-032013-02-01High dynamic range and depth of field depth camera

Publications (2)

Publication NumberPublication Date
US20130201288A1 US20130201288A1 (en)2013-08-08
US9294754B2true US9294754B2 (en)2016-03-22

Family

ID=48902539

Family Applications (1)

Application NumberTitlePriority DateFiling Date
US13/757,301Expired - Fee RelatedUS9294754B2 (en)2012-02-032013-02-01High dynamic range and depth of field depth camera

Country Status (1)

CountryLink
US (1)US9294754B2 (en)

Cited By (17)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20150116459A1 (en)*2013-10-252015-04-30Lips IncorporationSensing device and signal processing method thereof
US20160050346A1 (en)*2014-06-132016-02-18Lips CorporationIntegrated depth camera
US20160373722A1 (en)*2015-06-192016-12-22Amazon Technologies, Inc.Steganographic depth images
US9786080B1 (en)*2015-07-022017-10-10Yesvideo, Inc.2D/3D image scanning and compositing
US10212306B1 (en)2016-03-232019-02-19Amazon Technologies, Inc.Steganographic camera communication
US10241244B2 (en)2016-07-292019-03-26Lumentum Operations LlcThin film total internal reflection diffraction grating for single polarization or dual polarization
CN110312117A (en)*2019-06-122019-10-08北京达佳互联信息技术有限公司Method for refreshing data and device
WO2019199645A1 (en)2018-04-092019-10-17Sense Photonics, Inc.Automatic gain control for lidar for autonomous vehicles
WO2020058264A1 (en)*2018-09-192020-03-26Sony Semiconductor Solutions CorporationTime of flight apparatus and method
US10873738B2 (en)2016-03-032020-12-224D Intellectual Properties, LlcMulti-frame range gating for lighting-invariant depth maps for in-motion applications and attenuating environments
US10996320B2 (en)*2017-07-112021-05-04Sony Semiconductor Solutions CorporationElectronic device and control method of electronic device
US11001979B2 (en)2018-08-132021-05-11Vergence Automation, Inc.Methods and apparatus for ultrawide entrance angle reflective articles for use with autonomous vehicle machine vision systems
US11041957B2 (en)2018-06-252021-06-22Toyota Motor Engineering & Manufacturing North America, Inc.Systems and methods for mitigating effects of high-reflectivity objects in LiDAR data
US11514594B2 (en)2019-10-302022-11-29Vergence Automation, Inc.Composite imaging systems using a focal plane array with in-pixel analog storage elements
US11606517B1 (en)2021-06-072023-03-14Waymo LlcEnhanced depth of focus cameras using variable apertures and pixel binning
US11762133B1 (en)2018-09-132023-09-19Vergence Automation, Inc.Retroreflective materials and articles incorporating near-ideal total internal retroreflective elements
US12230649B2 (en)2020-05-192025-02-18Vergence Automation, Inc.AI system on chip (SOC) for robotics vision applications

Families Citing this family (58)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US9779546B2 (en)2012-05-042017-10-03Intermec Ip Corp.Volume dimensioning systems and methods
US10007858B2 (en)2012-05-152018-06-26Honeywell International Inc.Terminals and methods for dimensioning objects
TWI471630B (en)*2012-06-012015-02-01Hon Hai Prec Ind Co LtdAuto-focus system and method of a digital camera
US10321127B2 (en)2012-08-202019-06-11Intermec Ip Corp.Volume dimensioning system calibration systems and methods
US9939259B2 (en)2012-10-042018-04-10Hand Held Products, Inc.Measuring object dimensions using mobile computer
US20140104413A1 (en)2012-10-162014-04-17Hand Held Products, Inc.Integrated dimensioning and weighing system
US9080856B2 (en)2013-03-132015-07-14Intermec Ip Corp.Systems and methods for enhancing dimensioning, for example volume dimensioning
US10228452B2 (en)2013-06-072019-03-12Hand Held Products, Inc.Method of error correction for 3D imaging device
US10203399B2 (en)2013-11-122019-02-12Big Sky Financial CorporationMethods and apparatus for array based LiDAR systems with reduced interference
US9360554B2 (en)2014-04-112016-06-07Facet Technology Corp.Methods and apparatus for object detection and identification in a multiple detector lidar array
US9823059B2 (en)2014-08-062017-11-21Hand Held Products, Inc.Dimensioning system with guided alignment
US10810715B2 (en)2014-10-102020-10-20Hand Held Products, IncSystem and method for picking validation
US9779276B2 (en)2014-10-102017-10-03Hand Held Products, Inc.Depth sensor based auto-focus system for an indicia scanner
US10775165B2 (en)2014-10-102020-09-15Hand Held Products, Inc.Methods for improving the accuracy of dimensioning-system measurements
US9897434B2 (en)2014-10-212018-02-20Hand Held Products, Inc.Handheld dimensioning system with measurement-conformance feedback
US9752864B2 (en)2014-10-212017-09-05Hand Held Products, Inc.Handheld dimensioning system with feedback
US9557166B2 (en)2014-10-212017-01-31Hand Held Products, Inc.Dimensioning system with multipath interference mitigation
US9762793B2 (en)2014-10-212017-09-12Hand Held Products, Inc.System and method for dimensioning
US9330464B1 (en)2014-12-122016-05-03Microsoft Technology Licensing, LlcDepth camera feedback
US9581696B2 (en)*2014-12-222017-02-28Google Inc.Image sensor and light source driver integrated in a same semiconductor package
US9635231B2 (en)2014-12-222017-04-25Google Inc.Time-of-flight camera system and method to improve measurement quality of weak field-of-view signal regions
US10036801B2 (en)2015-03-052018-07-31Big Sky Financial CorporationMethods and apparatus for increased precision and improved range in a multiple detector LiDAR array
US9786101B2 (en)2015-05-192017-10-10Hand Held Products, Inc.Evaluating image values
US10066982B2 (en)2015-06-162018-09-04Hand Held Products, Inc.Calibrating a volume dimensioner
US20160377414A1 (en)2015-06-232016-12-29Hand Held Products, Inc.Optical pattern projector
US9857167B2 (en)2015-06-232018-01-02Hand Held Products, Inc.Dual-projector three-dimensional scanner
US9835486B2 (en)2015-07-072017-12-05Hand Held Products, Inc.Mobile dimensioner apparatus for use in commerce
EP3118576B1 (en)2015-07-152018-09-12Hand Held Products, Inc.Mobile dimensioning device with dynamic accuracy compatible with nist standard
US10094650B2 (en)2015-07-162018-10-09Hand Held Products, Inc.Dimensioning and imaging items
US20170017301A1 (en)2015-07-162017-01-19Hand Held Products, Inc.Adjusting dimensioning results using augmented reality
WO2017061104A1 (en)2015-10-092017-04-13パナソニックIpマネジメント株式会社Imaging apparatus and solid-state image element used for same
EP3156825B1 (en)*2015-10-162018-08-29Hand Held Products, Inc.Dimensioning system with multipath interference mitigation
US10249030B2 (en)2015-10-302019-04-02Hand Held Products, Inc.Image transformation for indicia reading
US10225544B2 (en)2015-11-192019-03-05Hand Held Products, Inc.High resolution dot pattern
US10025314B2 (en)2016-01-272018-07-17Hand Held Products, Inc.Vehicle positioning and object avoidance
US10339352B2 (en)2016-06-032019-07-02Hand Held Products, Inc.Wearable metrological apparatus
US9940721B2 (en)2016-06-102018-04-10Hand Held Products, Inc.Scene change detection in a dimensioner
US10163216B2 (en)2016-06-152018-12-25Hand Held Products, Inc.Automatic mode switching in a volume dimensioner
US10742390B2 (en)*2016-07-132020-08-11Novatek Microelectronics Corp.Method of improving clock recovery and related device
US10909708B2 (en)2016-12-092021-02-02Hand Held Products, Inc.Calibrating a dimensioner using ratios of measurable parameters of optic ally-perceptible geometric elements
US11047672B2 (en)2017-03-282021-06-29Hand Held Products, Inc.System for optically dimensioning
CN109314776B (en)*2017-05-172021-02-26深圳配天智能技术研究院有限公司Image processing method, image processing apparatus, and storage medium
US10733748B2 (en)2017-07-242020-08-04Hand Held Products, Inc.Dual-pattern optical 3D dimensioning
US11340339B2 (en)2017-12-222022-05-24Waymo LlcSystems and methods for adaptive range coverage using LIDAR
US10999524B1 (en)*2018-04-122021-05-04Amazon Technologies, Inc.Temporal high dynamic range imaging using time-of-flight cameras
US10584962B2 (en)2018-05-012020-03-10Hand Held Products, IncSystem and method for validating physical-item security
US10708484B2 (en)2018-06-202020-07-07Amazon Technologies, Inc.Detecting interference between time-of-flight cameras using modified image sensor arrays
US10681338B1 (en)2018-07-242020-06-09Amazon Technologies, Inc.Detecting interference in depth images captured using overlapping depth cameras
CN109286806A (en)*2018-09-112019-01-29Oppo广东移动通信有限公司 Deep acquisition module and mobile terminal
US10915783B1 (en)2018-12-142021-02-09Amazon Technologies, Inc.Detecting and locating actors in scenes based on degraded or supersaturated depth data
WO2020223879A1 (en)*2019-05-062020-11-12深圳市大疆创新科技有限公司Distance measurement apparatus and mobile platform
US11639846B2 (en)2019-09-272023-05-02Honeywell International Inc.Dual-pattern optical 3D dimensioning
WO2021065138A1 (en)*2019-10-012021-04-08パナソニックIpマネジメント株式会社Distance measurement device and control method
EP4014066A4 (en)*2019-12-112023-01-25Samsung Electronics Co., Ltd.Electronic apparatus and method for controlling thereof
WO2021193645A1 (en)*2020-03-242021-09-30株式会社小糸製作所Gating camera, sensing system, and vehicle lamp
US12025701B2 (en)*2020-10-122024-07-02Guangzhou Woya Laideling Technology Co., Ltd.Dynamic signal control in flash LiDAR
CN116868026A (en)*2021-03-222023-10-10深圳市大疆创新科技有限公司Range finding device, imaging device and cloud platform
DE102021130999A1 (en)*2021-11-252023-05-25Kuka Deutschland Gmbh Process, laser-optical detection system and robot workstation

Citations (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20020015144A1 (en)*2000-05-022002-02-07Asahi Kogaku Kogyo Kabushiki KaishaThree-dimensional image capturing device
US20090273770A1 (en)*2008-04-302009-11-05Honeywell International Inc.Systems and methods for safe laser imaging, detection and ranging (lidar) operation

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20020015144A1 (en)*2000-05-022002-02-07Asahi Kogaku Kogyo Kabushiki KaishaThree-dimensional image capturing device
US20090273770A1 (en)*2008-04-302009-11-05Honeywell International Inc.Systems and methods for safe laser imaging, detection and ranging (lidar) operation

Cited By (33)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20150116459A1 (en)*2013-10-252015-04-30Lips IncorporationSensing device and signal processing method thereof
US20160050346A1 (en)*2014-06-132016-02-18Lips CorporationIntegrated depth camera
US9420149B2 (en)*2014-06-132016-08-16Lips CorporationIntegrated depth camera
US20160373722A1 (en)*2015-06-192016-12-22Amazon Technologies, Inc.Steganographic depth images
US10158840B2 (en)*2015-06-192018-12-18Amazon Technologies, Inc.Steganographic depth images
US9786080B1 (en)*2015-07-022017-10-10Yesvideo, Inc.2D/3D image scanning and compositing
US10210644B1 (en)2015-07-022019-02-19Yesvideo, Inc.Image capture using target area illumination
US11838626B2 (en)2016-03-032023-12-054D Intellectual Properties, LlcMethods and apparatus for an active pulsed 4D camera for image acquisition and analysis
US10873738B2 (en)2016-03-032020-12-224D Intellectual Properties, LlcMulti-frame range gating for lighting-invariant depth maps for in-motion applications and attenuating environments
US12418719B2 (en)2016-03-032025-09-164D Intellectual Properties, LlcVehicle headlamp lighting control module for adas and autonomous vehicles
US11477363B2 (en)2016-03-032022-10-184D Intellectual Properties, LlcIntelligent control module for utilizing exterior lighting in an active imaging system
US12096121B2 (en)2016-03-032024-09-174D Intellectual Properties, LlcPrecision reflectivity and ambient light removal for a geiger mode/single photon active sensor system
US10778867B1 (en)2016-03-232020-09-15Amazon Technologies, Inc.Steganographic camera communication
US10212306B1 (en)2016-03-232019-02-19Amazon Technologies, Inc.Steganographic camera communication
US10802183B2 (en)2016-07-292020-10-13Lumentum Operations LlcThin film total internal reflection diffraction grating for single polarization or dual polarization
US10241244B2 (en)2016-07-292019-03-26Lumentum Operations LlcThin film total internal reflection diffraction grating for single polarization or dual polarization
US10996320B2 (en)*2017-07-112021-05-04Sony Semiconductor Solutions CorporationElectronic device and control method of electronic device
US12085675B2 (en)2018-04-092024-09-10Sense Photonics, Inc.Automatic gain control for LIDAR for autonomous vehicles
EP3756032A4 (en)*2018-04-092021-12-01Sense Photonics, Inc.Automatic gain control for lidar for autonomous vehicles
WO2019199645A1 (en)2018-04-092019-10-17Sense Photonics, Inc.Automatic gain control for lidar for autonomous vehicles
US11041957B2 (en)2018-06-252021-06-22Toyota Motor Engineering & Manufacturing North America, Inc.Systems and methods for mitigating effects of high-reflectivity objects in LiDAR data
US11001979B2 (en)2018-08-132021-05-11Vergence Automation, Inc.Methods and apparatus for ultrawide entrance angle reflective articles for use with autonomous vehicle machine vision systems
US11505903B2 (en)2018-08-132022-11-22Vergence Automation, Inc.Methods and apparatus for ultra wide entrance angle reflective articles for use with autonomous vehicle machine vision systems
US11762133B1 (en)2018-09-132023-09-19Vergence Automation, Inc.Retroreflective materials and articles incorporating near-ideal total internal retroreflective elements
US12072515B2 (en)2018-09-132024-08-27Vergence Automation, Inc.Retroreflective materials and articles incorporating near-ideal total internal retroreflective elements
WO2020058264A1 (en)*2018-09-192020-03-26Sony Semiconductor Solutions CorporationTime of flight apparatus and method
US12222455B2 (en)2018-09-192025-02-11Sony Semiconductor Solutions CorporationTime of flight apparatus and method
CN110312117B (en)*2019-06-122021-06-18北京达佳互联信息技术有限公司Data refreshing method and device
CN110312117A (en)*2019-06-122019-10-08北京达佳互联信息技术有限公司Method for refreshing data and device
US11514594B2 (en)2019-10-302022-11-29Vergence Automation, Inc.Composite imaging systems using a focal plane array with in-pixel analog storage elements
US12230649B2 (en)2020-05-192025-02-18Vergence Automation, Inc.AI system on chip (SOC) for robotics vision applications
US11606517B1 (en)2021-06-072023-03-14Waymo LlcEnhanced depth of focus cameras using variable apertures and pixel binning
US12143734B1 (en)2021-06-072024-11-12Waymo LlcEnhanced depth of focus cameras using variable apertures and pixel binning

Also Published As

Publication numberPublication date
US20130201288A1 (en)2013-08-08

Similar Documents

PublicationPublication DateTitle
US9294754B2 (en)High dynamic range and depth of field depth camera
US10949647B2 (en)System and method of efficient illuminator-sensor synchronization to capture images
US11330199B2 (en)Method and system of adaptable exposure control and light projection for cameras
US9516295B2 (en)Systems and methods for multi-channel imaging based on multiple exposure settings
US10863155B2 (en)Reduction of banding artifacts in image processing
CN104717435B (en)Camera device and image capture method
US10009554B1 (en)Method and system for using light emission by a depth-sensing camera to capture video images under low-light conditions
US9578224B2 (en)System and method for enhanced monoimaging
RU2627933C2 (en)Image capturing device and device control method
US20130278738A1 (en)Image processing apparatus and image processing method
JP2021500820A (en) Imaging control method and imaging device
US8743226B2 (en)Exposure adjustment method for night-vision camera
WO2016203727A1 (en)Medical image processing apparatus, medical image processing method, and medical observation system
US10200623B1 (en)Image capture setting determination in flash photography operations
US10663593B2 (en)Projector apparatus with distance image acquisition device and projection method
JP6950698B2 (en) Imaging control device, imaging control method and imaging device
US20120050490A1 (en)Method and system for depth-information based auto-focusing for a monoscopic video camera
EP3609175B1 (en)Apparatus and method for generating moving image data including multiple section images in electronic device
US9838667B2 (en)Image pickup apparatus, image pickup method, and non-transitory computer-readable medium
US20200145641A1 (en)Image processing apparatus and method
WO2016055090A1 (en)Microscope and method for obtaining a high dynamic range synthesized image of an object
US20210400252A1 (en)Imaging method, imaging system, manufacturing system, and method for manufacturing a product
US20100039561A1 (en)System, method, computer-readable medium, and user interface for displaying light radiation
CN116208851B (en) Image processing method and related device
US20110243442A1 (en)Video Camera for Acquiring Images with Varying Spatio-Temporal Resolutions

Legal Events

DateCodeTitleDescription
ASAssignment

Owner name:JDS UNIPHASE CORPORATION, CALIFORNIA

Free format text:ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BILLERBECK, BRYED;TIEN, AN-CHUN;MORALES, LUCAS;SIGNING DATES FROM 20130129 TO 20130329;REEL/FRAME:030151/0382

ASAssignment

Owner name:LUMENTUM OPERATIONS LLC, CALIFORNIA

Free format text:ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:JDS UNIPHASE CORPORATION;REEL/FRAME:036420/0340

Effective date:20150731

FEPPFee payment procedure

Free format text:PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

ASAssignment

Owner name:LUMENTUM OPERATIONS LLC, CALIFORNIA

Free format text:CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON PAGE A-A33 PREVIOUSLY RECORDED ON REEL 036420 FRAME 0340. ASSIGNOR(S) HEREBY CONFIRMS THE PATENT NUMBERS 7,868,247 AND 6,476,312 WERE LISTED IN ERROR AND SHOULD BE REMOVED;ASSIGNOR:JDS UNIPHASE CORPORATION;REEL/FRAME:037562/0513

Effective date:20150731

Owner name:LUMENTUM OPERATIONS LLC, CALIFORNIA

Free format text:CORRECTIVE ASSIGNMENT TO CORRECT INCORRECT PATENTS 7,868,247 AND 6,476,312 ON PAGE A-A33 PREVIOUSLY RECORDED ON REEL 036420 FRAME 0340. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT;ASSIGNOR:JDS UNIPHASE CORPORATION;REEL/FRAME:037562/0513

Effective date:20150731

ASAssignment

Owner name:LUMENTUM OPERATIONS LLC, CALIFORNIA

Free format text:CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON PAGE A-A33 PATENT NUMBERS 7,868,247 AND 6,476,312 WERE LISTED IN ERROR AND SHOULD BE REMOVED. PREVIOUSLY RECORDED ON REEL 036420 FRAME 0340. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT;ASSIGNOR:JDS UNIPHASE CORPORATION;REEL/FRAME:037627/0641

Effective date:20150731

Owner name:LUMENTUM OPERATIONS LLC, CALIFORNIA

Free format text:CORRECTIVE ASSIGNMENT TO CORRECT PATENTS 7,868,247 AND 6,476,312 LISTED ON PAGE A-A33 PREVIOUSLY RECORDED ON REEL 036420 FRAME 0340. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT;ASSIGNOR:JDS UNIPHASE CORPORATION;REEL/FRAME:037627/0641

Effective date:20150731

STCFInformation on status: patent grant

Free format text:PATENTED CASE

CCCertificate of correction
ASAssignment

Owner name:DEUTSCHE BANK AG NEW YORK BRANCH, AS COLLATERAL AGENT, NEW YORK

Free format text:PATENT SECURITY AGREEMENT;ASSIGNORS:LUMENTUM OPERATIONS LLC;OCLARO FIBER OPTICS, INC.;OCLARO, INC.;REEL/FRAME:047788/0511

Effective date:20181210

Owner name:DEUTSCHE BANK AG NEW YORK BRANCH, AS COLLATERAL AG

Free format text:PATENT SECURITY AGREEMENT;ASSIGNORS:LUMENTUM OPERATIONS LLC;OCLARO FIBER OPTICS, INC.;OCLARO, INC.;REEL/FRAME:047788/0511

Effective date:20181210

MAFPMaintenance fee payment

Free format text:PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment:4

ASAssignment

Owner name:LUMENTUM OPERATIONS LLC, CALIFORNIA

Free format text:RELEASE BY SECURED PARTY;ASSIGNOR:DEUTSCHE AG NEW YORK BRANCH;REEL/FRAME:051287/0556

Effective date:20191212

Owner name:OCLARO FIBER OPTICS, INC., CALIFORNIA

Free format text:RELEASE BY SECURED PARTY;ASSIGNOR:DEUTSCHE AG NEW YORK BRANCH;REEL/FRAME:051287/0556

Effective date:20191212

Owner name:OCLARO, INC., CALIFORNIA

Free format text:RELEASE BY SECURED PARTY;ASSIGNOR:DEUTSCHE AG NEW YORK BRANCH;REEL/FRAME:051287/0556

Effective date:20191212

FEPPFee payment procedure

Free format text:MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

LAPSLapse for failure to pay maintenance fees

Free format text:PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCHInformation on status: patent discontinuation

Free format text:PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FPLapsed due to failure to pay maintenance fee

Effective date:20240322


[8]ページ先頭

©2009-2025 Movatter.jp