Movatterモバイル変換


[0]ホーム

URL:


US10896492B2 - Cognitive load reducing platform having image edge enhancement - Google Patents

Cognitive load reducing platform having image edge enhancement
Download PDF

Info

Publication number
US10896492B2
US10896492B2US16/383,158US201916383158AUS10896492B2US 10896492 B2US10896492 B2US 10896492B2US 201916383158 AUS201916383158 AUS 201916383158AUS 10896492 B2US10896492 B2US 10896492B2
Authority
US
United States
Prior art keywords
gradient magnitude
subregions
image
mapping parameters
calculating
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
US16/383,158
Other versions
US20200151859A1 (en
Inventor
II John Davis Long
Omer Haciomeroglu
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Qwake Technologies Inc
Original Assignee
Qwake Technologies Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Qwake Technologies IncfiledCriticalQwake Technologies Inc
Priority to US16/383,158priorityCriticalpatent/US10896492B2/en
Assigned to QWAKE TECHNOLOGIESreassignmentQWAKE TECHNOLOGIESASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS).Assignors: HACIOMEROGLU, OMER, LONG, JOHN DAVIS, II
Publication of US20200151859A1publicationCriticalpatent/US20200151859A1/en
Assigned to QWAKE TECHNOLOGIES, LLCreassignmentQWAKE TECHNOLOGIES, LLCCORRECTIVE ASSIGNMENT TO CORRECT THE ASSIGNEE NAME PREVIOUSLY RECORDED AT REEL: 048874 FRAME: 0400. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT.Assignors: HACIOMEROGLU, OMER, LONG, JOHN DAVIS, II
Priority to US17/113,694prioritypatent/US11610292B2/en
Application grantedgrantedCritical
Publication of US10896492B2publicationCriticalpatent/US10896492B2/en
Assigned to QWAKE TECHNOLOGIES, INC.reassignmentQWAKE TECHNOLOGIES, INC.ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS).Assignors: QWAKE TECHNOLOGIES, LLC
Activelegal-statusCriticalCurrent
Anticipated expirationlegal-statusCritical

Links

Images

Classifications

Definitions

Landscapes

Abstract

Enhancing edges of objects in a thermal image comprises receiving a thermal image and generating a gradient magnitude image comprising a plurality of pixels having associated gradient magnitude values. The gradient magnitude image is partitioned into subregions and gradient magnitude statistics are calculated for each. Mapping parameters are calculated for each of the subregions that equalize and smooth a dynamic range of the corresponding gradient magnitude statistics across the subregions. The mapping parameters calculated for each of the subregions are applied to pixels in the subregions to generate enhanced gradient magnitude values having equalized luminosity and contrast, and a wireframe image is formed therefrom having enhanced edges of objects. The wireframe image is displayed on a display device, wherein the wireframe image appears as a decluttered line drawing where the enhanced edges have increased luminosity and contrast compared to the thermal image to reduce the cognitive load of the user.

Description

CROSS-REFERENCE TO RELATED APPLICATION
This application claims the benefit of provisional Patent Application Ser. No. 62/758,457, filed Nov. 9, 2018, assigned to the assignee of the present application, and incorporated herein by reference.
BACKGROUND
In high stress and oftentimes hazardous work environments-including firefighting, search & rescue, oil and gas, surgery, fighter pilots, mining, special ops, and the like, one false step has critical consequences, but so do too many slow steps. Go too fast and something life-threatening may be missed; go too slow and the results could be doubly devastating. The challenges of effectively and safely performing critical work in harsh and obscured environments have always existed. These challenges combine the physical strain imposed by hazardous terrain with the mental distress placed upon the individual operating within them. Critical human performance in high-stress environments is limited by how rapidly and effectively the brain can process impoverished or jumbled sensory inputs. Until now technology has been leveraged primarily to increase the amount of information provided to the senses, but not designed to specifically enhance the brain's existing (and unmatched) cognitive ability to make sense of that information.
For example, several emergency response systems are centered on the use of thermal imaging cameras (TICs) and augmented reality (AR) optics to provide a hands-free thermal display to the user. Current systems are typically carried by a crewmember who must iteratively scan, mentally process and communicate what they perceive. Current handheld and hands-free TICs lack the computational resources and software required to unobtrusively offer advanced image processing and data visualization features to all crewmembers in real-time. This capability and time gap in the visual understanding of hazardous environments has been identified as a significant causative factor in responder line of duty deaths. Such systems cause crewmembers, such as first responders, to operate in a Stop, Look, Process and Remember paradigm, which is cumbersome and time consuming.
Accordingly, there is a need for improved methods and systems for effectively for providing information to the senses of first responders operating in high-stress environments in a manner that reduces cognitive load to enhance performance.
BRIEF SUMMARY
The exemplary embodiment provides a cognitive load reducing platform for first responders that incorporates contextual and physiological visualizations delivered via electronic communications designed to reduce cognitive load and elevate performance. Aspects of the cognitive load reducing platform include enhancing edges of objects in a thermal image by receiving a thermal image and generating a gradient magnitude image comprising a plurality of pixels having associated gradient magnitude values. The gradient magnitude image is partitioned into subregions and gradient magnitude statistics are calculated for each. Mapping parameters are calculated for each of the subregions that equalize and smooth a dynamic range of the corresponding gradient magnitude statistics across the subregions. The mapping parameters calculated for each of the subregions are applied to pixels in the subregions to generate enhanced gradient magnitude values having equalized luminosity and contrast, and a wireframe image is formed therefrom having enhanced edges of objects. The wireframe image is displayed on a display device, wherein the wireframe image appears as a decluttered line drawing where the enhanced edges have increased luminosity and contrast compared to the thermal image to reduce the cognitive load of the user.
According to the method and system disclosed herein, the cognitive load reducing platform changes first responders' use of these technologies from the cumbersome, and information overloaded, Stop, Look, Process and Remember paradigm to a state of continuous assisted perception for all crewmembers.
BRIEF DESCRIPTION OF SEVERAL VIEWS OF THE DRAWINGS
FIG. 1 is a diagram illustrating one embodiment of a cognitive load reducing platform according to one embodiment.
FIGS. 2A and 2B are diagrams illustrating components of the cognitive load reducing platform in a firefighting vision system embodiment.
FIGS. 3A and 3B are diagrams illustrating an example of the enhanced images displayed on the display unit.
FIG. 4 is a diagram illustrating an example command and control interface of the cognitive load reducing platform.
FIG. 5 is a flow diagram illustrating the process performed by the edge enhancement engine on the thermal images to enhance the edges of objects therein to improve user vision and navigation a user in obscure environments, while reducing cognitive load of the user.
FIG. 6 is an illustration an example thermal image received from the TIC.
FIG. 7 is an illustration of an example gradient magnitude image generated from the thermal image ofFIG. 6.
FIG. 8 is a diagram graphically illustrating the gradient magnitude image partitioned into a set of non-overlapping subregions.
FIG. 9 is a diagram illustrating an example subregion histogram.
FIG. 10 is a diagram illustrating an example cumulative histogram generated from the gradient magnitude value histogram of one particular subregion.
FIG. 11 is a diagram graphically illustrating an example oflinear mapping1100 of subregion gradient magnitude values represented by line formed between control points (L, U) to predetermined reference gradient magnitude values.
FIG. 12 is a diagram graphically illustrating an example of one embodiment of the weight function used to interpolate the subregion mapping parameters before applying the mapping parameters to pixels (p) in a current subregion.
DETAILED DESCRIPTION
The exemplary embodiment relates to methods and systems for incorporating contextual and physiological visualizations into electronic communications via a cognitive load reducing platform having image edge enhancement. The following description is presented to enable one of ordinary skill in the art to make and use the invention and is provided in the context of a patent application and its requirements. Various modifications to the exemplary embodiments and the generic principles and features described herein will be readily apparent. The exemplary embodiments are mainly described in terms of particular methods and systems provided in particular implementations. However, the methods and systems will operate effectively in other implementations. Phrases such as “exemplary embodiment”, “one embodiment” and “another embodiment” may refer to the same or different embodiments. The embodiments will be described with respect to systems and/or devices having certain components. However, the systems and/or devices may include more or less components than those shown, and variations in the arrangement and type of the components may be made without departing from the scope of the invention. The exemplary embodiments will also be described in the context of particular methods having certain steps. However, the method and system operate effectively for other methods having different and/or additional steps and steps in different orders that are not inconsistent with the exemplary embodiments. Thus, the present invention is not intended to be limited to the embodiments shown, but is to be accorded the widest scope consistent with the principles and features described herein.
In many critical, high-stress activities, such as firefighting, specialized tools have been developed to support challenging environments and critical objectives of crewmembers engaged in the high-stress activities. For the most part, these tools have evolved to support the crewmembers' physical needs—heat protection, airway protection, forcible entry, fire suppression, and the like. In the past 10-15 years, a greater focus has been placed on supporting the crewmembers' informational needs, including hazardous environment detection, communication, and safety alerting. For example, hearing aids, binoculars, and seismic sensors all increase the collection of information, but don't increase crewmembers' abilities to process or critically discern that extra information. Polarized glasses, gas monitors, thermal imagers, and the like all collect information, but still do not address the time and stress penalty required to absorb and interpret all that information. This “more is better” approach is both distracting and inefficient.
Unfortunately, often times stress is the limiting factor to crewmembers successfully completing these critical and dangerous activities. These are, by definition, high-stress environments and the difficulty in absorbing more and more information is made worse by stress. The health of the crewmembers is also compromised by stress, and regrettably contributes to a majority of crewmember fatalities every year.
The exemplary embodiments provide a cognitive load reducing platform having edge enhancement that leverages the principles of neuroscience and the tools of computer vision to reduce the cognitive load of a user and elevate human performance in high stress environments. The principles of neuroscience are used to integrate sensor data into the natural senses in a manner that is optimized for the task at hand, e.g. search and rescue, and computer vision supplies the means in one embodiment. The cognitive load reducing platform significantly enhances the crewmembers' or user's ability to make well informed decisions rapidly when operating in complex environments where cognitive abilities decline. A premise of the cognitive load reducing platform is that if thinking and understanding are easier for crewmembers, then crewmember can achieve objectives more rapidly, spend less time in harsh conditions, and have potentially reduced stress levels because of the real-time assurance or reinforcement of a human sense, i.e., vision, hearing and or touch. Example users of the cognitive load reducing platform include, but are not limited to, firefighters, surgeons, soldiers, police officers, search and rescue and other types of first responders.
FIG. 1 is a diagram illustrating one embodiment of a cognitive load reducing platform according to one embodiment. In one embodiment, the cognitive load reducing platform is a wearable electronic system worn on the body of a user when the user is engaged in complex, high stress environments that reduce cognitive abilities.
The cognitiveload reducing platform10 comprises, one or more sensors12a-12n(collectively sensors12) that collect information about an environment as sensor data. The information collected about the environment refers primarily to sensor data that can be used for navigation and detecting hazards, but also to a user's health status. In one embodiment, the sensors are worn by the crewmembers. For example, multiple sensors may be incorporated into a sensor package that is worn by one or more crewmembers. In another embodiment, other sensors may be remote from the crewmembers, such as on a drone equipped with a camera, gas detector, and the like.
Example categories of sensors includes situational awareness sensors and biometric sensors for health status. The situational awareness sensors measure data about the user's external environment for hazard detection and navigation. Examples of situational awareness sensors for hazard detection may include, but are not limited to: cameras (e.g., a TIC, a radiometric thermal camera, a drone camera), a spectrometer, a photosensor, magnetometer, a seismometer, a gas detector, a chemical sensor, a radiological sensor, a voltage detector, a flow sensor, a scale, a thermometer, a pressure sensor, and an acoustic sensor (e.g., for selective active noise cancellation inside the mask to facilitate radio communication). Examples of situational awareness sensors used for user navigation may include, but are not limited to: an inertial measurement unit (IMU), a GPS sensor, a speedometer, a pedometer, an accelerometer, an altimeter, a barometer, attitude indicator, a depth gauge, a compass (e.g., a fluxgate compass), a gyroscope, and the like. Examples of biometric sensors that measure health conditions/status of the user may include, but are not limited to: a heart rate sensor, a blood pressure monitor, a glucose sensor, an electrocardiogram (EKG or ECG) sensor, an electroencephalogram (EEG) sensor, an electromyography (EMG) sensor, a respiration sensor, and a neurological sensor.
The platform also includes a high-speed processor complex14 coupled to the sensors12. The high-speed processor complex14 includes amemory16, acommunication interface19, and one or more processors18, such as graphics processor units (GPUs). The processor/GPUs18 execute one more software-basedcognitive enhancement engines20 to process the sensor data from the sensors12 into enhanced characterization data that incorporate contextual and physiological visuals, auditory and/or haptic cues. The cognitiveload reducing platform200 is sensor agnostic and as any type of sensor can be added to the platform as long a correspondingcognitive enhancement engine20 is provided to process and present that sensor data. Thememory16 may contain sensor data collected from the one or more sensors12, and may store thecognitive enhancement engines20 for implementing the processes described herein when executed by the one or more processors18. Although theedge enhancement engines20 are described as single modules, the functionality provided by theedge enhancement engines20 may be performed by multiple modules.
The cognitiveload reducing platform10 further includes one ormore output devices22 coupled to theprocessor complex14 to electronically communicate the enhanced characterization data to the wearer. In one embodiment, theoutput devices22 may be implemented as a visual display, headphones/ear buds and/or a haptic device. The enhanced characterization data is thus integrated into natural senses of the wearer in a manner that is optimized for the performance of a specific task of the user—navigating in limited, to no, view environments.
Prior solutions increase the amount of information provided to the user's senses without specifically enhancing the brain's existing (and unmatched) cognitive ability to make sense of that information. The cognitiveload reducing platform10, in contrast, filters, summarizes, and focuses sensor data into the enhanced characterization data comprising contextual and physiological visuals, audio and/or haptic cues to create a new category called “Assisted Perception” that significantly reduces complexity and cognitive load (and accompanying stress)—and decreases Time-To-Clarity required to save lives. The cognitiveload reducing platform10 is designed to reduce risk, improve human safety, and save lives.
The cognitive load reducing platform supports the introduction of life-saving, Assisted Perception solutions to high-stress environments. One example use of this new category of Assisted Perception is as a firefighting vision system. In this embodiment, the cognitive load reducing platform is a real-time computer vision engine designed to aid first responders as they navigate smoke filled, hazardous environments with little or no visible light. In this embodiment, the cognitive load reducing platform increases the speed and safety of first responders in the field with a focus upon navigation and visual communication applications. The Assisted Perception of the cognitive load reducing platform dramatically enhances one's ability to make well informed decisions rapidly when operating in complex environments where cognitive abilities decline. The platform has shown performance improvements of 267% (e.g., reducing the time to complete mission critical search and rescue tasks from 4.5 mins to 1.7 mins).
Several emergency response systems are based on the use of a thermal camera and AR optics to provide a hands-free imaging system to the user. However, the cognitive load reducing platform provides a novel integrated design of these hardware and software elements into a system that efficiently integrates into natural human visual perception in a manner that decreases stress in the field. In the first responder embodiment, the platform combines a unique combination of enhanced thermal imaging, augmented reality (AR) or virtual reality (VR), and environment visualization and mapping capabilities.
FIGS. 2A and 2B are diagrams illustrating components of the cognitive load reducing platform in the firefighting vision system embodiment. Referring toFIG. 2A, in one embodiment, the cognitiveload reducing platform200 comprises two components: i) an assistedperception module221 integrated with a face mask224 (or simply “mask”) of the SCBA worn by each user/crewmember; and ii) a command andcontrol interface226 displayed on adisplay device228 to a person of authority, such as an incident commander. The command andcontrol interface226 enables the person of authority to manage the incident and the crewmembers wearing respective assistedperception modules221 by displaying the enhanced characterization data from each of the respective assisted perception modules.
Each of the assistedperception modules221 comprises a modular set of components including aTIC212, aprocessor complex214 in communication with theTIC212 for executing anedge enhancement engine220, and adisplay unit222, which is removably attached to themask224. In relation toFIG. 1, theTIC212 inFIG. 2A comprises one of the sensors12 of theplatform200, thedisplay unit222 comprises theoutput device22, and theedge enhancement engine220 comprises thecognitive enhancement engine20 executed on theprocessor complex214.
In the embodiment shown, thedisplay unit222 may comprise an augmented reality (AR) display unit, a virtual reality (VR) display unit, or a head-mounted projection display unit. In the AR embodiment, the AR display unit may comprise optical see through glasses that can be either binocular or monocular.FIG. 2A shows an embodiment where thedisplay unit222 comprises an AR monocular display.
FIG. 2B is a diagram showing an exploded view of the assistedperception module221 showing additional components in a wireless embodiment. TheTIC212 captures thermal image data of obstacles and objects251 and sends the thermal image data via acable250 to theprocessor complex14. The exploded view of theprocessor complex14 shows that theprocessor complex14 is powered bybatteries252 and thecommunication interface19 ofFIG. 1 is coupled to anantenna254 to wirelessly transmit enhanced image data to thedisplay unit222. Thedisplay unit222 further includes aprocessing board256 with anantenna258 to receive the enhanced image data wirelessly transmitted from theprocessor complex14. Theprocessing board256 relays the enhanced image data to a digital signal processing board260, which processes the enhanced image data for display on thedisplay unit222 directly in front of the user's eye.
As stated above, in one embodiment, the cognitive load reducing platform is a wearable electronic system. As such, there are many placement embodiments for the components of the cognitive load reducing platform. In most embodiments, all components are located on, or otherwise carried by, a user. For example,FIGS. 2A and 2B illustrate an embodiment where the TIC212 (sensors12) and the display unit222 (output device22) are carried by the user by virtue of being integrated withmask224. Theprocessor complex214 is also worn by the user, such as being clipped to a belt or clothing, stowed in a pouch or a pocket, or attached to a back frame of the SCBA.
In some embodiments, however, the sensors12 and/or theprocessor complex14 may be located remote from the user. As an example, consider the use case where a remote gas sensor controlled by a third party sends gas data over a network (e.g., Internet) to the cognitiveload reducing platform10. In this embodiment, the cognitiveload reducing platform10 may be implemented at least in part as a website where theprocessor complex14 is implemented as one or more servers that wirelessly receives sensor data of various types. In this embodiment, the gas sensor data from the remote gas sensor is pushed to theprocessor complex14 where the sensor data is processed locally by a correspondingcognitive enhancement engine20, which converts and outputs a brain optimized visual format for display to the user on thedisplay unit22. In one embodiment, a third party could collect and push the gas sensor data into the cognitive load reducing platform in the cloud.
There are also many communication embodiments for the components of the cognitive load reducing platform. For example, in the embodiment shown inFIGS. 2A and 2B, theTIC212 communicates with theprocessor complex214 over a wired connection viacable250. However, in an alternative embodiment, theTIC212 wirelessly communicates with theprocessor complex14. In the embodiment shown inFIGS. 2A and 2B, theprocessor complex214 communicates with the AR display unit over a wireless connection. However, in an alternative embodiment, theprocessor complex214 communicates with the AR display unit over a wired connection. In yet a further embodiment, theTIC212 and theprocessor complex214 may be integrated together within a common housing, in which case theTIC212 is in direct communication with theprocessor complex214, and the communication between theprocessor complex214 and thedisplay unit222 could be over a wired or wireless connection. In either embodiment, thedisplay unit222 would still communicate wirelessly with the centralcommand display device228.
In one embodiment, the display unit222 (including digital signal processing board260, processingboard256, and antenna258) is mounted inside themask224. However, in an alternative embodiment, thedisplay unit222 is mounted outside themask224. For example, the display itself may be positioned outside themask224, while the digital signal processing board260, processingboard256 andantenna258, may be worn by the user, such as being clipped to a belt or clothing, stowed in a pouch or a pocket, or attached to a back frame of the SCBA.
According to one aspect of the disclosed embodiments, theedge enhancement engine220 executed by theprocessor complex214 in the firefighting embodiment performs high speed processing on the thermal images from theTIC212 in real time into enhanced characterization data comprising a stream of wireframe images where edges of objects appearing in the thermal images have increased luminosity and contrast in relation to the thermal images. The display unit is worn in a line of sight of the user and electronically displays the enhanced characterization data, where the stream of the wireframe images appear as decluttered, enhanced line drawings to enable the user to see and navigate in obscure conditions, while reducing the cognitive load of the user. In one embodiment, the enhanced characterization data comprises a stream of AR/VR or wireframe images having enhanced the edges or outlines of objects that are projected on the display device in the user's field of view, so the user can see and effectively navigate in dark or smoke-filled environments without overwhelming the user's ability to process the displayed information. In this embodiment, the enhanced line drawing images produced by the platform dramatically enhance the user's ability to make well-informed decisions rapidly when operating in complex environments where cognitive abilities decline, such as a first responder (e.g., a fire fighter or search and rescue personnel).
FIGS. 3A and 3B are diagram illustrating example enhanced wireframe images displayed on thedisplay unit222. The enhanced wireframe images are shown on a transparent AR or head mounted projection display placed directly within the user's line of sight to provide real time navigation. The enhanced wireframe images of the disclosed embodiments have enhanced edges and a decluttered appearance after processing of thermal images/video by the cognitive load reducing platform. In other words, the enhanced images clarify the TIC video by highlighting key details with bright edges and minimizing the rest. Notice that theedge enhancement engine220 is not optimized to recognize objects in the thermal images nor to create continuous lines outlining the objects, but is rather optimized to reduce visual complexity by filtering out faint elements and boosting contrast throughout the image. A close review of the enhanced wireframe images inFIGS. 3A and 3B reveals some or all the edges outlining objects may have small breaks and thus may not meet the mathematical definition of a continuous line, but when viewed by a human, the brain easily completes the scene and can recognize simple objects sufficiently enough to navigate. In an alternative embodiment, theedge enhancement engine220 may create wireframe images having continuous lines outlining the objects.
The enhanced wireframe images overlayed in front of a user's eye in real time are designed to save lives by providing the user (e.g., a firefighter) with just the information the user needs, instead of overloaded the user as can be easily done in stressful situations. As shown, the assistedperception module221 enables the user to see in dark, smoke-filled environments. However, seeing through smoke is a side benefit to the value of the cognitive load reducing platform, which is to reduce the visual complexity of hazardous environments, while allowing individuals to more easily make sense of their surroundings.
The Assisted Perception provided by the cognitive load reduction platform leverages the principles of neuroscience to enhance aggregated sensor data in real-time to allow first responders to do their jobs significantly faster and more safely. The closest competitor to an infrared sensor-based, extreme environment tool would be the handheld or helmet mounted infrared camera and display systems. However, none of these systems offer any context-specific interpretive processing of the output, nor are they designed as true augmented reality interfaces that reduce the cognitive load of the user.
Referring again toFIG. 2A, the wireless capabilities of theprocessor complex14 is also used create a shared experience between the crewmembers and the incident commander (IC) via the command andcontrol interface226. According to the present embodiment, the command andcontrol interface226 runs on thedisplay device228 to the incident commander and is associated or paired to a group of theAR modules221 worn by crewmembers. Through its wireless capabilities, each assistedperception module221 streams its enhanced image to the command andcontrol interface226 on thedisplay device228 so that the incident commander can see in real time what the crewmembers see. Thus, the platform improves situational awareness on the front lines as well as to create a shared operating picture with the commander on the backend through a cognitive load reducing command andcontrol interface226 that aggregates information and presents insights to the commander via a simplified GUI based on similar principals of neuroscience.
FIG. 4 is a diagram illustrating an example command and control interface of the cognitive load reducing platform. In one embodiment, thedisplay device228 comprises a tablet computer or large mobile phone, and the command andcontrol interface226 comprises multiple sub-panels or frames for displaying the streams from each of theAR modules221 of the crewmembers (e.g., up to 8). There are two communication modes. In the first communication mode, the incident commander can select any sub-panel to engage in audio or visual icon based communication. In the second communication mode, the incident commander can broadcast to all of theAR modules221 equipped crewmembers in the field.
Traditional emergency response tools to aid the incident commander focus upon the incident commander's ability to integrate information unavailable to the crewmembers, and to then communicate these insights via radio channels. In contrast, the cognitive load reducing platform allows the incident commander to see the moment to moment visual experience of their crewmembers and to communicate back to them using visual cues displayed to crewmembers equipped withAR modules221. Consequently, the connected nature of the platform (streaming visual data betweenAR modules221 to the central command display device228) elevates the safety of the entire workspace by providing a shared operating picture between individuals in the field and leaders monitoring workers from the periphery.
Edge Enhancement
Referring again toFIGS. 2A and 2B, in the firefighting embodiment, theedge enhancement engine220 performs high speed processing on the thermal images from theTIC212 to enhance the edges or outlines of objects therein. In one embodiment, theedge enhancement engine220 utilizes histogram equalization algorithms designed to facilitate the human brain's ability to make rapid and accurate inferences related to navigation. The histogram equalization algorithms reduce and declutter thermal image data to a simple “line drawing” or “cartoon image” that leverages the underlying neuro-structure of vision to aid in navigation.
FIG. 5 is a flow diagram illustrating the process performed by theedge enhancement engine220 on the thermal images to enhance the edges of objects therein to improve user vision and navigation a user in obscure environments, while reducing cognitive load of the user.
The process may begin by receiving a series of one or more thermal images and generating a gradient magnitude image from the thermal images, the gradient magnitude image comprising a plurality of pixels having associated gradient magnitude values (block500). A series of the thermal images form a video that is streamed out ofTIC212 and into theprocessor complex214 in real-time. In one embodiment, each pixel in thermal image may be 8-bits in length and have 256 possible intensities.
FIG. 6 is an illustration an example thermal image received from the TIC. Thermal images are normally grayscale where black objects represent cold temperatures, white objects represent hot temperatures, and the depth of gray indicates variations between the two. Some thermal cameras, such as radiometric thermal cameras that measure surface temperatures, add color to the images to aid users in identifying objects at different temperatures. In the example shown, thethermal image600 is an 8-bit grayscale image having grayed edges and faint contours.
The brightness or luminosity of object edges in thethermal image600 are not uniform, ranging from very bright to very faint. When displayed to a user in this form, thethermal image600 makes it difficult for first responders to quickly determine what they are seeing, and then determine a course of action, such as determining how to quickly and safely navigate the environment. One goal of the disclosed embodiments is to generate wireframe images were object edges throughout the image have uniform brightness. In one embodiment, the object edges may have uniform thickness as well. In a further embodiment, however, the object edges may have varying thickness related to a perceived distance of the object that from the TIC/user. For example, object edges closer to the user may be displayed with increased thickness compared with object edges farther from the user.
In one embodiment, the gradient magnitude of each pixel in the thermal image may be extracted using a 2D Gaussian filter to create a smoothed image. Next, x- and y-axis gradients (gxand gy) may be approximated by convolving the smoothed image with orthogonal (e.g., vertical and horizontal) Sobel filters to determine how much gray levels in the thermal image change in vertical and horizontal directions, respectively. The output of the Sobel filters may be combined pixel-wise through a l2-norm filter to produce the gradient magnitude image.
FIG. 7 is an illustration of an example gradient magnitude image generated from the thermal image ofFIG. 6. Thegradient magnitude image700 is a local estimate of contrast per pixel. Brighter pixel levels in thegradient magnitude image700 indicate a change from a dark value to a bright value in the thermal image, and darker pixel levels in thegradient magnitude image700 indicate a change from a bright value to a dark value in the thermal image. More specifically, white pixels indicate a high change in luminosity in the initial thermal image, while black pixels indicate no change at all. Thegradient magnitude image700 has a significant amount of pixel noise from the TIC and discontinuities due to the 8-bit space.
Referring again toFIG. 5, thegradient magnitude image700 is an estimate of contrast per pixel location, but an edge or line of an object in the image is not defined by a single pixel, but rather a region of space. Therefore, referring again toFIG. 5, theedge enhancement engine220 continues by partitioning thegradient magnitude image700 into a set of subregions to analyze luminosity and contrast in regions of space in the gradient magnitude image700 (block502). In one embodiment, the gradient magnitude image is partitioned into a set of non-overlapping subregions.
FIG. 8 is a diagram graphically illustrating thegradient magnitude image700 partitioned into a set ofnon-overlapping subregions800. In an example embodiment where the thermal image comprises 640×480 pixels or 307,200 total pixels, thegradient magnitude image700 may be partitioned into 300subregions800, each 32×32 pixels is size for a total of 1024 pixels.
Referring again toFIG. 5, theedge enhancement engine220 calculates gradient magnitude value statistics for each of the subregions (block504). In one embodiment, calculating the gradient magnitude statistics comprises for each of the subregions, generating a subregion histogram of gradient magnitude values that counts the number of pixels having each gradient magnitude value (block504A); generating cumulative gradient magnitude value counts (block504B); and generating upper and lower gradient magnitude percentile values (block504C). In some embodiments, the gradient magnitude value statistics for the subregions may be computed in parallel.
FIG. 9 is a diagram illustrating anexample subregion histogram900. Onesubregion histogram900 is generated for each of thesubregions800. The x-axis of thesubregion histogram900 represents possible gradient magnitude values, and the y-axis is the number of pixels at each different gradient magnitude value found in the subregion. InFIG. 9, the peak in the graph in thesubregion histogram900 may indicate, for example, that a number “x” of pixels in the subregion have a gradient magnitude value of “y”. For an 8-bit image, each pixel has a gradient magnitude value ranging from 0 to 255, which is reflected on the x-axis. As another example, in a 16-bit image, each pixel has a gradient magnitude value ranging between 0 and 65,535.
Referring again toFIG. 5, generating the cumulative histogram of the gradient magnitude value counts for each of the subregions (block504B) may comprise performing a cumulative sum operation on each subregion gradient magnitude value histogram that returns a cumulative sum of the gradient magnitude value histogram of each subregion. The cumulative sum operation is a sequence of partial sums of a given sequence. For example, the cumulative sums of the sequence {a, b, c . . . }, are a, a+b, a+b+c, . . . .
FIG. 10 is a diagram illustrating an example cumulative histogram generated from the gradient magnitude value histogram of one particular subregion. In one embodiment, thecumulative histogram1000 of the gradient magnitude value histogram is represented as ϕ|g|. The x-axis represents the cumulative sums of the gradient magnitude value counts, which range between 0 and the number of pixels within the subregion. The y-axis shows the proportion of counts less than or equal to “x”.
According to one aspect of the disclosed embodiments, generating upper and lower gradient magnitude percentile values (FIG. 5, block504C) may comprise applying a predetermined lower bound percentage1002 and upper boundpercentage1004 to the cumulative histogram ϕ|g| for each of the subregions. The goal is to determine the gradient magnitude value at the lower bound percentage1002 and upper boundpercentage1004, which are parameters that remain the same across all the subregions, and therefore are referred to as invariants. In one embodiment, the lower bound percentile1002 is set between a 5thpercentile and a 20thpercentile, or a 10thpercentile in a particular embodiment. In one embodiment, the upper boundpercentile1004 is set between an 80thpercentile and a 95thpercentile, or a 90thpercentile in a particular embodiment.
In a further embodiment, to compensate for background noise from the TIC, an absoluteminimum percentile1006 may be placed upon the gradient magnitude values to censor noise. Basically, while theedge enhancement engine220 analyzes how statistics for the gradient magnitude values in the different subregions differ, the absoluteminimum percentile1006 is used to minimize the effect of uniform noise in the scene. The predetermined upper and lower boundgradient percentiles1002 and1004 and the absoluteminimum percentile1006 are applied to the gradient magnitude values across all the subregions.
The gradient magnitude value at apoint1008 where the cumulative histogram crosses the lower bound percentile1002 and apoint1010 where the cumulative histogram crosses the upper boundpercentile1004 may then be extracted to determine the gradient magnitude value at the 10thpercentile and the 90thpercentile. In one embodiment, “L” is set equal to thegradient magnitude value1008 at the lower bound percentile1002 (also represented as IV, and “U” is set equal to thegradient magnitude value1010 at the upper bound percentile1004 (also represented as |g|u). The output of this process are gradient magnitude value control points (L, M) for each of thesubregions800.
Referring again toFIG. 5, according to the disclosed embodiments, theedge enhancement engine220 calculates mapping parameters for each of the subregions to equalize and smooth the dynamic range of the corresponding gradient magnitude statistics across the subregions (block506). By equalizing and smoothing the dynamic range of the great magnitude statistics across the subregions, minimum and maximum gradient magnitude values in each subregion are made substantially similar or uniform. As used herein the gradient magnitude values reflect luminosity and contrast.
In one embodiment, the process of calculating the mapping parameters may comprise determining mapping parameters necessary to linearly map the control points (L, U) of eachsubregion800 to a predetermined reference gradient magnitude representing a desired dynamic range (block506A). Using this process, a spatially non-local description of contrast within each subregion is equalized in that region to match the desired dynamic range across all the subregions.
FIG. 11 is a diagram graphically illustrating an example oflinear mapping1100 of subregion gradient magnitude values represented by line1104 formed between control points (L, U) to predetermined reference gradient magnitude values represented by1102. As inFIG. 10, the x-axis represent the cumulative histogram of the gradient magnitude values, which range between 0 and 255 for the 8-bit pixels. The y-axis shows percentiles of gradient magnitude values, including the invariant predetermined lower bound percentile1002 and the upper boundpercentile1004. The reference gradient magnitude1104, the lower bound percentile1002 and the upper boundpercentile1004 remain the same across allsubregions800.
The reference gradient magnitude is represented asdiagonal line1102 formed with positive slope between the intersection of the lower bound percentile1002 (10th percentile) and gradient magnitude value |g|l, and the intersection of the upper bound percentile1004 (90thpercentile) and gradient magnitude value |g|u.
A subregion gradient magnitude is represented as a line1104 formed between control points (L, M) of the current subregion. The subregion gradient magnitude line1104 shows that the subregion is in an area of the gradient magnitude image that is dim and has low contrast (shown by the steep slope). Thelinear mapping process1100 finds mapping parameters that moves the upper value control point M of the subregion gradient magnitude line1104 to the right to align with the intersection of the upper boundpercentile1004 and |g|uof thegradient magnitude line1102; and moves the lower value control point L to the left toward the intersection of lower bound percentile1002 and |g|lof thegradient magnitude line1102. The mapping parameters will now match the subregion gradient magnitude with the reference gradient magnitude to provide improved luminosity and contrast.
In one embodiment, the linear mapping process is performed by determining an intercept and slope of the referencegradient magnitude line1102 and using the intercept and slope to create linear mapping parameters to apply to each pixel in the subregion. In one embodiment, the linear mapping parameters are calculated as:
x′=f(α,β,x); andf(α,β,x)=βx+α,
where x=a current pixel gradient magnitude value, α=intercept, and β=slope. The result of this linear mapping is the equalized pixel value x′.
Referring again toFIG. 5, the linear mapping parameters (intercept α and slope β) to be applied to each pixel of the subregion are interpolated as a function of each pixel's computed distance to neighboring subregions to generate smoothed mapping parameters to compensate for arbitrary subregion boundaries (block506B). In one embodiment, a 2D nearest neighbor interpolation scheme may be used. In another embodiment, a convolution process using a sliding window may be performed, but with the penalty of computational cost. In one embodiment, the computation of the distances between pixels in a subregion to neighboring subregions may be performed prior to generating the mapping parameters.
In one embodiment, the 2D nearest neighbor interpolation uses a weight function when performing the mapping parameter interpolation to give gradient magnitude values from closest subregions more “weight” or influence on the gradient magnitude values of pixels in the current subregion over other subregions.
FIG. 12 is a diagram graphically illustrating an example of one embodiment of the weight function used to interpolate the subregion mapping parameters before applying the mapping parameters to pixels (p) in acurrent subregion1200. First, a set of anchor points (wi) are defined that are to be applied to all of the subregions. In one embodiment, a set of i=9 anchor points (w0-w8) are defined, which are located at a center, four corners, and the four midpoints of each side of thesubregion1200. Next, for each pixel (p) in each respective subregion, weights are calculated for each anchor point by calculating the distance (d0-d8) between the current pixel and each of the anchor points representing neighboring subregions. In one embodiment, the weights are inversely proportional to the calculated distances.
Finally, the smoothed mapping parameters (intercept α and slope β) are calculated for each pixel (p) in the respective subregion as a weighted sum of α and β for the anchor points wi:
p=Ei=08f(wi)*∝iand βp=Ei=08f(wi)*βi.
Referring again toFIG. 5, theedge enhancement engine220 applies the mapping parameters calculated for each of the subregions to pixels in the corresponding subregions to generate enhanced gradient magnitude values having equalized luminosity in contrast, and using the enhanced gradient magnitude values to form a wireframe image having enhanced edges of objects (block508). In one embodiment, the smoothed mapping parameters generated for each pixel (p) are applied to the respective pixels in the gradient magnitude image to create the wireframe image (See examples shown inFIGS. 3A and 3B).
The wireframe image is then displayed on a display device to the user, wherein the wireframe image appears as a decluttered line drawing where the enhanced edges have equalized luminosity and contrast compared to the thermal image to reduce the cognitive load of the user (block510). Preferably, the wireframe image is displayed as a video stream in real time on the display device, such asdisplay unit222, and in a line of sight of the user's eye.
In one embodiment, displaying the wireframe image on a display device, further comprises first translating the enhanced gradient magnitude values from a perspective of the TIC that captured the thermal image to a viewing perspective of an eye of the user (block510A). In one embodiment, a transformation operator may perform an affine transformation that is user configurable through a set of controls (e.g., a keyboard). The affine transformation (T) may be implemented as follows:
x′=Tt(x);Tt(x)=Tt-1(x)+δt(x),
where x=a current pixel enhanced gradient magnitude value, and Ttis the current transformation, which is incrementally adjusted from Tt-1(x) by δt(x) until appropriate for the user. Tt(x) is initialized to identity. The transformed gradient magnitude values may then be displayed as the wireframe image.
A method and system for a cognitive load reducing platform having image edge enhancement has been disclosed. The present invention has been described in accordance with the embodiments shown, and there could be variations to the embodiments, and any variations would be within the spirit and scope of the present invention. For example, the exemplary embodiment can be implemented using hardware, software, a computer readable medium containing program instructions, or a combination thereof. Accordingly, many modifications may be made by one of ordinary skill in the art without departing from the spirit and scope of the appended claims.

Claims (21)

We claim:
1. A computer-implemented method for enhancing edges of objects in a thermal image, the method comprising:
receiving, by a processor, the thermal image from a thermal camera and generating, by the processor, a gradient magnitude image comprising a plurality of pixels having associated gradient magnitude values, wherein the processor and thermal camera are worn by a user, the processor executing a software engine to process thermal images into wireframe images in real time, the software engine operational for;
partitioning the gradient magnitude image into subregions;
calculating gradient magnitude statistics for each of the subregions;
calculating mapping parameters for each of the subregions to equalize and smooth a dynamic range of the corresponding gradient magnitude statistics across the subregions, and interpolating the mapping parameters as a function of computed distances to neighboring subregions to generate smoothed mapping parameters to compensate for arbitrary subregion boundaries, wherein interpolating the mapping parameters comprises:
defining a set of anchor points (wi) to be applied to all of the subregions;
for each pixel (p) in a respective subregion, calculating weights for each anchor point by calculating the distance between a current pixel and each of the anchor points; and
calculating the smoothed mapping parameters α=intercept and β=slope for each pixel (p) in the respective subregion as a weighted sum of α and β for the anchor points wi:

αpi=08f(wi)*αiand βpi=08f(wi)*βi;
applying the mapping parameters calculated for each of the subregions to pixels in the corresponding subregions to generate enhanced gradient magnitude values having equalized luminosity and contrast, and using the enhanced gradient magnitude values to form a wireframe image having enhanced edges of objects; and
displaying the wireframe image on a display device to the user, wherein the wireframe image appears as a decluttered line drawing where the enhanced edges have increased luminosity and contrast compared to the thermal image to reduce the cognitive load of the user.
2. The method ofclaim 1, wherein calculating gradient magnitude statistics further comprises: generating for each of the subregions a histogram of gradient magnitude values and cumulative gradient magnitude counts, and upper and lower gradient magnitude percentile values.
3. The method ofclaim 1, wherein calculating the mapping parameters, further comprises: determining the mapping parameters necessary to linearly map gradient magnitude value control points of each subregion to a predetermined reference gradient magnitude representing a desired dynamic range.
4. The method ofclaim 3, wherein calculating the mapping parameters, further comprises: determining an intercept and slope of the reference gradient magnitude and using the intercept and the slope to create the mapping parameters to apply to each pixel in the subregion.
5. The method ofclaim 4, wherein calculating the mapping parameters, further comprises: calculating the mapping parameters as:

x′=f(α,β,x); andf(α,β,x)=βx+α,
where x=a current pixel gradient magnitude value, α=intercept, and β=slope, and x′=equalized pixel value.
6. The method ofclaim 1, wherein displaying the wireframe image on the display device further comprises: translating the enhanced gradient magnitude values from a perspective of a thermal camera that captures the thermal image to a viewing perspective of an eye of the user.
7. A system, comprising:
a memory containing a thermal image from a thermal camera;
a processor coupled to the memory, wherein the thermal camera and the processor are worn by a user; and
a software component executed by the processor that is configured to perform:
receiving a thermal image and generating a gradient magnitude image comprising a plurality of pixels having associated gradient magnitude values;
partitioning the gradient magnitude image into subregions;
calculating gradient magnitude statistics for each of the subregions;
calculating mapping parameters for each of the subregions to equalize and smooth a dynamic range of the corresponding gradient magnitude statistics across the subregions, and interpolating the mapping parameters as a function of computed distances to neighboring subregions to generate smoothed mapping parameters to compensate for arbitrary subregion boundaries, wherein interpolating the mapping parameters comprises:
defining a set of anchor points (wi) to be applied to all of the subregions;
for each pixel (p) in a respective subregion, calculating weights for each anchor point by calculating the distance between a current pixel and each of the anchor points; and
calculating the smoothed mapping parameters α=intercept and β=slope for each pixel (p) in the respective subregion as a weighted sum of for the anchor points wi:

αpi=08f(wi)*αiand βpi=08f(wi)*βi;
applying the mapping parameters calculated for each of the subregions to pixels in the corresponding subregions to generate enhanced gradient magnitude values having equalized luminosity and contrast, and using the enhanced gradient magnitude values to form a wireframe image having enhanced edges of objects; and
displaying the wireframe image on a display device to the user, wherein the wireframe image appears as a decluttered line drawing where the enhanced edges have increased luminosity and contrast compared to the thermal image to reduce the cognitive load of the user.
8. The system ofclaim 7, wherein calculating gradient magnitude statistics further comprises: generating for each of the subregions a histogram of gradient magnitude values and cumulative gradient magnitude counts, and upper and lower gradient magnitude percentile values.
9. The system ofclaim 7, wherein calculating the mapping parameters, further comprises: determining the mapping parameters necessary to linearly map gradient magnitude value control points of each subregion to a predetermined reference gradient magnitude representing a desired dynamic range.
10. The system ofclaim 9, wherein calculating the mapping parameters, further comprises: determining an intercept and slope of the reference gradient magnitude and using the intercept and the slope to create the mapping parameters to apply to each pixel in the subregion.
11. The system ofclaim 10, wherein calculating the mapping parameters, further comprises: calculating the mapping parameters as:

x′=f(α,β,x); andf(α,ρ,x)=βx+α,
where x=a current pixel gradient magnitude value, α=intercept, and β=slope, and x′=equalized pixel value.
12. The system ofclaim 7, wherein displaying the wireframe image on a display device further comprises: translating the enhanced gradient magnitude values from a perspective of a thermal camera that captures the thermal image to a viewing perspective of an eye of the user.
13. A non-transitory computer-readable medium containing program instructions for enhancing edges of objects in a thermal image, wherein the computer readable medium, a thermal camera that captures the thermal image, and a processor for processing the thermal image into a wireframe image in real-time, are worn by a user, the program instructions when executed by the processor for:
receiving the thermal image and generating a gradient magnitude image comprising a plurality of pixels having associated gradient magnitude values;
partitioning the gradient magnitude image into subregions;
calculating gradient magnitude statistics for each of the subregions;
calculating mapping parameters for each of the subregions to equalize and smooth a dynamic range of the corresponding gradient magnitude statistics across the subregions, and interpolating the mapping parameters as a function of computed distances to neighboring subregions to generate smoothed mapping parameters to compensate for arbitrary subregion boundaries, wherein interpolating the mapping parameters comprises:
defining a set of anchor points (wi) to be applied to all of the subregions;
for each pixel (p) in a respective subregion, calculating weights for each anchor point by calculating the distance between a current pixel and each of the anchor points; and
calculating the smoothed mapping parameters α=intercept and β=slope for each pixel (p) in the respective subregion as a weighted sum of α and β for the anchor points wi:

αpi=08f(wi)*αiand βpi=08f(wi)*βi;
applying the mapping parameters calculated for each of the subregions to pixels in the corresponding subregions to generate enhanced gradient magnitude values having equalized luminosity and contrast, and using the enhanced gradient magnitude values to form a wireframe image having enhanced edges of objects; and
displaying the wireframe image on a display device to the user, wherein the wireframe image appears as a decluttered line drawing where the enhanced edges have increased luminosity and contrast compared to the thermal image to reduce the cognitive load of the user.
14. The non-transitory computer-readable medium ofclaim 13, wherein calculating gradient magnitude statistics further comprises: generating for each of the subregions a histogram of gradient magnitude values and cumulative gradient magnitude counts, and upper and lower gradient magnitude percentile values.
15. The non-transitory computer-readable medium ofclaim 13, wherein calculating the mapping parameters, further comprises: determining the mapping parameters necessary to linearly map gradient magnitude value control points of each subregion to a predetermined reference gradient magnitude representing a desired dynamic range.
16. The non-transitory computer-readable medium ofclaim 15, wherein calculating the mapping parameters, further comprises: determining an intercept and slope of the reference gradient magnitude and using the intercept and the slope to create the mapping parameters to apply to each pixel in the subregion.
17. The non-transitory computer-readable medium ofclaim 16, wherein calculating the mapping parameters, further comprises: calculating the mapping parameters as:

x′=f(α,β,x); andf(α,β,x)=βx+α,
where x=a current pixel gradient magnitude value, α=intercept, and β=slope, and x′=equalized pixel value.
18. The non-transitory computer-readable medium ofclaim 13, wherein displaying the wireframe image on a display device further comprises: translating the enhanced gradient magnitude values from a perspective of a thermal camera that captures the thermal image to a viewing perspective of an eye of the user.
19. A method for enhancing edges of objects in a thermal image to improve user vision and navigation in obscure environments, comprising:
computing a gradient magnitude image from the thermal image, the gradient magnitude image comprising a plurality of gradient magnitude values;
partitioning the gradient magnitude image into a plurality of non-overlapping subregions;
for each of the subregions, calculating gradient magnitude statistics from the gradient magnitude values in the respective subregion;
applying predetermined upper and lower bound gradient percentiles and an absolute minimum percentile for the gradient magnitude values across all the subregions;
for each of the subregions, determining upper and lower gradient magnitude values at the upper and lower bound gradient percentiles, respectively;
for each of the subregions, calculating mapping parameters to apply to the pixels in the respective subregion to linearly map the upper and lower gradient magnitude values to a predetermined reference gradient magnitude representing a desired dynamic range applied to all the subregions;
for each of the pixels in each of the subregions, computing a distance to neighboring subregions and interpolating the mapping parameters generated for the respective subregion as a function of the computed distances to generate smoothed mapping parameters;
applying the smoothed mapping parameters to the pixels in the gradient magnitude image to generate a wireframe image; and
displaying the wireframe images in a line of sight of an eye of a user, wherein the wireframe image appears as a decluttered line drawing where edges have increased luminosity and contrast compared to the thermal image.
20. The method ofclaim 19, further comprising: performing the method of enhancing the edges of objects in a platform, comprising:
a thermal camera worn by the user to capture thermal images;
one or more processors in communication with the thermal camera, wherein the one or more processors execute an edge enhancement engine that processes the thermal images in real time into wireframe images; and
a display unit worn in the line of sight of of the user that displays the wireframe images, the display unit comprising one of an augmented reality (AR) display unit, a virtual reality (VR) display unit, or a head-mounted projection display unit.
21. The method ofclaim 20, further comprising: displaying the wireframe image in a command and control interface on a display device to a person of authority.
US16/383,1582018-11-092019-04-12Cognitive load reducing platform having image edge enhancementActiveUS10896492B2 (en)

Priority Applications (2)

Application NumberPriority DateFiling DateTitle
US16/383,158US10896492B2 (en)2018-11-092019-04-12Cognitive load reducing platform having image edge enhancement
US17/113,694US11610292B2 (en)2018-11-092020-12-07Cognitive load reducing platform having image edge enhancement

Applications Claiming Priority (2)

Application NumberPriority DateFiling DateTitle
US201862758457P2018-11-092018-11-09
US16/383,158US10896492B2 (en)2018-11-092019-04-12Cognitive load reducing platform having image edge enhancement

Related Child Applications (1)

Application NumberTitlePriority DateFiling Date
US17/113,694ContinuationUS11610292B2 (en)2018-11-092020-12-07Cognitive load reducing platform having image edge enhancement

Publications (2)

Publication NumberPublication Date
US20200151859A1 US20200151859A1 (en)2020-05-14
US10896492B2true US10896492B2 (en)2021-01-19

Family

ID=70550584

Family Applications (2)

Application NumberTitlePriority DateFiling Date
US16/383,158ActiveUS10896492B2 (en)2018-11-092019-04-12Cognitive load reducing platform having image edge enhancement
US17/113,694ActiveUS11610292B2 (en)2018-11-092020-12-07Cognitive load reducing platform having image edge enhancement

Family Applications After (1)

Application NumberTitlePriority DateFiling Date
US17/113,694ActiveUS11610292B2 (en)2018-11-092020-12-07Cognitive load reducing platform having image edge enhancement

Country Status (4)

CountryLink
US (2)US10896492B2 (en)
EP (1)EP3877899B1 (en)
CA (1)CA3119265A1 (en)
WO (1)WO2020096818A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US11346938B2 (en)2019-03-152022-05-31Msa Technology, LlcSafety device for providing output to an individual associated with a hazardous environment
US11493765B2 (en)*2019-06-142022-11-08Flir Systems AbImage relay systems and methods for wearable apparatuses
WO2023099010A1 (en)2021-12-032023-06-08Telefonaktiebolaget Lm Ericsson (Publ)Augmenting human vision in limited-visibility environments

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN112534467A (en)*2018-02-132021-03-19弗兰克.沃布林Method and apparatus for contrast sensitivity compensation
US11665319B2 (en)*2019-01-152023-05-30Longan Vision Corp.Augmented reality display device, and apparatus comprising same
US11842117B2 (en)*2020-01-312023-12-12Nec CorporationInformation display system and information display method
US20230236017A1 (en)*2020-06-152023-07-273M Innovative Properties CompanyPersonal protective equipment for navigation and map generation within a visually obscured environment
CN116824166B (en)*2023-08-292024-03-08南方电网数字电网研究院股份有限公司Transmission line smoke identification method, device, computer equipment and storage medium
WO2025074274A1 (en)*2023-10-062025-04-103M Innovative Properties CompanyThermal imaging systems and methods

Citations (61)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US5778092A (en)1996-12-201998-07-07Xerox CorporationMethod and apparatus for compressing color or gray scale documents
US6195467B1 (en)1999-03-252001-02-27Image Processing Technologies, Inc.Method and apparatus for sharpening a grayscale image
EP1168033A1 (en)2000-06-192002-01-02Aerospatiale Matra MissilesHelmet mounted viewing device
US6611618B1 (en)1997-11-132003-08-26Schepens Eye Research Institute, Inc.Wide-band image enhancement
US20030190090A1 (en)2002-04-092003-10-09Beeman Edward S.System and method for digital-image enhancement
US6891966B2 (en)1999-08-252005-05-10Eastman Kodak CompanyMethod for forming a depth image from digital image data
US6898559B2 (en)2000-12-082005-05-24Tracker R&D, LlcSystem for dynamic and automatic building mapping
US6909539B2 (en)2000-08-302005-06-21Science Applications International CorporationSystem and method for combining multiple energy bands to improve scene viewing
US20060023966A1 (en)1994-10-272006-02-02Vining David JMethod and system for producing interactive, three-dimensional renderings of selected body organs having hollow lumens to enable simulated movement through the lumen
US7085401B2 (en)2001-10-312006-08-01Infowrap Systems Ltd.Automatic object extraction
US7190832B2 (en)2001-07-172007-03-13Amnis CorporationComputational methods for the segmentation of images of objects from background in a flow imaging instrument
US20070257934A1 (en)2006-05-082007-11-08David DoermannSystem and method for efficient enhancement to enable computer vision on mobile devices
US20080092043A1 (en)2006-10-112008-04-17Trethewey Jim RControls and indicators with on-screen cognitive aids
US7369174B2 (en)*2001-12-272008-05-06Sage Technologies Ltd.Helmet-mounted thermal imaging system
US7377835B2 (en)2002-12-232008-05-27Sti Licensing Corp.Personal multimedia communication system and network for emergency services personnel
US20080146334A1 (en)2006-12-192008-06-19Accenture Global Services GmbhMulti-Player Role-Playing Lifestyle-Rewarded Health Game
US7430303B2 (en)*2002-03-292008-09-30Lockheed Martin CorporationTarget detection method and system
US7460304B1 (en)2007-03-122008-12-02Epstein Jay FThermal vision mask
EP1659890B1 (en)2003-08-192009-01-21Saab AbBreathing mask with display device
US7598856B1 (en)2006-01-312009-10-06Firesite LlcNavigation aid for low-visibility environments
US20110135156A1 (en)*2009-12-042011-06-09Chao-Ho ChenMethod of Locating License Plate of Moving Vehicle
US20110262053A1 (en)2010-04-232011-10-27Flir Systems AbInfrared resolution and contrast enhancement with fusion
US8054170B1 (en)2008-09-302011-11-08Adobe Systems IncorporatedCharacterizing and representing images
US8358307B2 (en)*2008-04-212013-01-22Sharp Kabushiki KaishaImage processing device, display device, image processing method, program, and storage medium
US8463006B2 (en)2007-04-172013-06-11Francine J. ProkoskiSystem and method for using three dimensional infrared imaging to provide detailed anatomical structure maps
US20130307875A1 (en)2012-02-082013-11-21Glen J. AndersonAugmented reality creation using a real scene
US8836793B1 (en)2010-08-132014-09-16Opto-Knowledge Systems, Inc.True color night vision (TCNV) fusion
US20150025917A1 (en)2013-07-152015-01-22Advanced Insurance Products & Services, Inc.System and method for determining an underwriting risk, risk score, or price of insurance using cognitive information
US20150067513A1 (en)2012-05-092015-03-05Apple Inc.Device, Method, and Graphical User Interface for Facilitating User Interaction with Controls in a User Interface
US20150163345A1 (en)2013-12-062015-06-11Digimarc CorporationSmartphone-based methods and systems
US20150202962A1 (en)2014-01-212015-07-23Honda Motor Co., Ltd.System and method for providing an augmented reality vehicle interface
US20150244946A1 (en)2013-11-042015-08-27Sos AgaianMethod and systems for thermal image / video measurements and processing
US20150302654A1 (en)2014-04-222015-10-22Ivan ArbouzovThermal imaging accessory for head-mounted smart device
US9177204B1 (en)2011-09-282015-11-03Rockwell Collins, Inc.Spectrally enhanced vision system for low visibility operations
US20150324989A1 (en)2013-09-032015-11-12Litel InstrumentsMethod & system for high accuracy & reliability registration of multi modal imagery
US20150334315A1 (en)2009-03-022015-11-19Flir Systems, Inc.Infrared imaging enhancement with fusion
US20150339570A1 (en)2014-05-222015-11-26Lee J. SchefflerMethods and systems for neural and cognitive processing
US20150338915A1 (en)2014-05-092015-11-26Eyefluence, Inc.Systems and methods for biomechanically-based eye signals for interacting with real and virtual objects
US20160260261A1 (en)2015-03-062016-09-08Illinois Tool Works Inc.Sensor assisted head mounted displays for welding
US20160295208A1 (en)2015-04-052016-10-06Hema Imaging LlcSystems and Approaches for Repeated Thermal Imaging Determinations
US9498013B2 (en)2014-09-192016-11-22Motorola Solutions, Inc.Wearable safety apparatus for, and method of, displaying heat source characteristics and/or hazards
US20160350906A1 (en)2013-12-192016-12-01Metaio GmbhMethod of tracking a mobile device and method of generating a geometrical model of a real environment using a camera of a mobile device
US20160360382A1 (en)2015-05-272016-12-08Apple Inc.Systems and Methods for Proactively Identifying and Surfacing Relevant Content on a Touch-Sensitive Device
US20170061663A1 (en)2015-08-272017-03-02Fluke CorporationEdge enhancement for thermal-visible combined images and cameras
US20170123211A1 (en)2015-10-072017-05-04Barry LavoieSelf-Contained Breathing Apparatus Face Piece Lens Vision System
US20170192091A1 (en)2016-01-062017-07-06Ford Global Technologies, LlcSystem and method for augmented reality reduced visibility navigation
WO2017130184A1 (en)2016-01-282017-08-03Israel Aerospace Industries Ltd.Systems and methods for detecting imaged clouds
US9728006B2 (en)2009-07-202017-08-08Real Time Companies, LLCComputer-aided system for 360° heads up display of safety/mission critical data
US9729767B2 (en)2013-03-222017-08-08Seiko Epson CorporationInfrared video display eyewear
US20170224990A1 (en)2012-11-262017-08-10Isy GoldwasserApparatuses and methods for neuromodulation
US20170251985A1 (en)2016-02-122017-09-07Newton HowardDetection Of Disease Conditions And Comorbidities
US9875430B1 (en)2016-03-302018-01-23Descartes Labs, Inc.Iterative relabeling using spectral neighborhoods
US9998687B2 (en)2012-09-122018-06-12Bae Systems Information And Electronic Systems Integration Inc.Face mounted extreme environment thermal sensor system
US9995936B1 (en)2016-04-292018-06-12Lockheed Martin CorporationAugmented reality systems having a virtual image overlaying an infrared portion of a live scene
US20180165978A1 (en)2016-12-092018-06-14The Boeing CompanySystem and method for interactive cognitive task assistance
US20180189957A1 (en)2016-12-302018-07-05Dassault SystemesProducing a segmented image of a scene
US10042164B2 (en)2013-03-152018-08-07Hudstar Systems, Inc.Self contained breathing apparatus (SCBA) electronics system
US20180241929A1 (en)2016-06-172018-08-23Huawei Technologies Co., Ltd.Exposure-Related Intensity Transformation
WO2018167771A1 (en)2017-03-152018-09-20Elbit Systems Ltd.Gradual transitioning between two-dimensional and three-dimensional augmented reality images
US10089547B2 (en)2015-03-122018-10-02Elbit Systems Ltd.Detection, characterization and presentation of adverse airborne phenomena
US10417497B1 (en)2018-11-092019-09-17Qwake TechnologiesCognitive load reducing platform for first responders

Family Cites Families (58)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US6088473A (en)*1998-02-232000-07-11Arch Development CorporationMethod and computer readable medium for automated analysis of chest radiograph images using histograms of edge gradients for false positive reduction in lung nodule detection
US7370983B2 (en)2000-03-022008-05-13Donnelly CorporationInterior mirror assembly with display
ITMI20030121A1 (en)2003-01-272004-07-28Giuseppe Donato MODULAR SURVEILLANCE SYSTEM FOR MONITORING OF CRITICAL ENVIRONMENTS.
KR20090120159A (en)*2008-05-192009-11-24삼성전자주식회사 Image Synthesis Device and Image Synthesis Method
US8773537B2 (en)*2008-10-272014-07-08Devcar, LlcNight vision system
US10244190B2 (en)2009-03-022019-03-26Flir Systems, Inc.Compact multi-spectrum imaging with fusion
US9451183B2 (en)2009-03-022016-09-20Flir Systems, Inc.Time spaced infrared image enhancement
US9171361B2 (en)2010-04-232015-10-27Flir Systems AbInfrared resolution and contrast enhancement with fusion
US10091439B2 (en)2009-06-032018-10-02Flir Systems, Inc.Imager with array of multiple infrared imaging modules
US10044946B2 (en)2009-06-032018-08-07Flir Systems AbFacilitating analysis and interpretation of associated visible light and infrared (IR) image information
US8984665B2 (en)2010-02-022015-03-24Wilcox Industries Corp.Helmet mounting system and mounting shoe interface
US9918023B2 (en)2010-04-232018-03-13Flir Systems, Inc.Segmented focal plane array architecture
EP2719167B1 (en)2011-06-102018-08-08Flir Systems, Inc.Low power and small form factor infrared imaging
US9509924B2 (en)2011-06-102016-11-29Flir Systems, Inc.Wearable apparatus with integrated infrared imaging module
US9143703B2 (en)2011-06-102015-09-22Flir Systems, Inc.Infrared camera calibration techniques
US9323325B2 (en)2011-08-302016-04-26Microsoft Technology Licensing, LlcEnhancing an object of interest in a see-through, mixed reality display device
US20160097857A1 (en)2012-02-072016-04-07Michael Cem GokayIntegrated Targeting Device
US10996542B2 (en)2012-12-312021-05-04Flir Systems, Inc.Infrared imaging system shutter assembly with integrated thermister
US9973692B2 (en)2013-10-032018-05-15Flir Systems, Inc.Situational awareness by compressed display of panoramic views
US10424038B2 (en)*2015-03-202019-09-24Digimarc CorporationSignal encoding outside of guard band region surrounding text characters, including varying encoding strength
US10192540B2 (en)2014-02-202019-01-29FLIR Belgium BVBACoordinated route distribution systems and methods
US10338800B2 (en)2014-02-212019-07-02FLIR Belgium BVBAEnhanced pilot display systems and methods
GB2542523B (en)2014-05-302020-12-30Flir SystemsTransmission signal shaping systems and methods
WO2016022525A1 (en)2014-08-052016-02-11Seek Thermal, Inc.Time based offset correction for imaging systems
US9924116B2 (en)2014-08-052018-03-20Seek Thermal, Inc.Time based offset correction for imaging systems and adaptive calibration control
US10230909B2 (en)2014-09-232019-03-12Flir Systems, Inc.Modular split-processing infrared imaging system
US10182195B2 (en)2014-09-232019-01-15Flir Systems, Inc.Protective window for an infrared sensor array
US10073516B2 (en)2014-12-292018-09-11Sony Interactive Entertainment Inc.Methods and systems for user interaction within virtual reality scene using head mounted display
US11505292B2 (en)2014-12-312022-11-22FLIR Belgium BVBAPerimeter ranging sensor systems and methods
WO2016118690A1 (en)2015-01-222016-07-28Siemens AktiengesellschaftSystems and methods for monitoring use of personal protective equipment
LU92673B1 (en)*2015-03-052016-09-06Iee Int Electronics & Eng SaMethod and system for real-time noise removal and image enhancement of high-dynamic range images
WO2016144712A2 (en)2015-03-062016-09-15Flir Systems, Inc.Anomalous pixel detection
US10349887B1 (en)2015-06-142019-07-16Facense Ltd.Blood pressure measuring smartglasses
ES2563098B1 (en)*2015-06-152016-11-29Davantis Technologies Sl IR image enhancement procedure based on scene information for video analysis
EP3350985B1 (en)2015-09-182022-04-20Teledyne Flir, LLCHigh dynamic range radiometric thermal video over low bitrate interface
WO2017099819A2 (en)2015-12-092017-06-15Flir Systems, Inc.Airborne inspection systems and methods
CN108780569B (en)2016-01-082023-05-05特利丹菲力尔有限责任公司System and method for image resolution enhancement
CN108780570B (en)2016-01-162022-12-06特利丹菲力尔有限责任公司 Systems and methods for image super-resolution using iterative collaborative filtering
WO2017222778A1 (en)2016-06-242017-12-28Flir Systems, Inc.Thermal-image based video compression systems and methods
CN109792508B (en)2016-09-292021-07-13菲力尔系统公司 Fail-safe inspection using thermal imaging analysis
CN109923580B (en)2016-11-042023-12-15特利丹菲力尔有限责任公司Dynamic range compression for thermal video
US11113791B2 (en)2017-01-032021-09-07Flir Systems, Inc.Image noise reduction using spectral transforms
US11302041B2 (en)2017-01-132022-04-12Teledyne Flir, LlcHigh visibility overlay systems and methods
US11087512B2 (en)2017-01-132021-08-10Flir Systems, Inc.High visibility overlay systems and methods
US11044422B2 (en)2017-01-132021-06-22Flir Systems, Inc.Pixel readout with partitioned analog-to-digital conversion systems and methods
CN110476416B (en)2017-01-262021-08-17菲力尔系统公司 System and method for infrared imaging in multiple imaging modalities
WO2018152071A1 (en)2017-02-152018-08-23Flir Systems, Inc.Systems and methods for efficient enhanced image filtering by collaborative sharpening in similarity domain
US10598550B2 (en)2017-02-172020-03-24Flir Systems, Inc.Radiometric correction and alignment techniques for thermal imager with non-contact temperature sensor
CN110622211B (en)2017-03-152023-11-14特利丹菲力尔有限责任公司System and method for reducing low frequency non-uniformities in images
WO2018223104A1 (en)2017-06-032018-12-06Flir Systems, Inc.Extensible architecture for surveillance and targeting imaging systems and methods
US10983206B2 (en)2017-11-072021-04-20FLIR Belgium BVBALow cost high precision GNSS systems and methods
US11733699B2 (en)2017-06-162023-08-22FLIR Belgium BVBAUltrasonic perimeter ranging sensor systems and methods
EP3659338B1 (en)2017-07-272022-10-12Teledyne Flir, LLCFlat field correction systems and methods for infrared cameras
KR102525546B1 (en)*2017-11-212023-04-26삼성디스플레이 주식회사Image processing method and image processor performing the same
CN111788602B (en)2017-12-292024-05-28泰立戴恩菲力尔有限责任公司Point cloud denoising system and method
US10962420B2 (en)2017-12-292021-03-30Flir Systems, Inc.Pulse detection and imaging systems and methods
US11890494B2 (en)2018-11-092024-02-06Qwake Technologies, Inc.Retrofittable mask mount system for cognitive load reducing platform
US11988513B2 (en)2019-09-162024-05-21FLIR Belgium BVBAImaging for navigation systems and methods

Patent Citations (61)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20060023966A1 (en)1994-10-272006-02-02Vining David JMethod and system for producing interactive, three-dimensional renderings of selected body organs having hollow lumens to enable simulated movement through the lumen
US5778092A (en)1996-12-201998-07-07Xerox CorporationMethod and apparatus for compressing color or gray scale documents
US6611618B1 (en)1997-11-132003-08-26Schepens Eye Research Institute, Inc.Wide-band image enhancement
US6195467B1 (en)1999-03-252001-02-27Image Processing Technologies, Inc.Method and apparatus for sharpening a grayscale image
US6891966B2 (en)1999-08-252005-05-10Eastman Kodak CompanyMethod for forming a depth image from digital image data
EP1168033A1 (en)2000-06-192002-01-02Aerospatiale Matra MissilesHelmet mounted viewing device
US6909539B2 (en)2000-08-302005-06-21Science Applications International CorporationSystem and method for combining multiple energy bands to improve scene viewing
US6898559B2 (en)2000-12-082005-05-24Tracker R&D, LlcSystem for dynamic and automatic building mapping
US7190832B2 (en)2001-07-172007-03-13Amnis CorporationComputational methods for the segmentation of images of objects from background in a flow imaging instrument
US7085401B2 (en)2001-10-312006-08-01Infowrap Systems Ltd.Automatic object extraction
US7369174B2 (en)*2001-12-272008-05-06Sage Technologies Ltd.Helmet-mounted thermal imaging system
US7430303B2 (en)*2002-03-292008-09-30Lockheed Martin CorporationTarget detection method and system
US20030190090A1 (en)2002-04-092003-10-09Beeman Edward S.System and method for digital-image enhancement
US7377835B2 (en)2002-12-232008-05-27Sti Licensing Corp.Personal multimedia communication system and network for emergency services personnel
EP1659890B1 (en)2003-08-192009-01-21Saab AbBreathing mask with display device
US7598856B1 (en)2006-01-312009-10-06Firesite LlcNavigation aid for low-visibility environments
US20070257934A1 (en)2006-05-082007-11-08David DoermannSystem and method for efficient enhancement to enable computer vision on mobile devices
US20080092043A1 (en)2006-10-112008-04-17Trethewey Jim RControls and indicators with on-screen cognitive aids
US20080146334A1 (en)2006-12-192008-06-19Accenture Global Services GmbhMulti-Player Role-Playing Lifestyle-Rewarded Health Game
US7460304B1 (en)2007-03-122008-12-02Epstein Jay FThermal vision mask
US8463006B2 (en)2007-04-172013-06-11Francine J. ProkoskiSystem and method for using three dimensional infrared imaging to provide detailed anatomical structure maps
US8358307B2 (en)*2008-04-212013-01-22Sharp Kabushiki KaishaImage processing device, display device, image processing method, program, and storage medium
US8054170B1 (en)2008-09-302011-11-08Adobe Systems IncorporatedCharacterizing and representing images
US20150334315A1 (en)2009-03-022015-11-19Flir Systems, Inc.Infrared imaging enhancement with fusion
US9728006B2 (en)2009-07-202017-08-08Real Time Companies, LLCComputer-aided system for 360° heads up display of safety/mission critical data
US20110135156A1 (en)*2009-12-042011-06-09Chao-Ho ChenMethod of Locating License Plate of Moving Vehicle
US20110262053A1 (en)2010-04-232011-10-27Flir Systems AbInfrared resolution and contrast enhancement with fusion
US8836793B1 (en)2010-08-132014-09-16Opto-Knowledge Systems, Inc.True color night vision (TCNV) fusion
US9177204B1 (en)2011-09-282015-11-03Rockwell Collins, Inc.Spectrally enhanced vision system for low visibility operations
US20130307875A1 (en)2012-02-082013-11-21Glen J. AndersonAugmented reality creation using a real scene
US20150067513A1 (en)2012-05-092015-03-05Apple Inc.Device, Method, and Graphical User Interface for Facilitating User Interaction with Controls in a User Interface
US9998687B2 (en)2012-09-122018-06-12Bae Systems Information And Electronic Systems Integration Inc.Face mounted extreme environment thermal sensor system
US20170224990A1 (en)2012-11-262017-08-10Isy GoldwasserApparatuses and methods for neuromodulation
US10042164B2 (en)2013-03-152018-08-07Hudstar Systems, Inc.Self contained breathing apparatus (SCBA) electronics system
US9729767B2 (en)2013-03-222017-08-08Seiko Epson CorporationInfrared video display eyewear
US20150025917A1 (en)2013-07-152015-01-22Advanced Insurance Products & Services, Inc.System and method for determining an underwriting risk, risk score, or price of insurance using cognitive information
US20150324989A1 (en)2013-09-032015-11-12Litel InstrumentsMethod & system for high accuracy & reliability registration of multi modal imagery
US20150244946A1 (en)2013-11-042015-08-27Sos AgaianMethod and systems for thermal image / video measurements and processing
US20150163345A1 (en)2013-12-062015-06-11Digimarc CorporationSmartphone-based methods and systems
US20160350906A1 (en)2013-12-192016-12-01Metaio GmbhMethod of tracking a mobile device and method of generating a geometrical model of a real environment using a camera of a mobile device
US20150202962A1 (en)2014-01-212015-07-23Honda Motor Co., Ltd.System and method for providing an augmented reality vehicle interface
US20150302654A1 (en)2014-04-222015-10-22Ivan ArbouzovThermal imaging accessory for head-mounted smart device
US20150338915A1 (en)2014-05-092015-11-26Eyefluence, Inc.Systems and methods for biomechanically-based eye signals for interacting with real and virtual objects
US20150339570A1 (en)2014-05-222015-11-26Lee J. SchefflerMethods and systems for neural and cognitive processing
US9498013B2 (en)2014-09-192016-11-22Motorola Solutions, Inc.Wearable safety apparatus for, and method of, displaying heat source characteristics and/or hazards
US20160260261A1 (en)2015-03-062016-09-08Illinois Tool Works Inc.Sensor assisted head mounted displays for welding
US10089547B2 (en)2015-03-122018-10-02Elbit Systems Ltd.Detection, characterization and presentation of adverse airborne phenomena
US20160295208A1 (en)2015-04-052016-10-06Hema Imaging LlcSystems and Approaches for Repeated Thermal Imaging Determinations
US20160360382A1 (en)2015-05-272016-12-08Apple Inc.Systems and Methods for Proactively Identifying and Surfacing Relevant Content on a Touch-Sensitive Device
US20170061663A1 (en)2015-08-272017-03-02Fluke CorporationEdge enhancement for thermal-visible combined images and cameras
US20170123211A1 (en)2015-10-072017-05-04Barry LavoieSelf-Contained Breathing Apparatus Face Piece Lens Vision System
US20170192091A1 (en)2016-01-062017-07-06Ford Global Technologies, LlcSystem and method for augmented reality reduced visibility navigation
WO2017130184A1 (en)2016-01-282017-08-03Israel Aerospace Industries Ltd.Systems and methods for detecting imaged clouds
US20170251985A1 (en)2016-02-122017-09-07Newton HowardDetection Of Disease Conditions And Comorbidities
US9875430B1 (en)2016-03-302018-01-23Descartes Labs, Inc.Iterative relabeling using spectral neighborhoods
US9995936B1 (en)2016-04-292018-06-12Lockheed Martin CorporationAugmented reality systems having a virtual image overlaying an infrared portion of a live scene
US20180241929A1 (en)2016-06-172018-08-23Huawei Technologies Co., Ltd.Exposure-Related Intensity Transformation
US20180165978A1 (en)2016-12-092018-06-14The Boeing CompanySystem and method for interactive cognitive task assistance
US20180189957A1 (en)2016-12-302018-07-05Dassault SystemesProducing a segmented image of a scene
WO2018167771A1 (en)2017-03-152018-09-20Elbit Systems Ltd.Gradual transitioning between two-dimensional and three-dimensional augmented reality images
US10417497B1 (en)2018-11-092019-09-17Qwake TechnologiesCognitive load reducing platform for first responders

Non-Patent Citations (16)

* Cited by examiner, † Cited by third party
Title
Bretschneider et al., "Head Mounted Displays for Fire Fighters" 3rd International Forum on Applied Wearable Computing 2006; 15 pages.
Chen, "Reducing Cognitive Load in Mobile Learning: Activity-centered Perspectives" Published in International Conference on Networking and Digital Society; DOI: 10.1109/ICNDS.2010.5479459; pp. 504-507 (2010).
Fan, et al., "Reducing Cognitive Overload by Meta-Learning Assisted Algorithm Selection" Published in 5th IEEE International Conference on Cognitive Informatics; DOI: 10.1109/COGINF.2006.365686; pp. 120-125 (2006).
Georgy Gimel'farb, "Part 3: Image Processing Digital Images and Intensity Histogram", Jun. 4, 2014 (Year: 2014).*
Gimel'Farb Part 3: Image Processing, Digital Images and Intensity Histograms; COMPSCI 373 Computer Graphics and Image Processing; University of Auckland, Auckland, NZ; Date unknown; 57 pages.
Haciomeroglu, "C-thru smoke diving helmet" Jan. 8, 2013, 14 pages; coroflot.com <https://www.coroflot.com/OmerHaciomeroglu/C-Thru-smoke-Diving-Helmet>.
Haciomeroglu, "C-thru smoke diving helmet" Jan. 8, 2013; 15 pages; behance.com <http://ww.behance.net/gallery/6579685/C-Thru-Smoke-Diving-Helmet>.
Khan et al., "Tracking Visual and Infrared Objects using Joint Riemannian Manifold Appearance and Affine Shaping Modeling" Dept. of Signals and Systems, Chalmers University of Technology, Gothenburg, 41296, Sweden; IEEE International Conference on Computer Vision Workshop (2011); pp. 1847-1854.
Mckinzie, "Fire Engineering: The Future of Artificial Intelligence in Firefighting" Oct. 25, 2018; available at <https://www.fireengineering.com/articles/2018/10/artificial-intelligence-firefighting.html>; 16 pages.
Omer Haciomeroglu, "C-Thru Smoke Diving Helmet", Apr. 29, 2014, (Year: 2014).*
Patent Cooperation Treaty: International Search Report and Written Opinion for PCT/US2019/058635 dated Jan. 15, 2020; 14 pages.
Reis, et al., "Towards Reducing Cognitive Load and Enhancing Usability Through a Reduced Graphical User Interface for a Dynamic Geometry System: An Experimental Study" Proceedings-2012 IEEE International Symposium on Multimedia, ISM 2012. 445-450. 10.1109/ISM.2012.91; pp. 445-450 (2012).
Reis, et al., "Towards Reducing Cognitive Load and Enhancing Usability Through a Reduced Graphical User Interface for a Dynamic Geometry System: An Experimental Study" Proceedings—2012 IEEE International Symposium on Multimedia, ISM 2012. 445-450. 10.1109/ISM.2012.91; pp. 445-450 (2012).
Thomsen-Florenus, "Thermal Vision System" Berlin, Germany; Dec. 2017; 7 pages.
Wu et al., "Contract-Accumulated Histogram Equalization for Image Enhancement", IEEE SigPort, 2017. [Online]. Available at <http://sigport.org/1837>.
Wu, "Feature-based Image Segmentation, Texture Synthesis and Hierarchical Visual Data Approximation" University of Illinois at Urbana-Champaign, Apr. 2006; 61 pages.

Cited By (4)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US11346938B2 (en)2019-03-152022-05-31Msa Technology, LlcSafety device for providing output to an individual associated with a hazardous environment
US12169234B2 (en)2019-03-152024-12-17Msa Technology, LlcSafety device for providing output to an individual associated with a hazardous environment
US11493765B2 (en)*2019-06-142022-11-08Flir Systems AbImage relay systems and methods for wearable apparatuses
WO2023099010A1 (en)2021-12-032023-06-08Telefonaktiebolaget Lm Ericsson (Publ)Augmenting human vision in limited-visibility environments

Also Published As

Publication numberPublication date
EP3877899A4 (en)2023-01-04
EP3877899B1 (en)2025-08-13
CA3119265A1 (en)2020-05-14
US11610292B2 (en)2023-03-21
US20210118107A1 (en)2021-04-22
EP3877899C0 (en)2025-08-13
WO2020096818A1 (en)2020-05-14
US20200151859A1 (en)2020-05-14
EP3877899A1 (en)2021-09-15

Similar Documents

PublicationPublication DateTitle
US11610292B2 (en)Cognitive load reducing platform having image edge enhancement
US11354895B2 (en)Cognitive load reducing platform for first responders
US20240165434A1 (en)Retrofittable mask mount system for cognitive load reducing platform
US11474348B2 (en)Method and device for eye tracking using event camera data
US11915376B2 (en)Wearable assisted perception module for navigation and communication in hazardous environments
US10559065B2 (en)Information processing apparatus and information processing method
CN110998666B (en)Information processing device, information processing method, and program
US10474411B2 (en)System and method for alerting VR headset user to real-world objects
US11137600B2 (en)Display device, display control method, and display system
US11726320B2 (en)Information processing apparatus, information processing method, and program
CN111856751B (en)Head mounted display with low light operation
KR101331055B1 (en)Visual aid system based on the analysis of visual attention and visual aiding method for using the analysis of visual attention
WO2021019360A1 (en)Fatigue evaluation system and fatigue evaluation device
Karim et al.A Novel Eye-Tracking Device Designed with a Head Gesture Control Module.
EP4441698B1 (en)Augmenting human vision in limited-visibility environments
US12136142B2 (en)Information processing apparatus, head-mounted display apparatus, information processing method, and non-transitory computer readable medium
US20250186260A1 (en)Protective eyewear for high intensity sources
US11181973B2 (en)Techniques related to configuring a display device

Legal Events

DateCodeTitleDescription
FEPPFee payment procedure

Free format text:ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: MICROENTITY

FEPPFee payment procedure

Free format text:ENTITY STATUS SET TO MICRO (ORIGINAL EVENT CODE: MICR); ENTITY STATUS OF PATENT OWNER: MICROENTITY

STPPInformation on status: patent application and granting procedure in general

Free format text:RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

ASAssignment

Owner name:QWAKE TECHNOLOGIES, LLC, CALIFORNIA

Free format text:CORRECTIVE ASSIGNMENT TO CORRECT THE ASSIGNEE NAME PREVIOUSLY RECORDED AT REEL: 048874 FRAME: 0400. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT;ASSIGNORS:LONG, JOHN DAVIS, II;HACIOMEROGLU, OMER;REEL/FRAME:054342/0851

Effective date:20200828

STPPInformation on status: patent application and granting procedure in general

Free format text:PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCFInformation on status: patent grant

Free format text:PATENTED CASE

ASAssignment

Owner name:QWAKE TECHNOLOGIES, INC., CALIFORNIA

Free format text:ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:QWAKE TECHNOLOGIES, LLC;REEL/FRAME:059954/0061

Effective date:20161130

MAFPMaintenance fee payment

Free format text:PAYMENT OF MAINTENANCE FEE, 4TH YEAR, MICRO ENTITY (ORIGINAL EVENT CODE: M3551); ENTITY STATUS OF PATENT OWNER: MICROENTITY

Year of fee payment:4


[8]ページ先頭

©2009-2025 Movatter.jp