Movatterモバイル変換


[0]ホーム

URL:


US20140320629A1 - Haptically-Enabled Co-Robotics for Underwater Tasks - Google Patents

Haptically-Enabled Co-Robotics for Underwater Tasks
Download PDF

Info

Publication number
US20140320629A1
US20140320629A1US14/164,115US201414164115AUS2014320629A1US 20140320629 A1US20140320629 A1US 20140320629A1US 201414164115 AUS201414164115 AUS 201414164115AUS 2014320629 A1US2014320629 A1US 2014320629A1
Authority
US
United States
Prior art keywords
light
virtual
image
camera
depth
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US14/164,115
Inventor
Howard Jay Chizeck
Fredrik Ryden
Andrew Stewart
Wei-Chih Wang
Payman Arabshahi
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Washington Center for Commercialization
Original Assignee
University of Washington Center for Commercialization
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Washington Center for CommercializationfiledCriticalUniversity of Washington Center for Commercialization
Priority to US14/164,115priorityCriticalpatent/US20140320629A1/en
Assigned to UNIVERSITY OF WASHINGTON THROUGH ITS CENTER FOR COMMERCIALIZATIONreassignmentUNIVERSITY OF WASHINGTON THROUGH ITS CENTER FOR COMMERCIALIZATIONASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS).Assignors: CHIZECK, HOWARD JAY, ARABSHAHI, PAYMAN, WANG, WEI-CHIH, RYDEN, FREDRIK, STEWART, ANDREW
Publication of US20140320629A1publicationCriticalpatent/US20140320629A1/en
Assigned to NATIONAL SCIENCE FOUNDATIONreassignmentNATIONAL SCIENCE FOUNDATIONCONFIRMATORY LICENSE (SEE DOCUMENT FOR DETAILS).Assignors: UNIVERSITY OF WASHINGTON
Abandonedlegal-statusCriticalCurrent

Links

Images

Classifications

Definitions

Landscapes

Abstract

Apparatus configured for operation in an underwater environment are provided. A device includes a camera and a transceiver. The camera can capture, within a predetermined interval of time, first light within a first frequency range of light and second light within a second frequency range of light. The camera can generate a first image based on the first light and a second image based on the second light. The first frequency range of light can differ from the second frequency range of light; for example, the first frequency range can include 420-450 nanometers (nm) and the second frequency range can include 830 nm. The transceiver can send the images and receive commands based on haptic feedback generated based on the first image and the second image. For example, an actuator can be configured to be controlled by one or more commands.

Description

    RELATED APPLICATIONS
  • The present application claims priority to U.S. Provisional Patent Application No. 61/756,132 entitled “Methods and Systems for Six Degree-of-Freedom Haptic Interaction with Streaming Point Clouds”, filed Jan. 24, 2013, U.S. Provisional Patent Application No. 61/764,908 entitled “Methods for Underwater Haptic Rendering Using Nontact Sensors”, filed Feb. 14, 2013, and U.S. Provisional Patent Application No. 61/764,921 entitled “Virtual Fixtures for Subsea Technology”, filed Feb. 14, 2013, all of which are entirely incorporated by reference herein for all purposes.
  • STATEMENT OF GOVERNMENT RIGHTS
  • This invention was made with government support under grant no. 0930930, awarded by the National Science Foundation and with support under grant MRSEED01-006 “Haptically-Enabled Co-Robotics for Remediation of Military Munitions Underwater” for the Strategic Environmental Research and Development Program (SERDP). The United States Government has certain rights in the invention.
  • BACKGROUND OF THE INVENTION
  • Unless otherwise indicated herein, the materials described in this section are not prior art to the claims in this application and are not admitted to be prior art by inclusion in this section.
  • Haptic rendering is the translation of forces in a virtual environment to a physical device that can provide touch-based, a.k.a. haptic, feedback to a user of the haptic rendering device. Both impedance type and admittance type haptic rendering devices are available.
  • To provide haptic feedback about the virtual environment, objects in the virtual environment are often represented as a collection of polygons, such as triangles, that can be operated upon using a haptic rendering device. The haptic rendering device can be controlled using a “Haptic Interaction Point” (HIP) in the virtual environment, which performs a similar function for the haptic rendering device as a mouse pointer does for a computer mouse. Ideally, the HIP should not be able to penetrate virtual environment objects.
  • FIGS. 1A-1C depict ascenario100 that illustrates a prior art technique of utilizing proxy130 to control interactions between HIP110 andpolygon120 in a virtual environment. Inscenario100, HIP110 starts as the position shown asHIP110ainFIG. 1A, moves throughposition HIP110bshown inFIG. 1B, and ends atposition HIP110cshown inFIG. 1C. In this technique, HIP110 and proxy130 are connected by a simulated spring not shown in the Figures.
  • InFIG. 1A, proxy130, shown atposition130a, is in “free motion”; e.g.,proxy130ais not touchingpolygon120. InFIG. 1B, proxy130, shown atposition130b, is “in contact” withpolygon120. Inscenario100, while HIP110 continues to move down fromposition110bintopolygon120, proxy130 is not permitted to enter intopolygon120. InFIG. 1C, proxy130, shown atposition130c, is still in contact with a surface ofpolygon120 after HIP110 has moved toposition110cinside ofpolygon120. As the distance increases between HIP110 and proxy130, the simulated spring exerts a proportionally larger force to draw HIP110 closer to proxy130.FIG. 1C shows the simulated spring force asforce140 exerted on HIP110. As shown inFIG. 1C,force140 is exerted in the direction of a normal of the surface in contact with the proxy130; e.g., a hypotenuse122 ofpolygon120.
  • FIG. 2 shows anexample coordinate system200 specifying six degrees of freedom fortool210.Coordinate system200 can be used for other entities as well, such as HIP. A position oftool210 can be specified as a point (x, y, z) in three-dimensional space specified usingcoordinate system200. For example, the point can defined in terms of three axes, such as an X axis, a Y axis, and a Z axis. Then, the position oftool210 can be specified as a (x, y, z) coordinate, with x, y, and z respectively specifying an X-axis coordinate, a Y-axis coordinate, and a Z-axis coordinate. If only position is taken into account,tool210 can be said to have three positional degrees of freedom, as each of x, y, and z can be specified to positiontool210.
  • Tool210 can be rotated about each of the X axis, Y axis, and Z axis, as shown inFIG. 2. If only rotations are taken into account,tool210 can be said to have three rotational degrees of freedom, as rotations about each of x, y, and z can be specified to rotate (or orient)tool210. Taking both positional and rotational degrees of freedom into account,tool210 can have up to 6 degrees of freedom, as listed onFIG. 2. These six degrees of freedom include respective degrees of freedom for selecting an X coordinate, a Y coordinate, a Z coordinate, an X rotation, a Y rotation, and a Z rotation.
  • Techniques for six degree-of-freedom haptic rendering in virtual environments consisting of polygons and/or voxels (volume pixels) have been specified. These efforts are typically divided into direct rendering- and virtual coupling methods where the latter can further be subdivided into penalty-, impulse- and constraint-based methods. The simplest 6-DOF haptic rendering method is the direct method, where the virtual tool perfectly matches the configuration of the haptic rendering device. The force sent to user is directly based on the amount of penetration in the virtual environment. Unfortunately the direct method suffers from problems with “pop through”. Pop through is an artifact that arises when the rendering algorithm erroneously penetrates a thin surface.
  • In virtual coupling methods, a virtual coupling, or connection, between the haptic rendering device and the virtual tool is utilized. In this method, the force on the haptic rendering device is simply calculated as a spring between the virtual tool, referred to also as ‘god-object’, ‘proxy’ or ‘IHIP’, and the configuration of the haptic rendering device. 6-DOF rendering methods using virtual couplings rely on rigid body simulations, since the virtual tool has to be simulated as a 6-DOF object, as compared to 3-DOF rendering where the rotational component can be ignored. In penalty-based methods, the configuration (position and rotation) of the virtual tool is calculated using penalty forces based on the tool's penetration depth into objects, similar to how penalty-costs are used in traditional optimization. These penalty forces are then integrated to produce the motion of the virtual tool. This method results in a virtual tool that actually penetrates objects in the environment. Fortunately this penetration is typically very small.
  • For impulse-based dynamics methods, a virtual object is moved by a series of impulses upon contact/collision (rather than forces based on penetration depth). In constraint-based methods, the virtual tool moves into contact with the environment but (ideally) never violates constraints imposed by the environment.
  • Other environments can be explored by robots, such as undersea, outer space, and hazardous environments. In some of these environments, robots can be controlled by human operators receiving video and/or audio information from the robot.
  • SUMMARY
  • In one aspect, a method is provided. A computing device receives first depth data about an environment. The computing device generates a first plurality of points from the first depth data. The computing device determines a virtual tool, where the virtual tool is specified in terms of a translation component for the virtual tool and a rotation component for the virtual tool. The computing device determines a first force vector between the virtual tool and the first plurality of points. The computing device sends a first indication of haptic feedback based on the first force vector.
  • In another aspect, an article of manufacture is provided. The article of manufacture includes a physical and/or non-transitory computer-readable storage medium storing instructions that, upon execution by a processor of a computing device, cause the computing device to perform functions including: receiving first depth data about an environment; generating a first plurality of points from the first depth data; determining a virtual tool specified in terms of a translation component for the virtual tool and a rotation component for the virtual tool; determining a first force vector between the virtual tool and the first plurality of points; and sending a first indication of haptic feedback based on the first force vector.
  • In yet another aspect, a computing device is provided. The computing device includes a processor and data storage. The data storage stores instructions that, upon execution by the processor, cause the computing device to perform functions including: receiving first depth data about an environment; generating a first plurality of points from the first depth data; determining a virtual tool specified in terms of a translation component for the virtual tool and a rotation component for the virtual tool; determining a first force vector between the virtual tool and the first plurality of points; and sending a first indication of haptic feedback based on the first force vector.
  • In one aspect, a method is provided. A computing device receives first depth data about an environment. The computing device generates a first plurality of points from the first depth data. The computing device determines a haptic interface point (HIP). The computing device defines a virtual fixture for the environment. The computing device determines a first force vector between the HIP and the first plurality of points. The first force vector is based on the virtual fixture. The computing device sends a first indication of haptic feedback based on the first force vector.
  • In another aspect, an article of manufacture is provided. The article of manufacture includes a physical computer-readable storage medium. The physical computer-readable storage medium stores instructions that, upon execution by a processor of the article of manufacture, cause the article of manufacture to perform functions. The functions include: receiving first depth data about an environment; generating a first plurality of points from the first depth data; determining a HIP; defining a virtual fixture for the environment; determining a first force vector between the HIP and the first plurality of points, where the first force vector is based on the virtual fixture; and sending a first indication of haptic feedback based on the first force vector.
  • In yet another aspect, a computing device is provided. The computing device includes a processor and data storage. The data storage stores instructions that, upon execution by the processor, cause the computing device to perform functions. The functions include: receiving first depth data about an environment; generating a first plurality of points from the first depth data; determining a HIP; defining a virtual fixture for the environment; determining a first force vector between the HIP and the first plurality of points, where the first force vector is based on the virtual fixture; and sending a first indication of haptic feedback based on the first force vector.
  • In one aspect, a device configured for operation in an underwater environment is provided. The device includes a camera. The camera is configured to capture, within a predetermined interval of time, first light within a first frequency range of light in the underwater environment and second light within a second frequency range of light in the underwater environment. The camera is configured to generate a first image based on the first light and a second image based on the second light, where the first frequency range of light differs from the second frequency range of light. The camera includes a communication interface. The communication interface is configured at least to send at least the first image and the second image and to receive one or more commands based on haptic feedback with the haptic feedback generated based on the first image and the second image.
  • In another aspect, a method is provided. A camera captures, within a predetermined interval of time, first light within a first frequency range of light in an underwater environment and second light within a second frequency range of light in the underwater environment. The camera generates a first image based on the first light and a second image based on the second light. The first frequency range of light differs from the second frequency range of light. The camera sends at least the first image and the second image from the camera
  • One advantage of this application is the ability to specify and providing haptic feedback for application using a virtual tool specified using six degrees of freedom interacting with objects in an environment specified using point data with depth information, such data for a plurality of points. For example, three degrees of freedom can specify a position of the virtual tool in a three-dimensional space, and three degrees of freedom can specify an orientation, or rotation, of the virtual tool with respect to the three-dimensional space. An example application for using the virtual tool is a task performed by a user-controlled robot, where haptic feedback is given to the user during remote control of the robot.
  • Another advantage of this application is that haptic feedback is generated at rapid rates based on depth data, such as a stream of frames with depth information. As disclosed herein, the use of a stream of frames with depth information permits haptic rendering, or providing haptic feedback, at a haptic rendering rate faster than a frame-reception rate. In some embodiments, a herein-described haptic rendering system can have a haptic rendering rate of at least 1000 Hz.
  • Another advantage of this application is that virtual fixtures can be defined based on objects recognized in the environment. That is, as an object changes within the environment, a virtual fixture associated with the object can dynamically adjust to the changes of the object. Another advantage is that virtual fixtures can be dynamically changed based on status of operations, such tasks listed and organized on a task list. One or more virtual fixtures can be associated with task(s) on the task list. Virtual fixtures associated with a task can change based on the status of the task. Thus, the virtual fixtures can change throughout execution of a task, and so guide completion of the task. Further, a level of automation can be specified that controls the virtual fixtures, and so allows for full automation, partial automation, or no automation (manual control) for completing the task.
  • Another advantage of this application is providing a camera that captures images and depth information underwater. The images can be captured in one range of frequencies, such as within a blue-green range of visible light frequencies. The depth information can be captured in a different range of frequencies, such as in a near-infrared range of frequencies. The visible light can be captured and a visible light image generated. At the same time, NIR radiation can be captured and an image with depth information generated. The visible light image and the image with depth information can be sent from the camera in succession, so that both visible-light information and depth information for a same time can be provided. The visible-light information and depth information can be used to generate haptic feedback for operators controlling devices to perform underwater tasks.
  • Providing haptic feedback from underwater sensors to an operator at the surface has the potential to transform subsea manipulation capability. Virtual fixtures will allow manipulators to precisely maneuver in sensitive environments where contact should not be made with surrounding objects or structures. Biological studies of animal colonies around hydrothermal vents could benefit greatly from such a capability—allowing scientists to carefully gather data with unprecedented resolution and proximity to sensitive organisms. Common tasks such as connector mating between instruments can be carried out very efficiently by creating guidance fixture near male and female connectors. Thus, time, expense, and equipment can be saved, and the environment preserved using the herein-described haptic feedback techniques to perform underwater tasks.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Various examples of particular embodiments are described herein with reference to the following drawings, wherein like numerals denote like entities, in which:
  • FIGS. 1A-1C depict a scenario illustrating a prior art technique of controlling interactions between a Haptic Interaction Point (HIP) and a polygon in a virtual environment;
  • FIG. 2 shows an example coordinate system specifying six degrees of freedom for a tool;
  • FIG. 3A is a diagram of a haptic rendering environment, in accordance with an example embodiment;
  • FIG. 3B shows an image based on captured depth data and the same image annotated with normal vectors, in accordance with an example embodiment;
  • FIG. 4A depicts a scenario of capturing depth data in an environment, in accordance with an example embodiment;
  • FIG. 4B depicts another scenario of capturing depth data in the environment shown inFIG. 4A, in accordance with an example embodiment;
  • FIG. 4C depicts a virtual tool and example points of depth data captured from an environment, in accordance with an example embodiment;
  • FIG. 5A depicts points derived from depth data and a forbidden region, in accordance with an example embodiment.
  • FIG. 5B depicts a virtual tool in accordance with an example embodiment;
  • FIG. 5C shows an example virtual tool whose movement toward a desired position is impeded by a forbidden region, in accordance with an example embodiment;
  • FIG. 5D shows an example virtual tool whose movement is in collision with a forbidden region, in accordance with an example embodiment;
  • FIG. 6 is an example gaming scenario with haptic rendering, in accordance with an example embodiment;
  • FIG. 7 is an example scenario for a haptic rendering session in a remote environment, in accordance with an example embodiment;
  • FIG. 8 is an example scenario for collaborative haptic rendering sessions in a remote environment, in accordance with an example embodiment;
  • FIG. 9 depicts a construction scenario, in accordance with an example embodiment.
  • FIG. 10 shows an underwater robot, in accordance with an example embodiment; and
  • FIG. 11A is a block diagram of a computing environment, in accordance with an example embodiment;
  • FIG. 11B is a block diagram of an example computing device, in accordance with an example embodiment;
  • FIG. 12 is a flowchart of a method, in accordance with an example embodiment;
  • FIG. 13 is a block diagram of an architecture for rendering adaptive virtual fixtures, in accordance with an example embodiment;
  • FIG. 14 is a block diagram of a pipeline for parallel processing and generation of virtual fixtures, in accordance with an example embodiment;
  • FIG. 15 depicts an environment with a robot arm and multiple virtual fixtures, in accordance with an example embodiment;
  • FIG. 16 is a flowchart of another method, in accordance with an example embodiment;
  • FIG. 17 is a block diagram of a system for a proximate operator to perform tasks in a remote physical environment, in accordance with an example embodiment;
  • FIG. 18A is a block diagram of a camera configured to provide images for underwater haptic rendering, in accordance with an example embodiment;
  • FIG. 18B is an exploded view of a camera configured to provide images for underwater haptic rendering, in accordance with an example embodiment;
  • FIG. 18C depicts a sensor system, in accordance with an example embodiment;
  • FIG. 19 is a near-infrared image of an object in clear water, in accordance with an example embodiment;
  • FIG. 20A is a near-infrared image of an object in murky water, in accordance with an example embodiment;
  • FIG. 20B is an image representing depth data related to the image ofFIG. 20A, in accordance with an example embodiment;
  • FIG. 21 shows a graph of light absorption in water versus light wavelength and a graph of light absorption in water versus light wavelength over a range of water temperatures, in accordance with an example embodiment; and
  • FIG. 22 is a block diagram of a time-multiplexed system of near-infrared cameras configured to capture images of an environment, in accordance with an example embodiment;
  • FIG. 23 is a block diagram of a frequency-multiplexed system of near-infrared cameras configured to capture images of an environment, in accordance with an example embodiment;
  • FIG. 24A shows a structure of a metal sensor, in accordance with an example embodiment;
  • FIG. 24B is a diagram of an interferometer-based metal detector, in accordance with an example embodiment;
  • FIG. 24C is a diagram of another interferometer-based metal detector, in accordance with an example embodiment;
  • FIGS. 25A,25B, and25C are each a diagram of a metal detection system that includes multiple metal detectors, in accordance with an example embodiment;
  • FIG. 26A depicts a sensor system using the metal detection system ofFIG. 25A, in accordance with an example embodiment;
  • FIG. 26B depicts another sensor system using the metal detection system ofFIG. 25A, in accordance with an example embodiment;
  • FIG. 27A is an image of example metal objects, in accordance with an example embodiment;
  • FIG. 27B is a graph of an output signal from an interferometer-based metal detector while scanning for metal objects, in accordance with an example embodiment;
  • FIG. 27C is an outline of metal objects determined based on the data inFIG. 27B, in accordance with an example embodiment;
  • FIG. 28 illustrates a scenario where a vessel detects and retrieves an underwater object, in accordance with an example embodiment.
  • FIG. 29 is a flowchart of a method, in accordance with an example embodiment.
  • DETAILED DESCRIPTIONOverview
  • Haptic interaction has traditionally been a purely virtual task. But recent advancements in depth sensing have made it possible to stream point data with depth information in real-time and at low cost; e.g., by using a depth-enabled camera such as the camera used by Kinect from the Microsoft Corporation of Redmond, Wash. For example, haptic interaction with 6-DOFs from streaming point data can enable applications that fully utilize a 6-DOF haptic rendering device, such as the ‘delta.6’ haptic interface from Force Dimension of Nyon, Switzerland. A 6-DOF haptic rendering method could be useful even though many haptic rendering devices only support 3-DOF actuation (but has 6-DOF sensing), such as the PHANTOM Omni® haptic rendering device from Sensable Inc. of Wilmington, Mass.
  • Example 6-DOF applications include situations for remotely touching a moving object, such as remotely petting a dog. For example, the depth-enabled camera can capture depth images of a remote environment where the dog is located. The depth images can be transmitted to a local computing device. The local computing device can generate a virtual three-dimensional environment showing the depth images (in this example, representing the dog) along with a virtual tool. The virtual tool can be controlled by a local user of the haptic rendering device connected to the local computing device during interaction with the virtual environment. For example, the virtual tool can represent a 6-DOF haptic rendering device controlled by the local user and the virtual tool can take the form of a virtual human hand in the virtual environment. The user can control the 6-DOF haptic rendering device to provide commands to move the virtual tool; e.g., move the virtual human hand to pet the dog. In response, a robot located in the remote environment can receive equivalent commands to those provided to the virtual tool; e.g., so that the robot can move and pet the dog.
  • When haptic interaction is combined with manual control of a robot (or robotic end effector) in co-robotic tasks, such as telerobotic surgery, a 6-DOF capability could add versatility and precision to this co-robotic interaction. For example, it has also recently been shown that the techniques of haptic rendering with 3 DOFs can be useful in telerobotics for implementation of virtual fixtures. That is, the user receives a force when the teleoperated robot is near collision with the environment.
  • In some embodiments, virtual fixtures related to point data can be specified. Examples of virtual fixtures include forbidden-region fixtures and guidance fixtures. A forbidden-region fixture can define an area represented by the point data where the virtual tool is not permitted to enter; i.e., the forbidden region. A guidance fixture can define an area represented by the point data where the virtual tool is permitted, and perhaps encouraged, to enter. Haptic feedback can be provided once the virtual tool attempts entry of a virtual fixture.
  • Another example use of haptic feedback is robotic or telerobotic surgery. In telerobotic surgery, a surgeon can be located at one location which can be remote from an operating room that holds the patient. During the operation, the surgeon can view the patient using depth images provided by a depth-enabled camera in the operating room and transmitted to a local computing device. The local computing device can receive the depth images, generate a virtual environment for the surgeon, receive commands via a 6 DOF haptic rendering device to control a virtual tool and corresponding surgical robot in the operating room, and generate haptic feedback that the surgeon can use to better treat the patient.
  • Haptic feedback can also be useful in remotely controlling exploration devices operating in hostile, dangerous, and/or unsafe environments. Example exploration devices include undersea and outer space exploration vehicles, explosive location devices, chemical sniffing (e.g., drugs, gas leaks, toxic waste) mechanisms, exploration robots, and/or other exploration devices. For example, using haptic feedback, including haptic feedback in forbidden regions, can provide additional information about an environment under exploration and/or protect the exploration device from entering a known dangerous, already-explored, or otherwise forbidden region.
  • Example Environments for Haptic Rendering
  • FIG. 3A is a diagram of ahaptic rendering environment300, in accordance with an example embodiment.Environment300 includescomputing devices320a,320bconnected vianetwork310.Computing device320ais inremote environment340 and is additionally connected to remote controlled device (RCD)336, and depth-enabled (DE) cameras346a-346d. Remote controlleddevice336 is connected totool314.Computing device320bis additionally connected to display312 andhaptic feedback device316.
  • In the example shown inFIG. 3A, each of depth-enabled cameras346a-346dis configured to capture depth data of remote controlleddevice336 and objects342,344 and provide the captured depth data tocomputing device320a. One example of depth data is a depth image having observed color and depth values. In some embodiments, each of depth-enabled cameras346a-346dcan be configured with an optical camera to capture image data and an infrared camera to capture depth data.
  • The depth image can be received as a NR×NC matrix of pixels, with each pixel having 24 RGB bits of color value information, including 8 bits of data for each of red, green and blue component colors. The depth values can be received as a NR×NC matrix, where each depth value has Bdbits. For one example depth-enabled camera, the Kinect camera uses NR=640, NC=480, and Bd=11 or 13, and so produces 640×480 matrices of image data and depth data, where each depth value has at least 11 bits. The Kinect camera can simultaneously provide both image and depth data at a frame rate fcof 30 Hz. The depth values can represent depth relative to the camera ranging from Dmin to Dmax, with Dmin representing a minimum distance from the camera, e.g., 0 mm; and Dmax, e.g., 2048 or 8192 mm. Color and/or depth values can have additional information, such as device information about the camera capturing the depth image.
  • In some embodiments,computing device320acan be connected to more or fewer depth-enabled cameras than shown inFIG. 3A, while remaining connected to at least one depth-enabled camera. In other embodiments, a depth-enabled camera can generate other types of depth data than depth images; e.g., depth-only information, point clouds. In yet other embodiments, devices other than depth-enabled cameras can provide depth data; e.g., depth sensors using radar or another technique to determine depth.
  • Computing device320acan use received depth data to generatevirtual environment330. In order to interpret the depth data,computing device320acan transform the depth data into a collection of points, each point specified in a Cartesian coordinate system in three dimensions; e.g., as an (x, y, z) point in coordinatesystem200 discussed above in the context ofFIG. 2.
  • Each (x, y, z) point from the depth data can represent one point in a “point cloud” or collection of points in three-dimensional Cartesian space; e.g., (x, y, z)ε
    Figure US20140320629A1-20141030-P00001
    3. In other embodiments, other data structures than point clouds or other collections of points can be generated from depth data. After generating the collection of points,computing device320acan rendervirtual environment330 as images and/or as a three dimensional visualization.FIG. 3A showsvirtual environment330 including virtual objects (VOs)332v,virtual valve334v, virtual remote controlleddevice336v, and a representativevirtual tool314v.
  • Computing device320aandremote environment340 can be physically distant fromcomputing device320b. In some scenarios,remote environment340 can be physically near or in the same environment as an environment aroundcomputing device320b. In particular, of these scenarios not shown in the Figures, one computing device can provide the herein-described functionality ofcomputing devices320aand320b.
  • Computing device320acan generate force vectors related totool314 and send indications of haptic feedback tocomputing device320b. Upon reception of the indications of haptic feedback,computing device320bcan utilizehaptic interface device316 to generate the haptic feedback. Additionally,computing device320aand/or320bcan generatevisualization318 withvirtual object332v,virtual valve334v, and/orvirtual tool314v. As also shown inFIG. 3A,visualization318 can include an indication ofvirtual tool314vwhich can correspond tovirtual tool314vinvirtual environment330 and/ortool314 inremote environment340.
  • Haptic interface device316 can be a controllable mechanism configured to receive indications of haptic feedback and provide the indicated haptic feedback based on the indications. Examplehaptic interface devices316 include, but are not limited to a delta.6 haptic interface from Force Dimension, a PHANTOM Omni® Haptic Device from Sensable Inc., other haptic devices, other haptic interfaces, haptic gloves, tactile displays, devices configured at least in part to provide haptic feedback such as laptop computers, desktop computers, mobile telephones, haptic suits, and/or other devices.
  • Ashaptic interface device316 is moved, indication(s) of movement ofhaptic interface device316 can be generated and sent fromcomputing device320b, such as tocomputing device320avianetwork310. Upon reception of the indication(s) of movement,computing device320acan update a position ofvirtual tool314v. Also or instead, computingdevice320acan send control signals to change movement and/or rotation; e.g., change speed, direction, acceleration, pitch, yaw, roll, or to stop) to remote controlleddevice336 to movetool314 accordingly. In other embodiments,virtual tool314vcan represent a position of tool(s), sensor(s), and/or other device(s) on remote controlleddevice336 other thantool314, and by sending control signals remote controlleddevice336, the tool(s), sensor(s), and/or other device(s) can be moved.
  • As depth data ofremote environment340 are captured, the captured depth data can correspond to images and points showing movement and/or rotation of remote controlleddevice336 and/ortool314 and thus showing movement and/or rotation ofvirtual tool314v. In some embodiments, remote controlleddevice336vand/orvirtual tool314vcan be moved withinvirtual environment330 based on the indication(s) of movement/rotation instead of or as well as based on captured image and depth data.
  • An Algorithm for Moving a 6-DOF Virtual Tool
  • An algorithm is described for haptic interaction, which can be classified as a constraint-based virtual coupling method. The haptic interaction algorithm can support real-time haptic interaction with discontinuous streaming point data derived from depth sensors by iteratively resolving collisions for each received set of streaming point data; e.g., a received depth image.
  • The haptic interaction can occur between an arbitrary voxelized virtual tool controlled by a haptic rendering device and an environment represented using streaming point data with depth information derived from a depth sensor. A user of a haptic interface can direct a virtual tool to interact with both static point data for real objects and dynamic point data captured in real-time. The haptic interaction method can provide realistic forces/force feedback for a six degree-of-freedom haptic interaction by performing a rigid body simulation of the virtual tool at a very high rate; in some embodiments, a haptic rendering rate of at least 1000 Hz can be supported.
  • The point data can be represented by one or more depth images representing a set of fixed and infinitely stiff points in space. Each depth image can be filtered and surface normals for objects represented as points in the depth image can be calculated in real-time. Direct interaction between the virtual tool and the point data can be performed without first converting the point data to an intermediate data structure to speed rendering.
  • A quasi-static (at rest) simulation can enable matching of the position and orientation between the haptic rendering device and the virtual tool; i.e., a 6 DOF configuration. Collision between the virtual tool and the virtual environment can be detected. In the virtual environment, the virtual tool can be in one of the following three states: free motion, in contact, or in collision. The free-motion state occurs when no points based on depth information about the environment constrain the virtual tool. The in-contact state occurs when there are points based on depth information about the environment on the boundary of, but not penetrating, the virtual tool. The in-collision state occurs when there are points based on depth information about the environment that penetrate the virtual tool.
  • The quasi-static simulation can involve moving the virtual tool at each of a number of time steps to enable the virtual tool to contact and then bounce off a surface represented contact point(s) in the point data. The virtual tool is be simulated as an object at rest at the beginning of a time step. Then, the state of the virtual tool can be determined. If the virtual tool does not collide with points in the environment, then the virtual tool is in free motion; e.g., the virtual tool is in the free-motion state. If a collision is detected between points in the environment and the virtual tool, then the virtual tool can either be in contact; e.g., the virtual tool is in the in-contact state, or the virtual tool can be in collision; e.g., the virtual tool is in the in-collision state.
  • When the virtual tool is in contact, the points in the environment in contact with the virtual tool can form motion constraints for the virtual tool. These constraints can be used to compute a constrained motion that will move the virtual tool towards the configuration of the haptic rendering device without violating any contact constraints. When the virtual tool is in collision, the collision can be resolved by moving the virtual tool away from the set of collision constraints.
  • Periodically or otherwise, new information about the virtual environment can be received. For example, when a new depth data, such as a depth image, depth frame, or other data including depth information is captured, the new depth data can be transformed into a point cloud or other structure representing points in the environment, where every point in Cartesian space corresponds to a pixel. The point cloud can be filtered; e.g., using a bilateral filter and a surface normal is calculated for every point in the filtered point cloud. This surface normal calculation can lead to collision constraint(s) that point in the correct direction for a surface, regardless of which direction is used to approach the surface. After the state of the virtual tool is determined, the force on the haptic rendering device is calculated as a function of the kinetic distance between the configuration of the virtual tool and the haptic rendering device.
  • Table 1 below includes example pseudo-code describing an example haptic interaction algorithm regarding moving a virtual tool in a virtual environment.
  • TABLE 1
    0001done = false;
    0002WHILE (done == FALSE) DO
    0003 // process depth data Din — see “Real-Time Processing of Streaming Depth Data”
     section.
    0004 IF (new depth data available) THEN
    0005  Receive depth data Din from capture device, with Din including depth information for
    0006   each of a number of pixels captured by captured device
    0007  FOR each pixel P in Din DO
    0008   Let Coords(P) = Cartesian coordinates for P;
    0009   Filter depth values of P;
    0010   Let Normal(P) = normal vector for point P pointed toward capture device;
    0011  END FOR
    0012 END IF
    0013
    0014 // determine motion constraints — see “Collision Detection with the Virtual Tool” section
    0015 // assume no points of Pin are in contact with VT
    0016 Let VOX = voxelization of virtual tool VT;
    0017 Let BB = bounding box of projection of VT onto Din;
    0018 Let state = free-motion;
    0019 Let points_of_contact = NULL;
    0020 render_virtual_environment(Din, VT); // render the virtual environment with VT
    0021
    0022 // now, find out if any points are in contact with VT
    0023 // if point is within bounding box, then see if pixel corresponds to point within
    0024 // voxelization VOX of VT. If point in VOX, then point is in contact or collides with VT
    0025
    0026 FOR each pixel P in Din DO
    0027  IF (Coords(P) is within BB) THEN
    0028   Let P’ = conversion of Coords(P) into coordinates used for VOX;
    0029   Query VOX to determine if P' is a point within boundary of VOX;
    0030   IF (P’ is within boundary of VOX) THEN
    0031    Add P to points_of_contact;
    0032   END IF
    0033  END IF
    0034 END FOR
    0035
    0036 // if there are any points in contact, see if any points penetrate VT. If point penetrates VT,
    0037 // then point is in collision with VT. Otherwise, VT is in-contact (just touching) the point.
    0038
    0039 IF (points_of_contact != NULL) THEN
    0040  Let state = in-contact
    0041  FOR each point PC in points of contact DO
    0042   IF (penetration depth of PC into VT > 0) THEN
    0043    Let state = in-collision
    0044   END IF
    0045  END FOR
    0046 END IF
    0047
    0048 // move VT based on state — see “Finding a Constrained Motion while In Contact”
    0049 // and “Resolving Collisions” sections below
    0050
    0051 initialize_matrix(Matrix_J); // create/zero out Matrix_J as necessary
    0052
    0053 IF (state ==free-motion) THEN // no constraints on moving VT
    0054  Move VT according to unconstrained acceleration of Equation (1) below
    0055 ELSE
    0056  // points in contact/collision add constraints
    0057  FOR each point PC in points_of_contact DO
    0058   Determine constraint(PC) using Equation (2) below
    0059   Add constraint(PC) to Matrix_J;
    0060  END FOR
    0061
    0062  // move VT based on state and constraints
    0063  IF (state == in-contact) THEN
    0064   Move VT according to constrained acceleration (see Equations (3) and (4))
    0065   using nearest point algorithm;
    0066  ELSE // state == in-collision
    0067   Move VT according to constrained acceleration using Equations (6) and (7)
    0068  END IF
    0069 END IF
    0070
    0071 // apply force to haptic rendering device — see “Calculating the Force” section
    0072 Let F = force determined by Equation (8)
    0073 Send indication of F to provide haptic feedback
    0074
    0075 // determine if we can terminate algorithm
    0076 done = are_we_done_yet( ); // or similar technique/function
    0077END WHILE
  • The algorithm shown in Table 1 is specified as mainly as a loop between lines 0002 and 0077 that can iterate one or more times until a variable done is TRUE. A loop iteration can begin at line 0005 by determining if new depth data Din is available. If so, the new depth data Din is processed at lines 0006-0013, and as further discussed in the “Real-Time Processing of Streaming Depth Data” section below.
  • As further discussed in “Collision Detection with the Virtual Tool” section below, lines 0014-0020 involve generating a voxelization VOX of a virtual tool VT and a bounding box BB of a projection of virtual tool VT into a space for Din, as well as initialing a state variable, representing a state of VT, to “free-motion” and a points_of_contact list to NULL (no points in list). Also, the virtual environment is rendered based on depth data Din and virtual tool VT. At lines 0021-0034, a determination is made whether any points represented by the depth data Din are in contact with virtual tool VT. At lines 0035-0046, the state of VT is determined: (a) if no points of Din are in contact with VT, the state of VT remains set as “free-motion”, (b) if a point of DIN is in contact with VT and if none of the in-contact points in the points_of_contact list has penetrated VT, the state of VT is set to “in-contact”, (c) else, at least one point has penetrated VT and so the state of VT is set to “in-collision”.
  • At lines 0047-0069, virtual tool VT is then moved based on VT's state, which is discussed below in the “Finding a Constrained Motion while In Contact” and “Resolving Collisions” sections below. If VT is in a state of free-motion (VT is not in contact with any points of Din and no interpenetrations), the movement is based on unconstrained acceleration specified using Equation (1) below. That is, the virtual tool can be moved in the direction of the unconstrained acceleration until first contact occurs. In some embodiments, the virtual tool can be moved in small discrete steps to ensure that no contact is missed.
  • Otherwise, VT is moved according to constrained acceleration specified using Equations (2)-(7) below. In some embodiments not shown in Table 1 above, bisection can be applied at first contact to further refine the point(s) of contact. In other embodiments not shown in Table 1 above, the constrained acceleration can be applied in small discrete steps, where the magnitude per step is limited to a predetermined amount. The virtual tool can be moved in small discrete steps such that no feature is missed (as for the case when the virtual tool is in free motion). This procedure of movement using small discrete steps can be repeated until collision is resolved. In particular of these embodiments, bisection can be applied again until the collision is resolved.
  • At lines 0070-0073, force feedback is calculated and provided to a user of a haptic tool operating VT, discussed below in the “Calculating the Force” section. At the end of the loop iteration on lines 0074-0077, a determination is made as to whether the algorithm should end or not. If the algorithm should end, the done variable should be set to TRUE to terminate the loop, and therefore the algorithm. Otherwise, done should be set to FALSE to lead to another loop iteration.
  • Real-Time Processing of Streaming Depth Data
  • FIG. 3B showsdepth image350 based on captured depth data anddepth image350 annotated with normalvectors including normals352,354,356,358, in accordance with an example embodiment. The upper image ofFIG. 3B showsdepth image350 of captured depth data representing a human hand and arm.
  • As mentioned above, Cartesian coordinates can be determined for each pixel in a depth image. For each pixel, color and depth values in the depth image can be used to calculate Cartesian coordinates for a corresponding point. With the pixels in the depth image transformed to points in a Cartesian coordinate system, the depth values can be filtered using a bilateral filter, where the points can be weighted using a Gaussian kernel as a function of Euclidean distances in a Cartesian frame. The points can be weighted to preserve depth discontinuities; i.e., a neighboring pixel only contributes to the filtered value if its depth value is sufficiently close.
  • A normal vector can be calculated by fitting a plane to the points in a small neighborhood around each point using a least squares technique. The neighboring points can be weighted using the smooth Wendland weighting function (e.g., with radius rw). This weighted total least squares problem has a closed form solution that can be solved with a 3×3 eigenvalue decomposition for every point. There can be two numerical solutions to the least squares problem, where each solution is a possible normal. The normal can be selected as the numerical solution pointing towards the depth sensor.
  • The lower image shown inFIG. 3B showsdepth image350 annotated by normal vectors; e.g.,normals352,354,356,358, where the normal vectors were calculated using the techniques discussed immediately above for each 1000thpoint derived fromdepth image350. In some embodiments, a normal vector can be calculated for every point in a depth image. In other embodiments, per-pixel data parallel work can be performed by a computing device utilizing one or more graphics processing units (GPUs). For example, the above-mentioned Cartesian transformation, bilateral filtering, and/or least squares solution/eigenvalue decomposition can be done for every pixel in parallel using OpenCL (Khronos Group Inc.) running on the GPU(s).
  • GPU performance can be optimized by copying several pixels at a time to local GPU memory. Then, the copied pixels can be processed in parallel by a work group of computational units; i.e., components of GPUs and/or entire GPUs. As local memory access by the computation units can be up to 2 orders of magnitude faster than remote memory access, performance can be significantly speeded by use of local GPU memory. Local copying can be useful in filtering operations where many neighboring pixels are accessed.
  • Collision Detection with the Virtual Tool
  • In order for the virtual tool to move with respect to the points derived from the depth image, a quick collision detection method is needed. In some embodiments, the virtual tool is voxelized. For example, the voxels can be of a uniform size, such as 0.5 mm on a voxel side. Voxels can be stored in local memory using a simple scan-line technique.
  • To determine whether a point P collides with a virtual tool VT, the point P can be transformed to a corresponding point P′ in a frame of reference for VT. Then, point P′ can be queried against the voxels used to voxelize VT. In some cases, the query can be equivalent to a look up in a collision/no collision lookup table. In some embodiments, collision lookup can be sped by checking for collision with points that are in the neighborhood of virtual tool VT. The neighborhood of VT can be approximated using a two-dimensional bounding box placed around the projection of the virtual tool onto the depth image. Then, a point Pn can be determined to be within the neighborhood of VT if a corresponding pixel Px to point Pn is a pixel within the bounding box projected on to the depth image.
  • FIG. 4A depictsscenario400 of capturing depth data in anenvironment410, in accordance with an example embodiment. The left side ofFIG. 4A shows an overhead view ofenvironment410, with depth enabledcamera412 configured to capture depth data, such as depth images, ofobjects414 and416 as well astool418. At a time T1 inscenario400,tool418 is in front ofobject416 and to the left ofobject414.
  • At time T1, depth enabledcamera412 capturesdepth image420 ofenvironment410.Depth image420 is shown in the upper-right portion ofFIG. 4A, withobject414 on the left side of the image,object416 in the central portion ofdepth image420, and a side view oftool418 shown in front ofobject416. Inscenario400, after conversion of pixels indepth image420 to corresponding points in a Cartesian coordinate system, each point is then checked for collision with a virtualtool representing tool418.
  • As discussed above, a bounding box surrounding the image oftool418 can projected on todepth image420. The lower-right portion ofFIG. 4A showsdepth image420 with projectedbounding box422 replacing a portion ofdepth image420 depictingtool418. Then, points corresponding to pixels ofdepth image420 that are within boundingbox422 can be queried for possible collision withtool418. Inscenario400, the corresponding points can be points captured fromobject416, whose depth values would be larger than those oftool418; i.e.,object416 is behindtool418. As such, no collisions would be detected betweentool418 and points representing the rest ofenvironment410 inscenario400.
  • FIG. 4B depictsscenario450 of capturing depth data in anenvironment410, in accordance with an example embodiment. The left side ofFIG. 4B shows an overhead view ofenvironment410, with depth enabledcamera412,objects414 and416, andtool418 as discussed above in the context ofFIG. 4A. At a time T2 inscenario400,tool418 is in front ofobject416 and just touchingobject414.
  • At time T2, depth enabledcamera412 capturesdepth image460 ofenvironment410.Depth image460 is shown in the upper-right portion ofFIG. 4B, withobject414 on the left side of the image,object416 in the central portion ofdepth image460, and a side view oftool418 shown in front ofobject416 and just touchingobject414. Inscenario450, after conversion of pixels indepth image460 to corresponding points in a Cartesian coordinate system, each point is then checked for collision with a virtualtool representing tool418.
  • As discussed above, a bounding box surrounding the image oftool418 can projected on todepth image460. The lower-right portion ofFIG. 4B showsdepth image460 with projectedbounding box462 replacing a portion ofdepth image460 depictingtool418. Then, points corresponding to pixels ofdepth image460 that are within boundingbox462 can be queried for possible collision withtool418. Inscenario450, some of the corresponding points can be points captured fromobject416 as discussed above in the context ofFIG. 4A. Other of the corresponding points can be points fromobject414, such as points at the left-most extent of boundingbox462 that just touchesobject414. The depth values can be similar to those depth values representing a position oftool418; i.e., object414 may be in contact with and/or penetrated bytool418. As such, collisions may be detected betweenobject414 andtool418 inscenario450.
  • Finding a Constrained Motion while in Contact
  • Using this collision detection method, the virtual tool is moved towards the configuration of the haptic rendering device and stopped at the first point of contact. Gauss' least constraints principle can provide a basis for find a feasible movement for the virtual tool that will not violate the contact constraint further.
  • The generalized unconstrained acceleration agucan be expressed using Equation (1) below:
  • agu=(auαu)(1)
  • where auis the unconstrained acceleration in the horizontal and vertical (X and Y) dimensions, and αuis the unconstrained acceleration in the depth (Z) dimension.
  • agucan be considered as an ideal unit step; e.g., a unit step that can be taken if the configuration of virtual tool VT were to match the configuration of the haptic rendering device. However if virtual tool VT is in contact, this ideal unit step cannot be taken. Each point of contact (as found by the collision detection) can introduce a linear constraint on accelerations a and a specified by Equation (2):

  • nTa+(r×n)Tα≧d  (2)
  • For equation (2), n is the normal vector at the point of contact, r is the vector from the center of mass to the point of contact, and d is the penetration depth of the point of contact with d=0 when virtual tool VT is in contact with, but not penetrated by, the contact point and d<0 when virtual tool VT is penetrated by the contact point.
  • Given agu, a constrained acceleration accan be calculated which minimizes the kinetic distance between the virtual tool and the haptic rendering device with respect to the linearized constraints. More formally, the constrained acceleration accan be found as
  • ac=12minag(agu-ag)TM(agu-ag)(3)
  • subject to

  • Jag≧0  (4)
  • where M is the mass matrix for the virtual tool and J is a matrix of constraints formed by (2) for each point in contact.
  • Equations (3) and (4) represent a quadratic programming problem solvable using Wilhelmsen's nearest point algorithm. The constrained acceleration acmay be valid for only a small neighborhood around the current configuration because of the linearized contact constraints J. Further, validity of accan be affected any movement by the virtual tool VT and/or changes in point positions introduced by new depth data (represented by Pin), as the movement can introduce new constraints on virtual tool VT.
  • To illustrate the depth value d of Equation (2),FIG. 4C depictsvirtual tool470 and example points (Ps)472a,474a, and476arepresenting depth data captured from an environment. For example, points472a,474a, and476acan be points in a Cartesian coordinate system calculated from pixels of a depth image, as discussed above. As mentioned above, a normal vector can be calculated by fitting a plane, or surface, to each ofpoints472a,474a, and476aa small neighborhood around each point. For example,FIG. 4C shows an example surface (S)472bcalculated forpoint472aand corresponding normal (N)472c.FIG. 4C also showssurfaces474b,476bandnormals474c,476cforrespective points474a,476a.
  • The depth d for a point P with respect to virtual tool VT can be specified as a distance from P to a portion of VT closest to P along the direction of a normal vector N associated with P. For example,FIG. 4C shows thatpoint472ais a positive distance from a portion ofvirtual tool470 closest to point472ain the direction of normal472c; that is, the depth ofpoint472awith respect tovirtual tool470 is positive. As the depth ofpoint472ais positive,virtual tool470 can be classified as in free motion, or unconstrained by,point472a.
  • As another example,FIG. 4C shows thatpoint474ais just touching a surface ofvirtual tool470. That is,point474ais a distance of 0 units away fromvirtual tool470 in the direction of normal474c, and so the depth ofpoint474ais 0. As the depth ofpoint474ais 0,virtual tool470 can be classified as being in contact withpoint474a.FIG. 4C also shows thatpoint476ais inside the surfaces ofvirtual tool470. That is,point476ais a negative distance away from a closest surface ofvirtual tool470 in the direction of normal476c, and so the depth ofpoint474cis negative. As the depth ofpoint476ais negative,virtual tool470 can be classified as being in collision withpoint476a. As such,virtual tool470 can be considered to be constrained bypoints474aand476a.
  • Resolving Collisions
  • One technique to resolve collisions between VT and points is based on the use of interval arithmetic. Generally speaking, interval arithmetic operates on ranges (or intervals) of values rather than specific values. For example, suppose bodies A and B are 6.5 feet apart. For interval arithmetic, the relative position of A and B can be specified based on a distance interval; e.g., an interval between 6 to 8 feet.
  • As part of this collision-resolution technique, the step size to move any part of the virtual tool can be constrained to be less than the smallest size of voxels used to represent the virtual tool. For example, the step size can be chosen to be half the tool voxel size. Bounding the virtual tool movement can ensure that no collisions are missed in a single point image, or when static point data is used. However, when dynamic point data is used, bounding the virtual tool's movement does not hold after the point data updates, as the locations of points represented by a new point image or other representation of the point data can be arbitrary.
  • A quasi-static collision resolution technique can be used to resolving collisions between virtual tools and virtual environments represented by dynamic point data. Initially, virtual tool VT is at rest; e.g., satisfies the condition specified by Equation (5):

  • agu=0  (5)
  • To move the virtual tool VT, the depth value d (discussed above in the context of Equation (2) and below in the context ofFIG. 4C) has to be non-zero in order for the minimization to move the virtual tool out of the constraint. The value of d should ideally match the penetration depth at each point of collision. However, a planar constraint specified in Equation (2) may only be valid in a small neighborhood of virtual tool VT and since movements of virtual tool VT might introduce new constraints.
  • An approximate solution to resolve collisions can be obtained by weighting the constraints equally and updating the configuration of the virtual tool iteratively. This approach can be summarized as:
  • ac=12minagagTMag(6)
  • subject to

  • Jag≧1  (7)
  • Note that the right hand side value “1” in Equation (7) can be chosen as any vector with identical elements. The choice of right hand side value can change the magnitude, but not the direction of the result from Equation (6).
  • Calculating the Force
  • The force f sent to the haptic rendering device can be calculated as a function of the difference between the virtual tool and haptic device configuration. For example, f can specified as f=(fTfR)T, with fTε
    Figure US20140320629A1-20141030-P00002
    3being a translational component, and fRε
    Figure US20140320629A1-20141030-P00002
    3being a rotational component. Then, fTcan correspond to three degrees of freedom for a position of virtual tool VT and fRcan correspond to three degrees of freedom for a rotation of virtual tool VT.
  • The force f can be calculated as:

  • f=KMagu  (8)
  • where K is a diagonal matrix containing spring constants between virtual tool VT and each contact point. For an example of spring constants, a virtual spring with a corresponding spring constant can be connected between virtual tool VT and a contact point. As the distance between VT and the contact point increases, the virtual spring exerts a proportionally larger force to draw VT closer to the contact point. The calculated force can be exerted in the direction of the normal at the contact point; i.e., corresponding to the virtual spring pulling VT and the contact point together.
  • Example Implementation
  • As an example, the herein-described 6-DOF haptic rendering algorithm was implemented on a desktop computer (AMD Phantom II X6 with a Radeon HD 6990 GPU) running Ubuntu 11.10. The force was calculated asynchronously at 1000 Hz in a separate thread. During typical interaction, the collision detection algorithm ran at 15 kHz. Point images were filtered and normal vectors were calculated for every point using the GPU.
  • Realtime processing was achieved using a neighborhood of 9×9 points for filtering as well as normal vector calculation. The position of the haptic rendering device was both controlled automatically (for purposes of producing accurate results) as well as with a Phantom Omni haptic rendering device. Using the latter, only translational forces could be perceived by the user since the Phantom Omni only provides 3 DOFs of sensation.
  • To evaluate the presented haptic rendering method in a noise free environment, a virtual box (with 5 sides but no top) was constructed. The normal vectors were set perpendicular to each side, pointing into the box. The position of the haptic rendering device was then automated to move 2 s into the box, rotate in the positive direction for 2 s, rotate in the opposite direction for 2 s, and then move for 2 s out of the box. In this example, the virtual tool became constrained by the sides of the box and interacted with up to 6 contact points simultaneously.
  • Two examples of haptic interaction with streaming point data were performed using arbitrary polygon models as virtual tools. The performance of the haptic rendering method can depend on the number of points in the neighborhood of the virtual tool. In some embodiments, most of the computational time can be spent on collision detection and forming movement constraints. In these embodiments, performance of the algorithm can scale approximately linearly in the number of neighboring points. To improve performance in interaction with large tools, the point data can be down-sampled. Further, performance can be enhanced by utilizing multiple CPUs of a computing device for collision detection. Also, the algorithm can be implemented partially or completely using interval arithmetic, which may improve performance as well.
  • Example Technique for Generating and Enforcing Virtual Fixtures
  • FIG. 5A depictspoints510 derived from depth data and forbidden-region fixture514, in accordance with an example embodiment.FIG. 5A showspoints510 including points512a-512i. Forbidden-region fixture514 can define a forbidden region that includes some ofpoints510; e.g., points512d-512fas shown inFIG. 5A, and surrounding space where a virtual tool is prohibited from entry.
  • A forbidden region can be generated by selecting points, planes, regions of space, and/or objects in a virtual environment to define the forbidden region. For each forbidden-region point picompletely within the forbidden region, a forbidden-region radius rf,iand a forbidden-region stiffness ks, ican be determined.FIG. 5A shows a two dimensional forbidden region feature defined by forbidden-region points512d-512fand where rf,1=rf,2=rf,3. During haptic rendering ofpoints510, a virtual tool will be prohibited from entering forbiddenregion514.
  • During haptic rendering, the forbidden regions can be considered to be defined by forbidden-region spheres of the forbidden-region fixture, each forbidden-region sphere defined by a forbidden-region point piand related forbidden-region radius rf,i. Then, the virtual tool can be constrained to only move on the surface of or away from the forbidden-region spheres.
  • FIG. 5B depictsvirtual tool520, in accordance with an example embodiment.Virtual tool520 has acenter522, shown with a V inFIG. 5B.Virtual tool520 also has contact point(s)524, defined as one or more points on an exterior surface ofvirtual tool520.
  • The definitions of virtual-tool states can be modified to account for forbidden regions. The virtual tool can be considered to be in a free-motion state with respect to one or more forbidden regions if the virtual tool is not on the boundary or inside of any of the one or more forbidden regions. The definition of depth can be similarly modified, so that a modified depth MD of a virtual tool VT with respect to a forbidden region F having outward normal FN can be specified as distance between contact point(s) of VT and a closest portion of forbidden region F moving in a direction of FN. In terms of the modified depth, VT is in a free-motion state with respect to forbidden region if the modified depth MD between VT and F is greater than 0.
  • The virtual tool can be considered to be in an in-contact state with respect to one or more forbidden regions if a portion of one or more of the forbidden region(s) is just touching contact point(s)524; e.g., the modified depth MD between the virtual tool and the forbidden region equals 0. The virtual tool can be considered to be in an in-collision state with respect to one or more forbidden regions if a portion of one or more of the forbidden region(s) is within contact points of the virtual tool e.g., the modified depth MD between the virtual tool and the forbidden region is less than 0.
  • When the virtual tool is in a free motion state, movement along the vector of unconstrained acceleration is allowed as long as the virtual tool does not come in contact with or collide with a forbidden region; e.g., as long as the modified depth remains greater than zero.
  • FIG. 5C shows an examplevirtual tool530 whose movement toward desiredvirtual tool position550 is impeded by forbiddenregion540, in accordance with an example embodiment. As shown in the example ofFIG. 5C, forbiddenregion540 centered at point pican constrain movement of free-motionvirtual tool530 toward desiredvirtual tool position550 in the direction of vector u. Specifically,FIG. 5C shows that when a center ofvirtual tool530 moves to a position P2,virtual tool530 will be in contact with forbiddenregion540 at point pc. The outward normal to forbidden region N(pc) indicates a direction of motion forvirtual tool530 to avoid forbiddenregion530.
  • FIG. 5D shows an examplevirtual tool534 in collision with forbiddenregion560 centered at point pj, in accordance with an example embodiment. As shown in the example ofFIG. 5D, a portion ofvirtual tool534 is inside of, and so is in collision with, forbiddenregion560. For example, forbiddenregion560 can be generated based on new depth information received from a depth-enabled camera or similar device. In particular, point pcof the contact points forvirtual tool534 is a point furthest inside forbiddenregion560; e.g., a point ofvirtual tool534 closest to point pjcentering forbiddenregion560. As such,virtual tool534 cannot travel along vector u2 toward desiredvirtual tool position570 without at least partially traversing forbiddenregion560. A vector u3 from center point pjto contact point pcindicates a direction of motion for virtual tool535 to escape forbiddenregion560; e.g.,virtual tool534 can move along a weighted vector sum of u2 and u3 to escape forbiddenregion560 while progressing toward desiredvirtual tool position570.
  • Example Applications Involving Haptic Rendering
  • FIG. 6 is anexample gaming scenario600 with haptic rendering inenvironment610, in accordance with an example embodiment. Inscenario600,players620 and630 are playing a game of “capture the flag” ingame environment610.FIG. 6 showsplayer620 as a black circle ingame environment610 and guardingflag622, and showsplayer630 as a grey circle ingame environment610 and guardingflag632.
  • In capture the flag, each player tries to be the first player to capture the opponent's flag and return the captured flag to a goal area. That is, during the game,player620 attempts to captureflag632 and returnflag632 togoal624 beforeplayer630 capturesflag622 and returnsflag622 togoal634. In the variation of capture the flag shown inscenario600, a player can lose if a shield level for the player goes to 0% as well.
  • Inscenario600, bothplayers620 and630 utilize haptic feedback devices, such as haptic gloves, body suits with haptic feedback generators, and/or other haptic devices. Also,players620 and630 each have computing devices configured to use game-playing software access a virtual environment generated from depth data generated by depth-enabledcameras650,652,654,656,658,660,662, and664 (labeled as C's inFIG. 6). In other scenarios, more or fewer cameras can be used than shown inFIG. 6. Also, inscenario600, bothplayers620 and630 have sensors configured to provide location and heading information for the player. For example, bothplayers620 and630 could have a haptic suit with a portable display that includes Global Positioning System (GPS) sensor(s), accelerometer(s), gaze detection sensors, and/or other sensors to both access the virtual environment, provide location and heading information, and receive haptic feedback.
  • Software for the game can determine a location and heading for each player withinenvironment610 and generate a display ofenvironment610 at the player's location as the player looks in the heading direction.FIG. 6 shows display670 generated forplayer620 with a view ofbox640 shown with a black star. Inscenario600, a “charging object”, or star having the player's color on an object withinenvironment610 increases or “charges” the player's shield level as long as the player touches the object. In contrast, if the player touches a “discharging object”, or object having a star of the color of the opponent, the player's shield level decreases or “discharges.” In some embodiments, a discharging object can produce a forbidden zone that discharges a shield level while a player is within the forbidden zone. The colored star used for identification of charging and discharging objects can be generated by the game-playing software to permit changes in colors of player and/or changing which objects act as charging and/or discharging objects. Also, in some embodiments, touching a discharging object can quickly or even immediately reduce a shield level to 0—e.g., the discharging object causes the player to (nearly) instantly lose the game. Depending on the scenario, such instant-discharge objects may or may not be identified to the player by the game-playing software.
  • Along with identification of charging and discharging objects, the game-playing software can generate slogans, images, etc. on objects in the environment. For example, inscenario600,player620 has a heading of due south, and the game-playing software has generated the slogan “Lose any flags?” on an image ofobject644 indisplay670. Similarly, inscenario600,display680 forplayer630 has slogans “Dare to pass?” and “Run!” display on generated images ofobjects644. In some scenarios, the game-playing software may provide images of objects without additional slogans, images, etc.
  • Displays670 and680 each include game-related data for each player, including a shield level, a number of flags taken, and a number of opponents inenvironment610. For example,FIG. 6 shows display670 forplayer620 with game-related data that indicates that player620: (a) has a shield level of 98% and that the shields are charging, (b) has not taken any flags, and (c) has one opponent inenvironment610.FIG. 6 also showsdisplay680 forplayer630 with game-related data that indicates that player630: (a) has a shield level of 100%, (b) has not taken any flags, and (c) has one opponent inenvironment610. In some embodiments, more, different, and/or less game-related data can be provided in a display for a player.
  • In some embodiments, the game-playing software can generate and/or display virtual objects as well as, or instead of, actual objects such asobjects640,642,644,646, and648 ofenvironment610. For example, objects648aand648b, each shown using dashed lines, can be virtual objects. As one example, a virtual object can represent a “regular” object similar to object646 or648 but is not physically present inenvironment610. In other examples, a virtual object can represent a trap such as a “covered pit” that an unwary player can fall into and lose shield strength, or a treasure such as “supercharger” that can immediately add shield strength to the player.
  • Many other examples of real and virtual objects are possible, including examples of other games that utilize haptic rendering. Some of these games can involve only one player and some games can involve more than two players. For example, a maze game with haptic feedback can involve a single player exploring the maze or can involve two or more players perhaps running a race through the maze.
  • FIG. 7 is an example hapticrendering session scenario700 in aremote environment710, in accordance with an example embodiment.Environment710 includes a dog “Fido”720, a depth-enabledcamera730, and a haptically-controlled device (HCD)732. Duringscenario700, a remote viewer interacts withFido720 viaHCD732. For example,HCD732 can be a mobile robot with a robot arm configured for remote control. Other HCDs are possible as well. In some scenarios,HCD732 is not present, which permits interaction with a virtual environment generated by data fromcamera730 without the ability to affect a real environment, such asenvironment710.
  • Display740 can be generated to visualize a virtual environment utilizing depth data generated by depth-enabledcamera730.FIG. 7 shows display740 withvisualization portion742 configured to display a view ofenvironment710 and textual portion744 configured to display both instructions, such as “Move your glove to pet Fido where the black hand touches him”, and other information such as “If you′re good to him, Fido might take you for a walk in the woods.”
  • A user conducting a haptic rendering session scenario can use a haptic glove or other haptic interface device to controlvirtual tool746 and touchFido720 using the robot arm ofHCD732. In some scenarios, the user can control movement ofHCD732, perhaps by certain movements of the haptic interface device; e.g., press a “move robot” button or otherwise signal a change of interpretation of movements of the haptic interface device from being interpreted as commands for moving the robot arm to commands for movingHCD732.
  • Inscenario700, the move robot button can be pressed to moveHCD732 alongpath734 and so walkingFido720 based on movements of the haptic interface device (e.g., move haptic interface device left or right to correspondingly moveHCD732 left or right alongpath734, rotate haptic interface device to better pet Fido). When the move robot button is not pressed, movements of the haptic interface device control the robot arm (e.g., for a haptic glove acting as the haptic interface device,Fido720 can be petted or scratched based on finger movements of the haptic glove).
  • In other scenarios, non-haptic interface devices can be used to controlHCD732; e.g., the haptic interface device controls the robot arm and a joystick or keyboard can be used to move the mobile robot. In still other scenarios,camera730 can be mounted onHCD732 to provide additional information aboutenvironment710, including information aboutFido720.
  • FIG. 8 is an example collaborative hapticrendering session scenario800 for aremote environment810, in accordance with an example embodiment.Environment810 includes pumpingstation812 that, duringscenario800, has failed. Due to the failure of pumpingstation812,waste814 has escaped, as shown byFIG. 8 as a black cloud inenvironment810.
  • Duringscenario810, fourrobots820,830,840, and850 have been deployed to fix pumpingstation810 and investigateenvironment810 to begin cleaningwaste814. Each ofrobots820,830,840, and850 has a depth-enabled camera facing forward and can be driven by a remote operator using a display, such asdisplay312 discussed above in the context ofFIG. 3A, and one or more haptic interface devices, such as haptic interface device(s)316 discussed above in the context ofFIG. 3A.
  • Inscenario800, each robot operator is in a different physical location. Each location is equipped with one or more haptic interface devices, one or more displays, and perhaps other interface devices, such as keyboards, keypads, touch screens, loudspeakers, microphones, and/or other interfaces. In other scenarios, some or all of the robot operators can be in the same remote location. In still other scenarios, a single robot can be used with a single robot operator. In even other scenarios, one or more of the robots can be controlled by multiple robot operators e.g., both a driver and a ladder operator can control a robotic fire engine. In yet other scenarios, one or more of the robot operators can be local; e.g., atenvironment810, perhaps riding on or within a robot.
  • FIG. 8 shows the displays generated the depth-enabled cameras for the robot operators, asdisplay860 forrobot820 headed along heading826,display870 forrobot830 headed along heading836,display880 forrobot840 headed along heading846, and display890 forrobot850 headed along heading856.Robot820 has two robot hands configured as haptic interface devices (HIDs)822,824,robot840 has tworobot hands842,844 configured as haptic interface devices, androbot850 hasrobot hand852 and probe854 configured as haptic interface devices. Each robot display shown inFIG. 8 includes a visualization portion and a textual portion, similar to display1040 shown inFIG. 10. For example,display860 includesvisualization portion862 andtextual portion864. InFIG. 8, virtual tools representing robot hands are shown using block diagrams of hands with two fingers and virtual tools representing probes are shown using slender triangles; e.g., pennant-shaped triangles.
  • Inscenario800,robot830 is in communication with theother robots820,840, and850, and is configured to a “communication coordinator” to send and receive text, voice, and perhaps other types of messages from the other robot operators. For example,robot830 can be controlled by asupervisor observing environment810 and coordinating the efforts ofrobots820,840, and850.
  • Inscenario800, a far end of each robot arm ofrobot820 is configured to be a virtual tool for the operator ofrobot820.FIG. 8 shows display860 forrobot820 including a location of a leftvirtual tool866, corresponding to a left robot hand on a left arm ofrobot820, and a rightvirtual tool868 corresponding to a right robot hand on a right arm ofrobot820. InFIG. 8, rightvirtual tool868 is position to move a handle of a door to pumpingstation812.
  • The operator ofrobot820 can use a non-haptic interface device, such as a keyboard or keypad to enter text that appears intextual portion864.FIG. 8 shows that the text “Joe, I'm trying to open the R door.” Is intextual portion864, preceded by an identifier “Chris” of a person who entered in the text; e.g., Chris is a first name of the operator ofrobot820. Inscenario820, text entered into atextual portion864 ofdisplay860,textual portion884 ofdisplay880, ortextual portion894 ofdisplay890 is displayed both in the entering textual portion plus intextual portion874, asdisplay870 andtextual portion874 are associated with the communication coordinator. In some embodiments, robot operators can send a message to one or more other operators directly without sending a copy of the message to the communication coordinator.
  • In other embodiments, the operator ofrobot820 can receive touch feedback while attempting to open the door of pumpingstation812. For example, as the operator ofrobot820 usesrobot hand822 to open the door of pumpingstation812, haptic rendering can provide feedback to indicate that the door is or is not resisting the efforts to open the door. Further, in embodiments whererobot hand822 is equipped with finger-like appendages, the operator ofrobot820 can use a haptic glove or other haptic interface device to move the appendages to grab the door handle, and pull down or push up with an amount of effort based on, e.g., proportional to, an amount of effort exerted by the operator in moving the haptic interface device.
  • Robot850 is equipped withprobe854. A probe can be equipped with one or more sensors for chemical, biological, and/or radiation testing.FIG. 8 showsrobot850 withprobe854 inwaste814.
  • Each of displays880 (for robot840) and890 (for robot850) shows a virtual tool associated with either a robot hand or a probe. For example, forrobot840,robot hand842 is associated withvirtual tool886 androbot hand844 is associated withvirtual tool888. Inscenario800, bothrobot hands842 and844 include probes.
  • Robot840 is engaged in measuring waste densities withrobot hands842 and844.FIG. 8 showstextual portion884 ofdisplay880 displaying various types of information, including location information for each probe and waste density values detected by each probe. Inscenario800, a waste density of “IP” indicates a probe is in the process of determining a waste density, such as shown inFIG. 8 forprobe844.
  • Both haptic and non-haptic commands can be used to control robots.Textual portion894 ofdisplay890 shows commands being communicated from an operator torobot850.FIG. 8 shows a “Cmd:” prompt for commands, such as “set left to rad”, which directsrobot850 to set aprobe854 to act as a radiation sensor.Robot850 responds with a confirmatory message from “Probe” of “Set for rad” to indicate thatprobe854 is ready for radiation testing. Inscenario800, the operator ofrobot850 sends another command “test probe” torobot850. In response,probe854 performs a radiation test and determines a value of “+1.3” for the radiation test. Inscenario800, the operator ofrobot850 can moverobot hand852 and probe854 using respective haptic interface devices.
  • FIG. 9 depictsscenario900 where construction is ongoing inenvironment910, in accordance with an example embodiment.Environment910 can be a virtual environment or an actual environment. In some embodiments,environment910 can represent an extra-terrestrial environment; e.g., in space or on a planet other than Earth. In these embodiments, values of various parameters, such as the speed of gravity (if any), planetary tilt, and relative location of other celestial bodies, can change from Earth-normal values.
  • Inscenario900, a solar power array is being constructed from at least two depicted sub-arrays: sub-array T1-182, shown on the left side ofenvironment910, and sub-array T2-182, shown on the right side ofenvironment910. To construct the solar power array, sub-array T1-182 is being manipulated bytool920 and sub-array T2-182 is being manipulated bytool922. In other scenarios, more or fewer tools can be utilized inenvironment910.
  • Inscenario900,display940 is used to monitor and control a haptic interaction session withtool920 anddisplay950 is used to monitor and control tool a haptic interaction session withtool922.Display940 includes environmental-display region942 and textual-display region944. Environmental-display region942 includes a display of sub-arrays T1-182 and T2-182,virtual tool946 representingtool920, and two control buttons. The “Bye” control button ends the haptic interaction session and the “Change Tool” button enables changingdisplay940 to monitor/control to a different virtual tool and/or to change operation oftool920; e.g., changing use of a robot hand as part oftool920 as shown inFIG. 9. More, fewer and/or different control buttons can be utilized in other scenarios. Textual-display region944 can provide text and perhaps other types of information (e.g., images, sounds, sensor information, communication sessions, etc.) for use as part ofdisplay940.
  • Virtual tool946, and subsequentlytool920, can be controlled by use of a 6 DOF (or other) haptic interface to enable control over location and rotation ofvirtual tool946 and to receive force feedback related to related to the virtual tool using the herein-described techniques.
  • In other scenarios,display940 can be used to monitor and control multiple haptic interaction sessions: for example, a robot arm can have three virtual tools corresponding to a shoulder, an elbow, and a wrist, with each virtual tool having up to 6 DOF. Then, ifdisplay940 were used to control the robot arm,display940 can control haptic interaction sessions for each of the virtual tools. As another example, a tool can have multiple controllable components; e.g., robot arms, robot hands, an engine/propulsion system, etc., where some or all of the controllable components can be directed using haptic interaction session(s). Then, ifdisplay940 were used to monitor and control such a tool,display940 can control one or more haptic interaction session(s) with the tool. Many other examples are possible as well.
  • Display950 is similar todisplay940.Display950 includes environmental-display region952 and textual-display region954. Environmental-display region952 includes a display of sub-arrays T1-182 and T2-182,virtual tool956 representingtool922, and “Bye” and “Change Tool” control buttons discussed above in the context ofdisplay940. Textual-display region954 can provide text and perhaps other types of information for use as part ofdisplay950.
  • Inscenario900,tools920 and922, controlled viarespective displays940 and950 and corresponding haptic interaction sessions, construct the solar power array using sub-arrays T1-182 and T2-182.Scenario900 can then end.
  • Underwater Haptic Rendering Applications
  • Virtual tools, haptic interaction sessions, and the herein-described techniques can be applied to underwater applications. For example, a underwater depth-enabled camera, similar in spirit to the Microsoft Xbox Kinect, but suitable for underwater conditions can be used for locating objects underwater and providing data for 6 DOF haptic feedback to a human operator of the robot. The haptic feedback can provide a human operator with a sense of touch for objects seen by the camera.
  • This system can: (a) permit the operator to ‘feel’ objects, such as underwater ordnance during remediation or objects being salvaged, through a haptic 6 DOF interface, based upon the camera system image and dynamic haptic rendering, (b) enable the operator to guide the robot end-effectors with the target during removal, via tele-operation, and (c) establish virtual ‘force fields’ around protected zones of objects (such as locations that might result in explosion of the ordnance). If the tele-operator tries to move an end-effector too close to a protected zone, he/she will feel resistance as the haptic interface “pushes back”. Imposition of protected zones, perhaps via virtual features, can prevent robot end effector contact with undesirable locations, either in the environment or on the ordnance. The end effector of a robot can be a device at an end of a robot arm or other appendage; e.g., a robotic hand at the end of robot arm. Combined with a tele-operated robotic device, this will allow human directed robotic capture and removal of objects under fresh or salt water.
  • FIG. 10 showsunderwater robot1000, in accordance with an example embodiment.Robot1000 is configured with twoscrew drives1002,1004, eightrobotic arms1010,1011,1012,1013,1014,1015,1016,1017,light sources1020,1024, andcamera1022.
  • Robot1000 can be a bottom-crawling vehicle with a screw-drive propulsion system.Robot1000 can move over an underwater object, such as ordnance, and the tele-operator uses robot arms1010-1017 to grab the underwater object.Light sources1020,1024 andcamera1022 can operate together as a camera system.
  • As shown inFIG. 10,camera system1020,1022,1024 can be integrated into a chassis ofrobot1000 and image the region withinservice bay1030.Camera system1020,1022,1024 can generate a three-dimensional image of object(s) withinservice bay1030, such asobject1032. The images generated bycamera system1020,1022,1024 can be provided to a processor at the surface to haptically rendering images withinservice bay1030. A human tele-operator interacts withrobot1000, arms1010-1017, and objects such asobject1032, using one or more haptic interfaces and perhaps display(s). The tele-operator can control arms1010-1017 to moveobject1032, pick upobject1032, attach a sling and cable to object1032 for future retrieval, and/or some other action such as mark the object for better identification.
  • Robot1000 can use a system of screw drives1002,1004 for propulsion. Using screw drives maintains negative buoyancy throughout a mission and thusrobot1000 need not manage buoyancy. The use of screw drives, in contrast to thrusters, lowers the risk of stirring up silt or sand, which can deteriorate visibility. Additionally, silt poses a challenge to traditional track-driven vehicles not faced by screw drives.
  • FIG. 10 shows that screw drives1002,1004 are mounted on each side ofrobot1000. Eachscrew drive1002,1004 can be driven at varying speeds.Robot1000 can move sideways and thrust may be vectored so that non-holonomic constraints do not limit operation. Thrust vectoring allows for increased maneuverability and simplifies the role of the tele-operator, allowing for natural modes of operation and quick adjustments to the platform's location on the sea, river, or lake bed. Screw drives1002,1004 can also be articulated, allowing for a range of maneuvers and for convenient movements such as vertical translation of the chassis andservice bay1032 ofrobot1000. In some scenarios, screw drives1002,1004 can dig into mud, getting therobot1000 deeper into areas that may have buried ordnance.
  • Robotic manipulation and removal of munitions requires a stable platform that can remain fixed in an inertial frame with respect to the object. To carefully and firmly grip an object identified for removal, the system can use a number of robotic manipulators designed for the task, such as arms1010-1017.
  • Arms1010-1017 can have a series of linkages and the necessary actuation to achieve motion derivatives that accomplish desired gripping maneuvers to handle underwater objects. In some embodiments, such as shown inFIG. 10, each of arms1010-1017 is identical and is designed for planar motion. When gripping an object, the trajectory of each of arms1010-1017 is designed to achieve optimal grasping with respect to the goals and constraints defined by a human operator.
  • In some embodiments, some of arms1010-1017 can be controlled using a low number of degrees of freedom. Even with low degree-of-freedom arms, intricate shapes can still be achieved with the coordinated manipulator system to allow for a high level of dexterity using simple components. As shown inFIG. 10, arms1010-1017 can be mounted along either side of theservice bay1030, allowing them to oppose on another so that simple motions can result in stable gripping of munitions. The use of eight arms allows for a lower contact pressure on an object from each contacting arm. Also, a subset of arms1010-1017 can be used, such as when a forbidden-region fixture onobject1032 prohibits one or more arms from grippingobject1032 at a particular location. In some embodiments,robot1000 and arms1010-1017 can operate in several modes of operation and redundancy, allowing an operator to carry out tasks even when a manipulator suffers damage.
  • The camera system can use a ‘range gated’ approach to allow both visible (blue green) and NIR images to be captured in combined stream of visible (blue green) and NIR images.Light source1020 can be configured with blue-green filters and/or blue-green light-emitting diodes. The blue-green filters are configured to pass through light in a blue-green frequency range of 450-480 nm, while the blue-green diodes are configured to emit light in the blue-green frequency range.
  • Light source1024 can include a NIR laser and a diffraction grating. The NIR laser and diffraction grating project a pseudo-random of dots of varying intensities.Camera1022 can measure depth from the distortion between the projected and observed dot patterns. In some embodiments, the pulse duration for the NIR laser can be shorter than the time to travel to the target to reduce back scatter.
  • Camera system1022 can be configured to capture the visible (blue green) and NIR images alternatively; that is illumination and camera frames are synchronized to capture the visible (blue green) and NIR images alternatively. That is,light sources1020,1024 can be alternatively activated at frame speed ofcamera1022 to capture visible and NIR images in a single image stream. Dual-wavelength camera system1022 can allowrobot1000 to obtain 2D images with depth measurements in low light conditions using a single camera. In some embodiments,camera system1022 can be configured as an endoscope.Camera1022 can include one or more adapters to combine the projected mask pattern from thelight source1024 withvisible light source1020, and recover the depth information from the reflected light with a wavelength-specific beam splitter.
  • In some embodiments,camera1022 can be a modified 0° or 30° 10 mm endoscope with light sources and detectors separated by fixed distances. In particular embodiments,camera1022 can output frames with VGA resolution (720×900 pixels) at 30 frames per second, in RGB+Depth (RGB+D) format.
  • To calibrate the visible and depth images, planar calibration ofcamera system1020,1022,1024 can first be performed on a planar surface with a checkerboard pattern. Withcamera system1020,1022,1024 configured for close depth detection, the checkerboard can be produced with photolithographic techniques for both scale and accuracy. After planar testing is complete, optical testing ofcamera system1020,1022,1024 can be performed on an irregular non-planar surface to determine a depth resolution, and any effect of geometric distortion on the optical distortion ofcamera system1020,1022,1024. If distortion is detected, a distortion-correction technique that automatically calculates correction parameters without precise knowledge of horizontal and vertical orientation can be used to correct the detected distortion.
  • Camera system1020,1022,1024 is discussed in more detail below in the context of at leastFIGS. 17 through 29.
  • The above-mentioned haptic rendering process can usecamera system1020,1022,1024 to generate depth data, depth images, and haptic feedback. The haptic feedback can give a tele-operator a “feel” for underwater objects that are within the field of the underwater video+depth cameras, such as underwater ordnance during remediation. The tele-operator's console (to transmit operator control actions as well as haptic feedback) can use two haptic interface devices such as discussed above in the context ofFIG. 3A. As the tele-operator moves the robot through the haptic interface devices, he/she can feel a “force field” of increasing impedance, as proximity of the tool tip to the virtual fixture boundary decreases.
  • Virtual fixtures can be established around the portion of protected structure; that is, structures not to be touched during a remediation procedure. Force-feedback virtual fixtures designed can improve the economy, speed and accuracy of user motions in the tele-operated environment. In particular, forbidden-region fixtures (FRFs) driven by haptic rendering information obtained from depth-enabled camera(s) can be used in two feedback control paths in this co-robotic system: by the tele-operator, and by a position control system forrobot1000.
  • Virtual fixtures around critical parts of a target (e.g., locations that would trigger explosion of the ordnance) can be designated by operator input or through image recognition. For operator designation of a virtual fixture, the tele-operator will specify the boundaries of virtual fixture either using a haptic interface device, or by using mouse, touch screen, or other input on a video display. At any time during operation ofrobot1000, a tele-operator ofrobot1000 can re-define virtual fixtures by drawing on a real time image, or by repeating the virtual fixture designation process. For automatic recognition, an image recognition capability could be used to specify ‘no touch’ zone(s) based on objects detected from images captured bycamera system1020,1022,1024.
  • Effectors at ends of robot arms1010-1017 can be tracked in real time. Using the haptic rendering algorithms described herein, haptic feedback can be provided to the tele-operator, such as pushing back if an effector gets too close to a protected location; i.e., near or within a forbidden region defined by a virtual fixture, such as a forbidden-region fixture. Additionally or instead, if an effector gets too close to a protected location, robot control actions can be modified to lock out certain motions. For example, suppose a forbidden region had been defined that was directly astern fromrobot1000. Then, if the tele-operator ofrobot1000 attempted movement of one or more of arms1010-1017 backwards close to or within the forbidden region, haptic feedback, such as resistance, can be used to inform the tele-operator about the forbidden region. Additionally or instead, backward movements ofrobot1000 can be inhibited while the forbidden region remains directly astern ofrobot1000. In some embodiments, both providing haptic information and dynamic robot responses can be modified; such as both providing resistance to the tele-operator and slowing down motion ofrobot1000 that is near or within a boundary of a virtual fixture. Many other examples are possible as well.
  • An Example Computing Network
  • FIG. 11A depicts anetwork1100 in accordance with an example embodiment. InFIG. 11A,servers1108 and1110 are configured to communicate, via anetwork1106, withclient devices1104a,1104b, and1104c. As shown inFIG. 11A, client devices can include apersonal computer1104a, a laptop computer1104b, and a smart-phone1104c. More generally, client devices1104a-1104c(or any additional client devices) can be any sort of computing device, such as a workstation, network terminal, desktop computer, laptop computer, wireless communication device (e.g., a cell phone or smart phone), and so on.
  • Thenetwork1106 can correspond to a local area network, a wide area network, a corporate intranet, the public Internet, combinations thereof, or any other type of network(s) configured to provide communication between networked computing devices.Servers1108 and1110 can share content and/or provide content to client devices1104a-1104c. As shown inFIG. 11A,servers1108 and1110 are not physically at the same location. Alternatively,servers1108 and1110 can be co-located, and/or can be accessible via a network separate fromnetwork1106. AlthoughFIG. 11A shows three client devices and two servers,network1106 can service more or fewer than three client devices and/or more or fewer than two servers.
  • An Example Computing Device
  • FIG. 11B is a block diagram of anexample computing device1120 including user interface module1121, network-communication interface module1122, one ormore processors1123, anddata storage1124, in accordance with embodiments of the invention.
  • In particular,computing device1120 shown inFIG. 11A can be configured to perform one or more functions of client devices1104a-1104c,networks310,1106,1720,servers1108,1110,computing devices320a,320b,1724,robot1000,1784, software1310, task-management software, master console1710,camera1780,1800,depth sensor1782,sensor system1860,2600,2630,system2200,2300,metal detector2420,2460, and/ormetal detector system2500,2520,2540, and/or to implement part or all ofarchitecture1300,pipeline1400, remotevirtual environment1704, and/orscenario2800.Computing device1120 can include a user interface module1121, a network-communication interface module1122, one ormore processors1123, anddata storage1124, all of which may be linked together via a system bus, network, orother connection mechanism1125.
  • Computing device1120 can be a desktop computer, laptop or notebook computer, personal data assistant (PDA), mobile phone, embedded processor, or any similar device that is equipped with at least one processing unit capable of executing machine-language instructions that implement at least part of the herein-described techniques and methods, including but not limited tomethod1200 described in more detail below with respect toFIG. 12.
  • User interface1121 can receive input and/or provide output, perhaps to a user. User interface1121 can be configured to send and/or receive data to and/or from user input from input device(s), such as a keyboard, a keypad, a touch screen, a computer mouse, a track ball, a joystick, and/or other similar devices configured to receive user input from a user of thecomputing device1120. User interface1121 can be configured to provide output to output display devices, such as, one or more cathode ray tubes (CRTs), liquid crystal displays (LCDs), light emitting diodes (LEDs), displays using digital light processing (DLP) technology, printers, light bulbs, and/or other similar devices capable of displaying graphical, textual, and/or numerical information to a user ofcomputing device1120. User interface module1121 can also be configured to generate audible output(s), such as a speaker, speaker jack, audio output port, audio output device, earphones, and/or other similar devices configured to convey sound and/or audible information to a user ofcomputing device1120. As shown inFIG. 11B, user interface module1121 can be configured withhaptic interface1121athat can receive inputs related to a virtual tool and/or a haptic interface point (HIP), a remote device configured to be controlled byhaptic interface1121a, and/or other inputs, and provide haptic outputs such as tactile feedback, vibrations, forces, motions, and/or other touch-related outputs.
  • Network-communication interface module1122 can be configured to send and receive data overwireless interfaces1127 and/orwired interfaces1128 via a network, such asnetwork1106. Wired interface(s)1128, if present, can comprise a wire, cable, fiber-optic link and/or similar physical connection to a data network, such as a wide area network (WAN), a local area network (LAN), one or more public data networks, such as the Internet, one or more private data networks, or any combination of such networks. Wireless interface(s)1127 if present, can utilize an air interface, such as a ZigBee, Wi-Fi, and/or WiMAX interface to a data network, such as a WAN, a LAN, one or more public data networks (e.g., the Internet), one or more private data networks, or any combination of public and private data networks.
  • In some embodiments, network-communication interface module1122 can be configured to provide reliable, secured, and/or authenticated communications. For each communication described herein, information for ensuring reliable communications (i.e., guaranteed message delivery) can be provided, perhaps as part of a message header and/or footer (e.g., packet/message sequencing information, encapsulation header(s) and/or footer(s), size/time information, and transmission verification information such as CRC and/or parity check values). Communications can be made secure (e.g., be encoded or encrypted) and/or decrypted/decoded using one or more cryptographic protocols and/or algorithms, such as, but not limited to, DES, AES, RSA, Diffie-Hellman, and/or DSA. Other cryptographic protocols and/or algorithms can be used as well as or in addition to those listed herein to secure (and then decrypt/decode) communications.
  • Processor(s)1123 can include one or more central processing units, computer processors, mobile processors, digital signal processors (DSPs), GPUs, microprocessors, computer chips, and/or other processing units configured to execute machine-language instructions and process data. Processor(s)1123 can be configured to execute computer-readable program instructions1126 that are contained indata storage1124 and/or other instructions as described herein.
  • Data storage1124 can include one or more physical and/or non-transitory storage devices, such as read-only memory (ROM), random access memory (RAM), removable-disk-drive memory, hard-disk memory, magnetic-tape memory, flash memory, and/or other storage devices.Data storage1124 can include one or more physical and/or non-transitory storage devices with at least enough combined storage capacity to contain computer-readable program instructions1126 and any associated/related data structures.
  • Computer-readable program instructions1126 and any data structures contained indata storage1126 include computer-readable program instructions executable by processor(s)1123 and any storage required, respectively, to perform at least part of herein-described methods, including but not limited tomethod1200 described below with respect toFIG. 12,method1600 described below with respect toFIG. 16, and/ormethod2200 described below with respect toFIG. 22. Computer-readable program instructions1126 can include instructions that when executed by processor(s)1123 to perform the herein-described functionality of software, such as, but not limited, to software1310 and/or task-management software.
  • An Example Method for Six Degree of Freedom Haptic Rendering
  • FIG. 12 is a flowchart depicting example functional blocks of anexample method1200.Method1200 begins atblock1210, where a computing device can receive first depth data about an environment, such as discussed above in detail in the context of at least Table 1 andFIGS. 3A and 3B.
  • Atblock1220, the computing device can generate a first plurality of points from the first depth data, such as discussed above in detail in the context of at least Table 1 andFIGS. 3A-4C.
  • In some embodiments, such as discussed above in the context of at least Table 1 andFIGS. 3A-4C, the first depth data can include data about an input plurality of points in the environment. Then, determining the first plurality of points can include: generating a filtered plurality of points by filtering the input plurality of points using a bilateral filter; and determining a normal vector for each point in the filtered plurality of points.
  • In particular of these embodiments, the computing device can include a graphics processing unit (GPU). Then, determining the normal vector for each point in the filtered plurality of points can include determining the normal vector for each point in the filtered plurality of points using the GPU.
  • Atblock1230, the computing device can determine a virtual tool, such as discussed above in detail in the context of at least Table 1 andFIGS. 3A,5A-5D,7,8, and9. The virtual tool can be specified in terms of a translation component for the virtual tool and a rotation component for the virtual tool. In some embodiments, the translation component can be specified in terms of up to three degrees of freedom, the rotation component can be specified in terms of up to three degrees of freedom distinct from the three degrees of freedom used for the translation component, and thus the virtual tool can be specified in terms of up to six degrees of freedom.
  • In other embodiments, the translation component can relates to a position of the virtual tool and the rotation component can relate to a rotation of the virtual tool about at least one axis. In still other embodiments, determining the virtual tool can include representing the virtual tool using a plurality of voxels, where each voxel has a voxel size in each of three dimensions, and where the voxel size for each of the three dimensions is a same size.
  • Atblock1240, the computing device can determine a first force vector between the virtual tool and the first plurality of points, such as discussed above in the context of at least Table 1 andFIGS. 3A,5A-5D,6,7, and10.
  • In some embodiments, determining the first force vector can include determining a bounding box for the virtual tool based on a projection of the virtual tool onto the first depth image; and determining neighboring points of the first plurality of points, wherein each neighboring point is within the bounding box. In particular of these embodiments,method1200 can also include determining whether there are zero neighboring points or more than zero neighboring points; and in response to determining that there are zero neighboring points: determining that the virtual tool is in a free-motion state, and moving the virtual tool in a direction based on the translation component and by a distance bounded by a same voxel size.
  • In other particular of these embodiments,method1200 can also include: in response to determining that there are more than zero neighboring points: determining a depth for each neighboring point with respect to the virtual tool; determining in-contact neighboring points from the neighboring points, where each in-contact neighboring point has a corresponding depth of zero; determining in-collision neighboring points from the neighboring points, where each in-collision neighboring point has a corresponding depth less than zero; in response to determining that no in-collision neighboring points are determined and to determining at least one in-contact neighboring point is determined: determining a first constrained acceleration for the virtual tool based on the at least one in-contact neighboring point, and moving the virtual tool in a direction based on the first constrained acceleration.
  • In still other particular of these embodiments,method1200 can also include: in response to determining that at least one in-collision neighboring point is determined: determining a second constrained acceleration for the virtual tool based on the at least one in-collision neighboring point; and moving the virtual tool in a direction based on the second constrained acceleration.
  • Atblock1250, the computing device can send a first indication of haptic feedback based on the first force vector, such as discussed above in the context of at least Table 1 andFIGS. 3A,5A-5D,6,7, and10.
  • In some embodiments, such as discussed above in the context of at least Table 1,method1200 can include: receiving second depth data about the environment at the computing device, where the second depth data can differ from the first depth data; generating a second plurality of points from the second depth data using the computing device; determining a second force vector between the virtual tool and the second plurality of points using the computing device; and sending, from the computing device, a second indication of haptic feedback based on the second force vector.
  • In other embodiments, such as discussed above in the context of at leastFIG. 3A, the computing device can be configured to communicate with a controllable mechanism. Then, sending the indication of haptic feedback based on the force vector from the computing device can include sending the indication of haptic feedback based on the force vector from the computing device to the controllable mechanism. In these embodiments,method1200 can also include providing haptic feedback based on the indication of haptic feedback using the controllable mechanism. In particular of these embodiments, the controllable mechanism can include a manipulator configured for surgical operation.
  • In still other embodiments, such as discussed above in the context of at leastFIG. 3A, the first depth data can include first image data and corresponding first depth values. In these embodiments,method1200 can also include generating a virtual environment based on the first image data and corresponding depth values. In particular of these embodiments, such as discussed above in the context of at least Table 1 andFIGS. 4A-4C, the virtual environment can include a first object. In these embodiments,method1200 can also include determining a contact between the first object and the virtual tool.
  • In even other embodiments,method1200 can also include defining a forbidden region within the environment. Then, determining the first force vector between the virtual tool and the first plurality of points can include inhibiting the virtual tool from moving within the forbidden region. In yet even other embodiments,method1200 can also include controlling a motion of an object or character within a virtual environment, wherein the virtual environment comprises at least one virtual object and at least one real surface.
  • Virtual Fixtures for Human/Autonomous Manipulation Tasks
  • In haptic interaction, an operator can use a physical haptic device to interact with objects in a virtual environment. This virtual environment can consist of representations of physical objects, as captured by sensor data, virtual objects that are entirely computer simulated, or combinations of representations of physical objects and virtual objects. A Haptic Interface Point (or HIP) can be thought of as the position of a 3-DOF end effector in virtual space that matches the position of the haptic device. For 6-DOF end effectors, a virtual tool can represent the end effector. When the user moves the haptic device, the HIP or virtual tool can move accordingly in the virtual environment.
  • Depth data can be specified as a point cloud can be thought of as an unstructured set of points representing the boundary of physical objects. Depth data can be obtained using stereo cameras, laser scanners or depth cameras in which case the data will be sorted in the image plane. A depth image can augment the depth data with color information. Haptic rendering as described herein can utilize streaming point cloud data represented in depth images for 3 DOF and 6 DOF applications.
  • Complex telerobotic manipulation tasks require high precision and repetitive actions for which automated systems are highly skilled. However, it also requires critical judgment and decision-making, which is best done by a human operator. Thus, herein is described a combined manipulator system (human+automation) that affords the operator a great amount of freedom but also enhances performance via computer-generated auditory-, visual- and haptic-feedback. The resulting technology will improve the outcome by preventing mistakes and increasing efficiency in manipulation tasks. Rather than using the feedback to solve the task (by forcing the human operator along a pre-defined trajectory) we use the feedback to communicate “intent” of the computer system during sequential operations.
  • The feedback can be related to “virtual fixtures” added to the virtual environment provided to the operator utilizing haptic feedback to operate effector(s) in a remote environment, such as an operator performing tasks such as remote search and rescue, sensor placement, ordnance placement/removal, remote movement of materiel, and remote construction. The remote environment can be hazardous; e.g., a cleanup site, a search and rescue location, an area with extreme weather conditions, an underwater environment, an outer space environment. Underwater applications of this technology include human-guided surveillance, underwater equipment maintenance and repair, environmental cleanup as well as tasks in the oil and gas industries. In the latter, commercial interest is high, given that expensive equipment is manipulated in time-critical environments, where mistakes can be both fatal and disastrous to the environment.
  • A virtual fixture can be defined as an overlay of abstract sensory information on a workspace in order to improve the telepresence of a remotely manipulated task. Virtual fixtures can be arbitrary in both appearance and feedback. In some scenarios, virtual fixtures can act as one or more virtual rulers guiding a remote manipulator along a pre-defined path. The operator can be notified of deviations from the path defined by the virtual fixture using audio, visual or haptic feedback in particular. Haptic feedback (force feedback applied to the operator control stick) can be used to guide the operator (and hence the remote end effector) back to the desired path.
  • Virtual fixtures can also include forbidden-region fixtures and guidance fixtures. As mentioned above, a forbidden-region fixture can define an area represented by the point data where the virtual tool is not permitted to enter; i.e., the forbidden region. A guidance fixture can define an area represented by the point data where the virtual tool is permitted, and perhaps encouraged, to enter.
  • Virtual fixtures, such as but not limited to forbidden-region fixtures and guidance fixtures, can be specified based on hard constraints that prevent end effector violation of constraints, or soft constraints where undesired motions are resisted but not prevented, using an associated penetration-based penalty cost. An operator can be notified about virtual fixtures using any combination of visual-, auditory-, and haptic feedback. For instance, the haptic feedback can take the form of repelling force feedback, vibrotactile feedback, or rendered as a viscous resistance.
  • Forbidden-region fixtures and guidance fixtures can be part of a set of standardized virtual fixtures with well-defined visual-, auditory-, and haptic properties. This standard set of virtual fixtures can be used to specify complex manipulation tasks in terms of common manipulation subtasks. Example subtasks include:
      • Grasping—This virtual fixture will be used to aid the operator in positioning the remote manipulator for a successful grasp of an object of interest.
      • Rotational manipulation and translational movements—This guidance virtual fixture will be used for valve turning and is important to make sure that no equipment is damaged during operation.
      • Insertion/Removal—Many operations requires coupling/decoupling of connectors or subsea equipment, so called hot stabbing, and it is therefore important to aid the operator in for instance linear insertions/removals.
      • Placement—This virtual fixture will aid in operations where equipment needs to be positioned accurately with respect to the environment.
  • The virtual fixtures can be defined with respect to absolute location(s) in task space and implicitly defined by data from sensors in the remote environment. The virtual fixture can adapt to moving targets and that external disturbances automatically will be rejected. For example, an underwater current can move, and thus disturb, an underwater manipulator. In another example, an external disturbance in a surgical context can occur when a patient moves (or is moved) during a surgical procedure. Many other examples are possible as well.
  • Haptic feedback can be generated based on data from non-contact sensors such as depth images including data from radar or sonar sensors. Using non-contact sensors, haptic feedback can be generated before contact has been made using 3D images of objects. This can be done before physical contact with obstacles thus avoiding loop delays in operation. In addition, virtual fixtures can be defined around objects depicted in the depth images.
  • By sensing range and effectively geometries near the manipulator, a guidance fixture can guide the operator to a specific location or region and/or as a forbidden region fixture to block the operator from a specific location or region. Virtual fixtures can be implemented to account for in-motion objects in the environment. Complex tasks, particularly sequential tasks, can benefit from multiple, adaptive virtual fixtures. The virtual fixtures can change as tasks in a series of sequential tasks are completed—for example, a virtual fixture can start as a guidance fixture for guiding operator to a region to perform a task and then, once the task is complete, the virtual fixture can change to a forbidden-region fixture to keep the operator from the region to avoid undoing previously-completed work. The herein-described system is capable of rendering multiple virtual fixtures and switching between them in an intuitive, fluid fashion in real time, in response to the environment and performance needs.
  • Virtual fixtures can aid the human operator in complex manipulation tasks. These tasks can be planned and carried out using automated and sequential transitioning between virtual fixtures. To perform part or all of the tasks, the operator can use a manipulation system with haptic rendering capabilities. During execution of the tasks, a level of automation can change from manual operation to fully autonomous operation, perhaps with one or more intervening levels of semi-autonomous operations possible. The human operator can adjust the level of automation and also intervene as necessary during autonomous operation. For example, during human-only or mostly-human operation, a flexible guidance fixture can be used that allows for small deviations from a guided path, where the deviations can be subject to guiding but non-constrained feedback, so to provide the operator increased flexibility and control over the situation.
  • While the operator directs the system, the operator can also respond to system performance and conditions. Auditory, visual, and haptic virtual fixtures can communicate intent of the system to the operator. For example, virtual fixtures can be used to describe basic tasks related to manipulation such as positioning, grasp and screwing/unscrewing tasks. A set of virtual fixtures can be added to the environment to avoid mistakes, such as a forbidden-region fixture, a guidance fixture, a standoff fixture, which is a virtual fixture for maintaining a fixed safety distance between the manipulator and an arbitrary object, and/or a velocity-limited fixture, which is a virtual fixture limiting the maximum velocity of the manipulator.
  • FIG. 13 is a block diagram ofarchitecture1300 for rendering adaptive virtual fixtures, in accordance with an example embodiment. A computing device,such computing device1120 executing software1310 can manage the virtual fixtures in a virtual environment based on data from a remote environment. Software1310 can process and generate the virtual fixtures in real-time based on data in live sensor feed1332 from video and range sensor package1330 in the remote environment. Sensor package1330 can be mounted on a device in the remote environment; e.g., aboard a remote manipulator or robot.
  • Software1310 can render virtual fixtures implicitly defined by data inlive sensor feed1332. That is, software1310 can track and adapt virtual fixtures to changes in the task space; e.g., as tasks are added, completed, updated, and removed; and to changes in the remote environment; e.g., as objects move or change shape. Software1330 can communicate position commands1342 fromhuman operator1340 and/orplans1322 from high-level planner1320 toremote manipulator1350; e.g., as desiredposition instructions1354.
  • Remote manipulator1350 can reportactual position information1352 to software1310. Software1310 can receiveactual position information1352 and data from live sensor feed1332 to generatefeedback1344 tohuman operator1340 about virtual fixtures and perhaps real objects in the remote environment. As software1310 determines tasks are completed and/or determines changes in the remote environment from data inlive sensor feed1332, software1310 can providechanges1324 to plans maintained byhigh level planner1320. Further, software1310 can autonomously provide desiredposition information1354 in some circumstances; e.g., positions to be reached while following a guidance feature, to avoid a collision, and/or in response to reaching forbidden region virtual fixture.
  • Table 2 below illustrates how virtual fixtures can be used in semi-autonomous mode (with both a human and an automated control loop) and in fully autonomous operation. In the latter, system intent is still communicated to the operator using our haptic rendering algorithms, thereby providing an intuitive method for the operator to supervise (but not control) the task. Conversely, the virtual fixtures can be used to communicate the outcome of the operation in manual/human control.
  • TABLE 2
    Level of AutomationRole of Virtual Fixtures
    Fully autonomousDuring fully autonomous operation, virtual fixtures operated by
    operationsoftware 1310 can communicate the intent of the
    autonomously-operating system to the human operator. The intent
    of the system can be shown with visual cues/displays (e.g.,
    highlighting target position(s) in the workspace), auditory signal(s),
    and/or haptic feedback.
    Semi-autonomousDuring semi-autonomous operation, the system can use the
    operationexperience and cognition of the human operator linked with the
    precise spatial awareness of the computer system.
    Software 1310 can let the operator make decisions within ranges of
    boundaries determined by likelihood estimates. The human
    operator can change the ranges of boundaries to effectively modify
    the level of automation. Controlling ranges of boundaries for
    decision gives the human operator an intuitive control to the level
    of automation during semi-autonomous operation.
    The human operator can make higher-level decisions; i.e. ability to
    change a planned path within the range of boundaries, or to
    intervene to avoid impending difficulties. Virtual fixtures can
    reside on a lower level such that a positioning task always will be
    made exactly, and such that unintended collisions with the
    manipulator are prevented, mainly using force feedback.
    Manual operationDuring manual operation, all manipulation tasks can be done
    manually; e.g., without virtual fixtures. Instead of completely
    turning off the virtual fixtures, the virtual fixtures can operate in the
    background without sending feedback to the human operator.
    Virtual fixtures can communicate the outcome of the manual
    operator commands to the temporarily deactivated autonomous
    system. By keeping the autonomous system up to date, the amount
    of time taken during switchover from manual to (semi-)
    autonomous operation can be reduced.
  • Operator1340 and/or software1310 can vary the level of automation. For example, simpler manipulation tasks may be controlled by software1310 in the fully autonomous mode withoperator1340 providing position commands1342 that reflect high-level decisions. If an unusual or exceptional event occurs (hence complexity increases),operator1340 can reduce the level of automation. Combining the experience and skill of a human operator with the spatial awareness of the manipulation system, along with an adjustable autonomy framework, can both increase efficiency and reliability of operations and allow execution more complex tasks than would otherwise be feasible with traditional manual or autonomous solutions.
  • A complex operation can be planned using standardized virtual fixtures; e.g., the operation can be specified as a sequence or other set of tasks, with each task shaped by virtual fixtures. When defined this way, both the human operator and the autonomous system will be able to readily interpret this plan. Additionally, since the autonomous system also tracks the outcome of each task, procedural compliance can be observed. If important steps are skipped, the human operator can be notified and/or other actions taken; e.g., remote actuators can be put into hibernation or a reduced-performance mode until the correct task is begun. This information can further be used to generate log files that are sparse representations of already completed tasks. Log files can later be used to generate performance statistics and maintenance schedules.
  • The system can support flexible planning, allowing for some steps to be loosely defined (a necessary feature when operating in un-structured or dynamic environments). For instance, the system might be aware that a lever needs to be pulled, but not which one. When the uncertain part of the plan is reached, the human operator will be asked to guide the system to the right lever using a point-and-click interface.
  • To best perform the task, the human operator should not perceive that the virtual fixtures are taking over completion the task and that the human is just riding along with an automated system. If the human operator is not sufficiently engaged, the human operator might be overly passive and perhaps reduce focus and performance. For forbidden-region virtual fixtures, preventing passivity is straightforward since forces are only sent when an operator is attempting entry into a forbidden region. That is, haptic feedback can be reasonably expected once the operator realizes the interaction with the forbidden region.
  • For a guidance virtual fixture, however, the human operator may not always anticipate the activation of a guidance force. Thus, virtual fixtures can use auditory and visual feedback, in addition to haptic to increase the operator's situational awareness during manipulation tasks. This multi-sensory feedback can give the human operator a clear indication when a guidance force is about to be activated. For instance, visual virtual fixture(s) can provide visual indications and/or auditory virtual fixture(s) can emit sounds to indicate regions in task space in which the guidance force will be activated.
  • On-the-fly definition of virtual fixtures can enhance the high-level architecture of planned operations constructed by standardized virtual fixtures. For this, a combined range and video sensor can be used in the system. The sensor can provide data at a first rate; e.g., at least 30 Hz, and based on the sensor data, force feedback can be generated at a second, faster rate; e.g., at least 1000 Hz. The system can utilize processors, such as general-purpose graphics processing (GPGPU) hardware and a customized parallelized low-level GPU, to process the sensor data at least the first rate and generate force feedback at the second rate.
  • FIG. 14 is a block diagram ofpipeline1400 for parallel processing and generation of virtual fixtures, in accordance with an example embodiment.Pipeline1400 can begin atblock1410, where range and video image data can be captured. For example, a computing device carrying out operations ofpipeline1400 can receive one or more depth image atblock1410, use cameras and/or sensors to obtain range and video image data, or otherwise obtain range (depth) and video image data.
  • Atblock1420, a point cloud can be generated from the range and video image data. The point cloud can be a collection of points in a frame of reference, such as a collection of Cartesian points in a three-dimensional space where the points correspond to objects whose images are captured atblock1410. In some embodiments, other data structures, such as depth images, can be utilized and/or generated atblocks1410 and1420.
  • Atblock1430, a bilateral filter can be applied. Bilateral filters can be used to smooth data within the point cloud by weighted a value associated with point; e.g., a depth value, with a weighted average of values from nearby points. For example, the points can be weighted using a Gaussian kernel as a function of Euclidean distances in a Cartesian frame. In some embodiments, block1430 can be omitted. In other embodiments, other filters can be applied instead of or along with the bilateral filter.
  • In some embodiments, normal values can be determined for points in the point cloud. For example,pipeline1400 can be utilized to carry out part or the entire example haptic interaction algorithm for 6-DOF haptic rendering discussed above in the context of Table 1.
  • Atblock1440, the image obtained atblock1410 can be segmented. That is, each pixel in the video image can be assigned a label. Each label can represent visual characteristics; e.g., a range of colors, a location on a boundary (or within the interior) of an object, etc. Two pixels assigned to the same label have the same visual characteristics.
  • Atblock1450, objects can be recognized from the segmented image. In some embodiments, recognized objects can be associated with collections of points within the point cloud. For example, suppose an apple is recognized in the segmented image and that the apple is depicted in a set of pixels SP of the segmented image. Let SPt be the set of points in the point cloud associated with pixels SP. Since SP is associated with the apple, then SPt can be associated with the apple as well.
  • Atblock1460, virtual fixtures can be defined. The virtual fixtures can be defined with respect to the object(s) recognized atblock1450. To continue the example fromblock1450, a virtual fixture VFa can be associated with SP and SPt, and thus be associated with the recognized apple. Then, if the apple moves in a later image captured atblock1410, the apple can be re-recognized in the later image atblock1450, and virtual fixture VFa reassociated with the moved apple. Thus, virtual fixtures can be dynamically associated atblock1460 with objects in an environment captured in the range and video image data obtained atblock1410 and recognized atblock1450.
  • In other embodiments, the virtual fixtures can be specified in terms of geometry in a remote environment and/or in a virtual environment. For example, a virtual object can be introduced into the virtual environment; e.g., a virtual cube or sphere. Then, a virtual fixture can be associated with the virtual object. As another example, the virtual fixture can be defined with respect to a geometric region, volume, or other entity in either the remote or virtual environment without respect to another object. Other examples ofpipeline1400 are possible as well.
  • In some scenarios, the captured range and video image data is insufficient; e.g., in cases of significant measurement noise or sensor occlusion. Measurement noise may be suppressed using filtering such as discussed above with respect to block1430, To account for sensor occlusion, occlusion by a remote manipulator can be predicted based on detailed models of the remote manipulator, kinematic equations, and data about manipulator positioning. Then, sensor data can be rejected that correspond to regions predicted to be occluded by the remote manipulator. In the rejected regions, sensor data can be interpolated using previously captured data. Further, by predicting a position of the remote manipulator for occlusion, the predicted position can also be used to verify the state of the manipulator during runtime, and so enhance operational safety.
  • FIG. 15 depicts anenvironment1500 with a robot arm and multiple virtual fixtures, including forbidden-region fixtures (FRFs)1530,1532 and guidance fixtures (GFs)1534,1536 in accordance with an example embodiment.Environment1500 is a virtual environment containing representations of real objects, such asrobot arm1510 and toxic material container, and representations of virtual objects, such asvirtual fixtures1530,1532,1534,1536,1538.
  • In some embodiments, some or all ofvirtual fixtures1530,1532,1534,1536,1538 can be associated with a task list. Table 3 below shows an example task list related toenvironment1500 forrobot arm1510, where the task list is associated withvirtual fixtures1530,1532,1534,1536,1538 as also shown in Table 3.
  • TABLE 3
    Task ListFRF 1530FRF 1532GF 1534GF 1536GF 1538
    1. Go to ValveInitially FRF.Initially FRF.Use GF toInitiallyInitially
    1524EnableEnablemoveguide awayguide
    entry/exitentry/exittowardfrom valveaway from
    based on validbased onvalidvalve 15241526material
    authorizationauthorization
    1522
    2. ObtainIf valid
    Authorization toauthorization
    Enter Regionprovided
    1532(during task),
    deactivateFRF
    3. Turn ValveUponUpon
    1524completioncompletion
    of task,of task,
    change tochange to
    guide awayguide
    from valvetowardvalve
    15241526
    4. Go to ValveUse GF to
    1526move
    towardvalve
    1526
    5. Turn ValveUpon
    1526completion
    of task,
    change to
    guide away
    fromvalve
    1526
    6. Open DoorDoor Handle/path to open door may be controlled by guidance fixture not
    1528shown in FIG. 15. This fixture may be activated by completion ofTask 5
    and/or deactivated by completion ofTask 6.
    7. Enter AreaPathway toarea 1534 can be provided by guidance fixture notUpon
    1530shown in FIG. 15. This fixture may be activated bycompletion
    completion ofTask 6.of task,
    change to
    guide
    toward
    material
    1522
    8. Go to ToxicUse GF to
    Material 1522move
    toward
    material 1522
    9. Get ToxicUpon
    Material 1522completion
    of task, GF
    can change
    to guide
    away from
    material
    1522
    10. Remove ToxicUse GF to
    Material frommove
    Region
    1532away from
    material
    1522.
    Other GFs
    can be
    activated.
    11. Wait for
    Transport from
    outside Region
    1530
    12. Put ToxicGuidance features not shown in FIG. 15 can guiderobot arm 1510 toward
    Material ontransport. In some embodiments, part of guidance feature 1530 can be
    Transportdeactivated to letrobot arm 1510 and/or transport work together.
    13. Close DoorGuidance features not shown in FIG. 15 may guiderobot arm 1510 todoor
    1528and to close door (similar to those for Task 6).
    14. LeaveDeactivateActivate FRF
    Perimeter RegionFRF untilonce robot arm
    1530robot arm 1510leaves region
    leavesregion1532
    1530.
  • Each ofvirtual fixtures1530,1532,1534,1536,1538, is shown inFIG. 15 associated with a number. For example,virtual fixtures1530 and1532 have associatedrespective indicators1550 and1552 showingrespective numbers 2 and 12, while each ofvirtual fixtures1534,1536,1538 is shown displayingrespective numbers 1, 4, and 8. In the environment shown inFIG. 15, where tasks in the task list of Table 3 have not yet begun, each number correspond to the next task that the virtual fixture is associated with. So,virtual fixture1534 is associated withTask 1 of the task list,virtual fixture1532 is associated withTask 2, and so on. The indicator can be updated as tasks complete; e.g., upon completion ofTask 1, the virtual fixture indicator forvirtual fixture1534 can be updated to refer toTask 3, which is the succeeding task for that virtual fixture. In some embodiments, an indicator of a virtual fixture can indicate multiple tasks associated with the virtual fixture; e.g., the next N tasks, with N>1, or all tasks referring to the virtual fixture. In other embodiments, upon completion of the last task involving a virtual fixture, an indicator for the virtual fixture can change to “Complete” or a similar indication of completion. In still other embodiments, upon completion of the last task involving a virtual fixture, the virtual fixture can be removed fromenvironment1500.
  • In another scenario, consider a manipulation task where two valves need to be turned, such asvalves1524 and1526. In this scenario, the order in which the valves are turned is irrelevant. The human operator can seevalves1524,1526 on a display with both highlighted as valid targets. In addition to the valves the operator also sees virtual cones; e.g.,guidance fixtures1534 and1536, extending out from each handle. Since the computersystem providing environment1500 cannot be sure which valve the human operator will approach first, no force feedback is given. Instead, if the operator commands the manipulator into the cone extending outvalve1524, force feedback can be automatically activated to guide the manipulator into a perfect configuration for a valve turn. That is, the human operator makes the high level decision and the computer system makes sure that it is being executed with high accuracy. To the operator the whole procedure will feel effortless, just as if the manipulator ‘just happened’ to slide into the right configuration by accident. Similar user interfaces exist in many document processors, where the user drags a figure to roughly the right position in the document and the software automatically aligns it with the margins.
  • Virtual fixtures can be programmatically controlled. For example, task-management software can control one or more associated virtual fixtures. In some embodiments, the task-management software can be part of software1310 discussed above in the context ofFIG. 13.
  • The task-management software can include software that manages tasks on a task list with associated virtual fixtures such as shown in Table 3. The task-management software can allow the operator to define a task list, define virtual fixtures as needed for the task list, define tasks on the task list, and then aid the operator and associated remote devices; e.g.,robot arm1510 ofFIG. 15, to carry out the tasks. The task-management software can ensure that tasks on the task list are carried out in order. In some embodiments, a task can be defined in terms of multiple operators acting in parallel.
  • The task-management software can use the task list in combination with real-time sensor data to automatically transition between the different virtual fixtures, such as the transition between grasping and turning a valve. In cases where multiple choices are possible, the task-management software will have the ability to let the operator decide while in semi-autonomous levels of automation. In cases when no operator is present and/or in a fully autonomous level of operation, the task-management software can use a number of techniques to make a decision, such as conditional probabilities and maximum likelihood votes. The task-management software can avoid undesirable transients that can arise from virtual fixture transitions as well as from task state changes by careful matching of terminal and initial conditions of successive control epochs as well as enforcement of protective constraints on control states and outputs.
  • The task-management software, in addition to serving as an interface and control system, can constantly monitor operations toward task completion and continually tracking of operation outcomes with associated likelihoods. The task-management software can adaptively generate and transition between virtual fixtures identified for a given task. Although the task-management software can execute the transitions between virtual fixtures autonomously, autonomous operation is not always desirable. Rather, a level of automation can be modified, either by the computer system or by the operator as discussed above in the context ofFIG. 13 and Table 2.
  • As another example, some or all virtual fixtures in a virtual environment can be associated with a script or other software that controls part or all of the operation of the virtual fixture. The script can include instructions, such as but not limited to instructions that can operate based on conditions of the virtual fixture; e.g., a type of virtual fixture (forbidden-region fixture, guidance fixture), conditions in the environment; e.g., time, temperature, weather conditions, water currents, wind condition, chemical conditions, biological conditions, nuclear conditions, sensory input related to conditions in the environment, task lists and/or tasks, and/or other conditions. The instructions can change the virtual fixture, perhaps based on a condition in the environment.
  • For example, a forbidden-region fixture that allows entry to a forbidden region to a device D if proper authorization is provided by D at time of entry can have a virtual-fixture script such as shown in Table 4 below.
  • TABLE 4
    ID = myID( );
    send_authentication_request( );
    R = get_authentication_response( );
    If (valid_entry(ID, R)) then
     Open(ID);
  • The virtual-fixture script can be invoked when the entry to the forbidden region associated with the virtual fixture is sought, when an object comes in contact with the virtual fixture, or at some other time. The virtual-fixture script first assigns an identifier “ID” to the virtual fixture using a myID( ) function—in some embodiments, this can be performed prior to script execution. Then, an authentication request can be sent and a response “R” to the authentication request received. The virtual-fixture script then determines if R provides valid entry to the virtual fixture using the valid_entry( ) function—if R does provide valid entry to the forbidden region associated with the virtual fixture, then the virtual fixture is opened using the open( ) function. Virtual-fixture scripts can have other characteristics; e.g., a virtual-fixture script can be written in other computer languages, can utilize many other functions, can be compiled and/or interpreted, and/or can perform other operations. Many virtual-fixture scripts are possible as well.
  • FIG. 16 is a flowchart ofmethod1600, in accordance with an example embodiment.Method1600 can begin atblock1610, where a computing device can receive first depth data about an environment. Atblock1620, the computing device can generate a first plurality of points from the first depth data. Atblock1630, the computing device can determine a haptic interface point.
  • Atblock1640, the computing device can define a virtual fixture for the environment. In some embodiments, wherein determining the virtual fixture for the environment can include: determining one or more objects in the environment from the first depth data; determining a first object of the one or more objects, the object having a first location in the environment; and associating the virtual fixture with the first object and the first location.
  • In other embodiments, the virtual fixture is at least one fixture selected from the group of fixtures consisting of a forbidden-region fixture and a guidance fixture. In particular of the other embodiments, the virtual fixture can include a forbidden-region fixture. Then, determining the first force vector can include: determining an initial force vector between the HIP and the first plurality of points; determining whether the HIP is within a forbidden region defined by the forbidden-region fixture; and in response to determining that the HIP is within the forbidden region: generating a forbidden-region-force vector away from the forbidden region, and determining the first force vector based on the initial force vector and the forbidden-region-force vector.
  • In other particular of the other embodiments, the virtual fixture can include a guidance fixture. Then, determining the first force vector comprises: determining an initial force vector between the HIP and the first plurality of points; determining whether the HIP is within a guidance region defined by the guidance fixture, wherein the guidance region comprises a target; and in response to determining that the HIP is within the guidance region: generating a guidance-region-force vector toward the target; and determining the first force vector based on the initial force vector and the guidance-region-force vector.
  • In still other embodiments, defining the virtual fixture for the environment can include: determining one or more instructions related to the virtual fixture and defining the virtual fixture based on the one or more instructions. In particular of the still other embodiments, the one or more instructions can be configured to change the virtual fixture based on a condition in the environment. In more particular of the still other embodiments, the condition in the environment comprises a condition selected from the group of conditions consisting of a time condition, a temperature condition, a weather condition, a water-current condition, a wind condition, a chemical condition, a biological condition, and a nuclear condition.
  • Atblock1650, the computing device can determine a first force vector between the haptic interface point and the first plurality of points. The first force vector is based on the virtual fixture. In some embodiments, the virtual fixture can include a standoff fixture. The standoff fixture can define a target and a safety region. The safety region can be associated with the target. Then, determining the first force vector can include: determining an initial force vector between the HIP and the first plurality of points; determining whether the HIP is within the safety region; and in response to determining that the HIP is within the safety region: generating a safety-region-force vector away from the target, and determining the first force vector based on the initial force vector and the safety-region-force vector.
  • In other embodiments, the HIP can have a velocity. The virtual fixture can include a velocity-limited fixture. The velocity-limited fixture can define a velocity-limited region and a maximum velocity. Then, determining the first force vector comprises: determining an initial force vector between the HIP and the first plurality of points; determining whether the HIP is within the velocity-limited region; in response to determining that the HIP is within the velocity-limited region: determining whether the velocity of the HIP exceeds the maximum velocity, and in response to the velocity of the HIP exceeding the maximum velocity, determining a velocity-limiting vector opposing the velocity of the HIP, and determining the first force vector based on the initial force vector and the velocity-limiting vector.
  • Atblock1660, the computing device can send a first indication of haptic feedback based on the first force vector. In some embodiments,method1600 can further include: receiving second depth data about the environment at the computing device, where the second depth data differs from the first depth data; determining one or more objects in the environment from the second depth data; determining whether the one or more objects includes the first object; and in response to determining that the one or more objects include the first object: determining a second location in the environment for the first object, and associating the virtual fixture with the first object and the second location.
  • In other embodiments, the environment can include a virtual tool associated with the HIP. The virtual tool can be defined in terms of a tool-translation component and a tool-rotation component. The target can be configured to be defined in terms of a target-translation component and a target-rotation component. Then,method1600 can further include: in response to determining that the HIP is within the guidance region, aligning the tool-rotation component with the target-rotation component.
  • In still other embodiments,method1600 can further include: determining a task list that can be associated with a plurality of tasks to be performed in the environment, where the task list comprises a first task. Then, determining the virtual fixture can include determining the virtual fixture based on the task list. In particular of these embodiments, determining the virtual fixture based on the task list can include: determining the virtual fixture initially to be a first virtual fixture; determining whether the first task is completed; and in response to determining that the first task is completed, changing the first virtual fixture to a second virtual fixture, wherein the first virtual fixture and the second virtual fixture differ.
  • In some particular of these embodiments, the first virtual fixture can be a guidance fixture and the second virtual fixture can be a forbidden-region fixture. In other particular of these embodiments, the first virtual fixture can be a forbidden-region fixture and wherein the second virtual fixture can be a guidance fixture. In other particular of these embodiments, the task list can further include a second task. The virtual fixture can include a first virtual fixture having a first type of virtual fixture. Then, determining the virtual fixture based on the task list can include: determining whether the first task is completed; in response to determining that the first task is completed, changing the first type of virtual fixture to a second type of virtual fixture, wherein the first type of virtual fixture and the second type of virtual fixture differ; determining whether the second task is completed; and in response to determining that the second task is completed, changing the second type of virtual fixture to a third type of virtual fixture, wherein the second type of virtual fixture and the third type of virtual fixture differ.
  • In some of the other particular of these embodiments, the first type of virtual fixture can be a forbidden-region type of virtual fixture, the second type of virtual fixture can be a guidance-region type of virtual fixture, and wherein the third type of virtual fixture can be the forbidden-region type of virtual fixture.
  • Haptic Rendering Related to Underwater Tasks
  • The herein-described techniques for 3DOF and 6DOF dynamic haptic rendering can be utilized to carry out tasks underwater. Operations in the deep sea, or hazardous marine environments, typically require manipulation from remotely operated vehicles (ROVs). For example, haptic rendering can be used in the context of an omnidirectional mobile robot with underwater grippers; e.g.,robot1000 discussed above in the context of at leastFIG. 10. These underwater tasks can include tasks for search and rescue, underwater infrastructure maintenance, repair, mining, military operations, and waste cleanup.
  • Subsea manipulation is a challenging task, and one that is necessary in a number of applications. Examples include the construction of cabled ocean observatories, biological studies of hydrothermal vent colonies, oil and gas operations, subsea mining, and other activities that require precise movement on the seafloor. In these applications, a human operator's perception is important, especially in the presence of disturbances due to currents, low visibility situations, or other scenarios that require adaptation. Another difficulty in performing manipulation tasks with remote robots arises in making precise motions and interactions in the presence of dynamically changing environments. These include objects or structures that must be avoided or protected. These objects can be moving, and possibly changing in size. For example, in underwater manipulation for repair (e.g., oil platforms) or for biological sampling, precision manipulation is required despite disturbances from water flow and turbulence.
  • Underwater tasks can involve manipulation in the vicinity of sensitive or delicate structures. Visual features, such as scaling and tremor dampening can improve performance. However, operator error, communication delay and limited visual feedback can cause a robot to be accidentally driven into sensitive structures and result in irreparable damage. The problem is further amplified in the absence of haptic feedback to the operator. If the contact is not visually detected by the operator, a resulting collision can cause serious damage.
  • Consider, for example, the common task of mating electrical connections underwater. In principle this is a simple maneuver; however, operators often find it difficult to successfully mate connections using underwater manipulators, such as robot arms including the ARM 5E MINI five-function electric manipulator configured for subsea use. At best, this difficulty results in excessively long duration operations that drive up the cost of subsea operations. At worst, hardware is destroyed, often requiring expensive replacements. Further, operational costs for ROVs can be expensive; e.g., $1 per second or more. Thus, there is great incentive to maximize the efficiency and capability of human/co-robot manipulation tasks in subsea environments.
  • An underwater depth-enabled camera can be used provide depth images for 3DOF and 6DOF haptic rendering to carry out underwater tasks. The depth images can provide haptic feedback to a human operator, giving with a sense of touch along with a visual representation objects seen by the camera. For example, a human operator can use a tele-operated robotic device equipped with the underwater depth-enabled camera to direct robotic removal of ordnance from lake, river or sea bottoms. An added feature of the herein-described techniques is prevention of robot end effector contact with undesirable locations through imposition of virtual fixtures, such as forbidden-region fixtures.
  • In some embodiments, the depth-enabled camera can be used in combination with a metal detector or metal detector array. This combination of devices can provide a human tele-operator with information about both depth (distance to target) and metal profile(s) within the area of operation. The metal detector (array) can be mounted on an underwater craft near manipulator(s), such as robotic arms, to allow 2D or 3D metal profiles to be mapped without using a visual or acoustical detector. An array of small metal detectors can be configured in a cross pattern, arranged around the camera system to allow 2D metal profiling data to be taken simultaneously with the image capture. The metal profile will be generated from the array moving across the area while scanning. An alternate metal detector design is to obtain the metal profile system using a phase array design where radiation pattern of the B field can be shifted by adjusting the phase difference between driving coils inside each metal sensor.
  • In an example system, a depth-enabled camera and metal detector array camera optimized for reconnaissance and surveying can be placed in a Sea Ray-like ROV operable from a surface craft, such as a barge. The ROV can be linked to the surface craft by a tether, ensuring ROV recovery as well as uninterruptible power and data lines. The ROV can be equipped with additional video cameras and lights in addition to the depth-enabled camera and metal detector array. The ROV can provide the data necessary to construct a complete 3D picture of what is ahead of the surface craft and update a 3D underwater map in real time.
  • Providing haptic feedback from underwater sensors to an operator at the surface has the potential to transform subsea manipulation capability. Virtual fixtures will allow manipulators to precisely maneuver in sensitive environments where contact should not be made with surrounding objects or structures. Biological studies of animal colonies around hydrothermal vents could benefit greatly from such a capability—allowing scientists to carefully gather data with unprecedented resolution and proximity to sensitive organisms. Common tasks such as connector mating between instruments can be carried out very efficiently by creating guidance fixture near male and female connectors. Thus, time, expense, and equipment can be saved, and the environment preserved using the herein-described haptic feedback techniques to perform underwater tasks.
  • FIG. 17 is a block diagram ofsystem1700, in accordance with an example embodiment.System1700 is shown with respect to three environments: proximatephysical environment1702, remotevirtual environment1704, and representing remotephysical environment1706. As one of many possible examples, remotephysical environment1706 can be an underwater environment.
  • In proximatephysical environment1702, master console1710 can be, or include, a computing device configured to enable a human operator (not shown inFIG. 17) to interact with remotephysical environment1706. For example, master console1710 can have a computing device, haptic interface device, and display configured to provide visualizations, such ascomputing device320b,haptic interface device316,display312, andvisualization318 discussed above with respect toFIG. 3A.
  • Master console1710 can communicate withnetwork1720.FIG. 1700 shows that master console1710 can receive feedback fromnetwork1720, such as visual feedback (VF)1712 and haptic force feedback1714 (FH), and can send commands, such as commands with a haptic device, or instructed, end effector position1716 (PH) of a remotely-controlled device in remotephysical environment1706, such asrobot1784.
  • Network1720 can connect computer devices in proximatephysical environment1702 and remotephysical environment1706; e.g., connect master console1710 withcomputing device1720. For example,network1720 can have some or all of the functionality ofnetwork310 ofFIG. 3A.FIG. 17 illustrates thatnetwork1720 can communicate data, including but not limited to,visual feedback1712,haptic force feedback1714, andhaptic device position1716, between proximatephysical environment1702 and remotephysical environment1706.
  • Remotevirtual environment1704 can include computing device(s)1724 configured to communicate withnetwork1720 and remotephysical environment1706. For example, computing device(s)1724 can have some or all of the functionality ofcomputing device320aofFIG. 3A.FIG. 17 indicates that computing device(s)1724 can includevisualization module1730,virtual fixture module1732,inverse kinematics module1734,registration module1750,forward kinematics module1752, andcontroller module1754. Each or all ofmodules1730,1732,1734,1750,1752,1754 can be or comprise software executable on processor(s) of computing device(s)1724.
  • FIG. 17 shows that remotephysical environment1706 can include sensors, such ascamera1780 anddepth sensor1782, androbot1784. For examples,camera1780 can be or include one or more of depth enabledcameras346a,346b,346c,346dshown inFIG. 3A androbot1784 can be or include remote controlleddevice336 also shown inFIG. 3A. In some embodiments,camera1780 can includedepth sensor1782 and be configured for underwater use; e.g.,camera1800 discussed below in the context ofFIGS. 18A and 18B.
  • In operation,camera1780 can capture light within remotephysical environment1706 and generateimages1760. At the same time,depth sensor1782 can capture a 3D depth map of remotephysical environment1706 in the form ofpoint cloud1782 obtain depth information about and generate point cloud (PC)1762.Robot1784 can obtain data about actual robot joint angle(s) for actuator(s), effector(s), robot arm(s), and/or other component(s) ofrobot1784. The data about actual robot joint angle(s) is shown inFIG. 17 as robot joint angle (A)1764.Images1760,point cloud1762, and robotjoint angle1764 can be communicated tocomputing device1724 in remotevirtual environment1704.
  • More specifically,FIG. 17 shows thatimages1760 are communicated toregistration module1750 andvisualization module1730,point cloud1762 is communicated toregistration module1750, and robotjoint angle1764 is communicated to forwardkinematics module1752 andcontroller module1754.
  • Forpoint cloud1762,point cloud1762 can be registered in the robot frame to aid processing withinsystem1700.FIG. 17 shows thatregistration module1750registers point cloud1762 to generate registered point cloud (RPC)1742. To registerpoint cloud1762,registration module1750 can generate a transformation betweenpoint cloud1762 and a point of view ofrobot1784 based on image(s)1760.Registered point cloud1742 can then be communicated fromregistration module1750 tovisualization module1730 andvirtual fixture module1732.
  • Visualization module1730 can use visual data from image(s)1712 and depth data from registeredpoint cloud1742 to generatevisual feedback1712.Visual feedback1712 can also include visual information related tohaptic force feedback1714; i.e., ifhaptic force feedback1714 indicates a force due to collision with an object, then corresponding visual feedback; e.g., a visual collision alert or other indication of collision, can be added tovisual feedback1712. Similarly, end effector position commands, such asend effector position1716, can be used in visual feedback; e.g., ifend effector position1716 indicates that an end effector moves in a direction D, thenvisual feedback1712 can show corresponding movement of (a visualization of) an end effector ofrobot1784.
  • Forward kinematics module1752 can convert robotjoint angle1764 to end effector coordinates (P)1744.Virtual fixture module1732 can employ registeredpoint cloud1742, end effector coordinates1744, and endeffector position1716 to compute and providehaptic force feedback1714 to the operator via master console1710 andnetwork1720.Virtual fixture module1732 also can compute and communicate desired end effector position (PD)1740 toinverse kinematics module1734.Inverse kinematics module1734 can convert desiredend effector position1740 to desired joint angle(s) (θD)1746, which in turn can be converted into motor torque commands (τ)1766 bycontroller module1754. Upon reception of motor torque commands1766,robot1784 can move motor(s) in accord with the received commands1766. In some embodiments,robot1784 can have multiple controllable components; e.g., have multiple arms and/or end effectors. In still other embodiments,robot1784 can be mobile and move in accord with commands provided from master console1710. In these embodiments, motor torque commands1766 can be generated to move multiple components ofrobot1784 and/or to moverobot1784 itself. In still other embodiments,system1700 can usemultiple cameras1780,depth sensors1784, and/orrobots1784 in remotephysical environment1706, which can be controlled by one or more operators at master console1710 and using information provided by computing device(s). Other configurations ofsystem1700 are possible.
  • Depth Cameras for Underwater Virtual Fixtures
  • FIG. 18A is a block diagram ofcamera1800 configured to provide images for underwater haptic rendering, in accordance with an example embodiment.Camera1800 is a video camera built for underwater use and configured to capture depth data and light and provide the captured delight. The captured depth data and light can be provided as images, such as depth images that can be used to generate haptic feedback and virtual fixtures. In some embodiments,camera1800 can provide frames at VGA resolution (720×900 pixels) at a frame rate of 30 frames/second in RGB plus Depth (RGBD) format.Camera1800 can determine depth displacement based on variations of the geometry of a projected pseudo-random pattern of near-infrared (NIR) light. Generally, visible light can be light with wavelength(s) in a visible-light range between 400 and 700 nanometers (nm), and near-infrared light can be light with wavelength(s) in a near-infrared range between 700 and 2000 nm.
  • FIG. 18A shows thatcamera1800 includes near-infrared light source1810,diffraction grating1812,infrared camera1814,visible light source1820,polarizer1822,visible light camera1824,processor1830, andcommunication interface1840. Near-infrared light source1810 can generate infrared light using one or more light-emitting diodes (LEDs), such as an array of 2 W laser diodes having a wavelength (λ) of 830 nm, and a change in wavelength (Δλ) of 3 nm, such as laser diodes made by JDSU.
  • Camera1800 can rely on a structured light principle where the displacement of an object is determined by variations of the geometry of a projected pattern. For example, infraredlight source1810 anddiffraction grating1812 can be used to project a pseudo-random pattern having varying intensities.Camera1800 can also determine an expected pseudo-random pattern based on the projected pseudo-random pattern, and then determine variations in geometry by comparing the captured and expected pseudo-random patterns.
  • Camera1800 can incorporate a near-infrared monochrome sensor (e.g. Microsoft/X853750001/VCA379C7130, 1200×1600 pixels), perhaps as part ofinfrared camera1814, for measuring depth from the distortion between the projected and observed pseudo-random patterns. For reduce the back scattering, a pulse duration of infraredlight source1810 can be set much shorter than the time expected for emitted infrared light to travel to the target.Infrared camera1814 can capture light in the near-infrared range and generate image(s) using the captured light, including but not limited to images of projected pseudo-random patterns of near-infrared light emitted by infraredlight source1810 anddiffraction grating1812. In some embodiments, infrared camera can generate images having a predetermined depth resolution or range of depth resolutions; e.g., a depth resolution of 1.5-2.5 mm.
  • Visiblelight source1820 can use one or more light sources, such as visible light LEDs, to generate visible light. In some embodiments,visible light source1820 can have blue-green filters and/or blue-green LEDs to generate visible light in the 450 to 480 nm range. For example,visible light source1820 can be, or include, a Sea-View SV-Q10K 500W Halogen Unit. Visible light in 450-480 nm range can match a blue-green transmission window of water, thus improving the contrast and range of operation. Polarizer1822 can reduce glare and undesirable reflections from the environment. Polarizer1822 also can detect changes in a polarization state of light scattered by water and suspended particles, allowing for enhanced contrast.Visible light camera1814 can capture light in the visible light range and generate image(s) using the captured light.
  • Images generated byinfrared camera1814 andvisible camera1824 can be received byprocessor1830 and provided tocommunication interface1840. For example,processor1830 can interleave depth-related images received frominfrared camera1814 with visible light images fromvisible light camera1824 into pairs of images, where each pair of images includes a depth-related image and a visible-light image captured at substantially the same time; e.g., within 1/30th of a second of each other.
  • In some embodiments,processor1830 can alternate activation oflight sources1810,1820 at a rate based on a frame rate ofcamera1800. For example, supposecamera1800 is configured to generate N video frames per second, with N>1, and let a frame speed be X seconds, where X≦1/N. Then,light source1810 can be activated to emit near-infrared light for 0.5× seconds and a depth-related image captured by near-infrared camera1814 using the emitted near-infrared light, thenlight source1820 can be activated to emit visible light for 0.5× seconds and a visible-light image captured byvisible light camera1824 using the emitted visible light. In particular of these embodiments, a single camera configured to capture light in both the visible-light range and near-infrared range can perform the functionality ofinfrared camera1814 andvisible light camera1824 by capturing both images of the pair of images within the frame speed of X seconds.
  • In other embodiments, the pair of images includes a depth-related image and a visible light image captured; e.g.,processor1830 can activate bothlight sources1810 and1820 at the same time and bothinfrared camera1814 andvisible light camera1824 can capture light emitted at the same time in the respective near-infrared and visible-light ranges. Thus, each pair of images can provide depth and visible light information about an environment captured at substantially the same time.
  • FIG. 18B is an exploded view ofcamera1800, in accordance with an example embodiment.FIG. 18B showscamera1800 with infraredlight source1810,infrared camera1814,visible light source1820,visible light camera1824,communication interface1840,watertight body1850, andlight shield1852.Watertight body1850 can hold components ofcamera1800 on and/or inside ofbody1850.Light shield1852 can shieldcameras1814,1824 from direct light emitted fromlight sources1810,1820, and so reduce glare on images taken bycameras1814,1824.
  • For underwater applications,camera1800 can be waterproofed.Camera1800 can be enclosed inwatertight body1850 configured to isolatecamera1800 from the harm of water using materials that have little or no impact on optical performance.Camera1800 can include water-tight connectors that allow external communication via optical fibers connected tocommunication interface1840. In some embodiments not shown inFIG. 18A,camera1800 can include a battery configured for poweringcamera1800. In some embodiments, one or more adaptors can enable connection ofcamera1800 to a borescope. The adapter(s) can combine near-infrared images and visible-light images, and recover the depth information from the near-infrared images using a wavelength-specific beam splitter.
  • FIG. 18C depicts asensor system1860, in accordance with an example embodiment.Sensor system1860 includes metal detector(s)1870 and a scanning camera system, where the scanning camera system has alaser scanner1862 andline scan camera1864. Metal detectors, such as metal detector(s)1870, are discussed below in more detail in the context of at leastFIGS. 24A through 27B.
  • In the example shown inFIG. 18C,sensor system1860 is scanning inregion1880 along ascan direction1882 upward and rightward.Laser scanner1862 generates an illumination field using laser beams such aslaser beam1872FIG. 18C show laser beams generated bylaser scanner1862 using dashed and dotted lines.Laser scanner1862 can use using high power collimated laser lights to generate collimated laser beams with minimal cross sections, such aslaser beam1872.FIG. 18C shows thatlaser scanner1860 is configured to emit laser beams, such aslaser beam1872 along scan lines, such asscan lines1874a,1874b,1874c,1874d, and1874e, withinregion1880.
  • Sensor system1860 allows small individual elements of a target scene, such as a portion ofregion1880 swept out by scan lines1874a-1874e, to be illuminated one at a time with minimal backscatter. For example,FIG. 18C shows a field ofview1876 forline scan camera1864 as a relatively small portion ofregion1880. At the same time, a narrow field of view photodetector (not shown inFIG. 18C) can track field ofview1876 and reduce forward scatter from light reflected from the target scene.
  • FIG. 19 shows near-infrared image1900 ofobject1910 in clear water, in accordance with an example embodiment. To captureimage1900, a near-infrared camera and light source, such as near-infrared camera1814 and near-infrared light source1810, were suspended about 85 cm (about 33.5 inches) above a flat surface. A container with clear water about 20 cm (about 7.9 inches) deep was placed on the flat surface below the near-infrared camera and light source. The near-infrared camera was aimed at the body of water. Then, the near-infrared light source generated a pseudo-random pattern and a near-infrared image, nowimage1900, was captured by the near-infrared camera. The pseudo-random pattern is observable inFIG. 19 as dots scattered throughoutimage1900; e.g., dots visible below and to either side of a white circle used to outlineobject1900.
  • FIG. 20A is near-infrared image2000 of an object in murky water, in accordance with an example embodiment.Image2000 was captured using the same setup used for FIG.19—a near-infrared camera and light source, such as near-infrared camera1814 and near-infrared light source1810, were suspended about 85 cm above a flat surface, about 20 cm of water was placed on a container above the flat surface, and the near-infrared camera aimed at the water. Then, as withimage1900, a pseudo-random pattern of near-infrared light was projected onto the water and a near-infrared image, nowimage2000, was captured.Object2010aofimage2000 is the same object asobject1910 ofimage1900.
  • However, forimage2000, murky water was used rather than the clear water used forimage1900 ofFIG. 19.FIG. 20A indicates that only part of the projected pattern makes it through the body of murky water.
  • FIG. 20B isimage2050 representing depth data related toimage2000 ofFIG. 20A, in accordance with an example embodiment. The depth data was obtained by processing the partial projected pattern obtained fromimage2050. The depth data can be distorted due to the index difference between air and water, but as long as the depth data can be obtained through a body of water, it can be corrected by software post-processing.Image2050 showsobject2010bclearly, indicating that valid depth data forobject2010bwas obtained even though only a partial projected pattern onobject2010awas obtainable from image2010.
  • FIG. 21 showsgraph2100 of light absorption in water versus light wavelength andgraph2150 of light absorption in water versus light wavelength over a range of water temperatures, in accordance with an example embodiment.Graph2100, shown in the upper portion ofFIG. 21, shows that the water absorption spectrum has a minimum near the NIR wavelength, indicating that the NIR wavelength is suitable for underwater.Graph2150, shown in the lower portion ofFIG. 21, shows water absorption for light in a range of wavelengths between 700 and 900 nm as a function of temperatures ranging from 30° C. to 40° C. witharrows2160 and2162 indicating a direction of temperature.
  • Resolving Dot Pattern Interference in Multiple Structured Light Cameras
  • The simultaneous use of multiple depth-enabled cameras; e.g.,camera1800 can extend coverage regions for haptic rendering, overcome occlusions, and enable complete full-circle captures of environments. In some cases, structured light sensing can degrade when multiple depth-enabled cameras are pointing at the same scene, due to continuous, non-modulated projection of pseudo-random patterns; e.g., as discussed above in the context ofdiffraction grating1812 ofFIG. 18A. The continuous, non-modulated projection of pseudo-random patterns can cause crosstalk when dot patterns of devices interfere with one another.
  • Herein are provided techniques and devices that can mitigate structured light crosstalk. In some embodiments, these techniques and devices can be utilized without modification of the internal electronics, firmware, or software of a depth-enabled camera and without degradation of the depth-enabled camera's frame rate.
  • One technique to eliminate crosstalk involves time multiplexing where a near-infrared light source associated with each camera in a system of cameras can be modulated and synchronized at a predetermined modulated time frequency.
  • FIG. 22 is a block diagram of time-multiplexedsystem2200 of near-infrared cameras2224,2234 configured to capture images ofenvironment2212, in accordance with an example embodiment.Timed switch2210 can generate two periodic signals,T1 signal2214 andT2 signal2216, with both signals having the same period but only one signal is active at any specified time. For example,timed switch2210 can use generate a square wave with equal active/inactive periods forT1 signal2214 and an inverted version of the square wave for T2 signal2216 (or vice versa). Other techniques for generatingT1 signal2214 andT2 signal2216 can be used as well.
  • Then, whenT1 signal2214 is active, near-infrared light source2220 can be activated. When activated, near-infrared light source2220 can shine near-infrared light2226 throughdiffraction grating2222 and ontoenvironment2212. Subsequently, near-infrared camera2224 can capture near-infrared light2228 reflected fromenvironment2212 as an image ofT1 images2240. Near-infrared camera2224 can placeT1 images2240 ontoconnector2260.
  • When T2 signal2216 is active, near-infrared light source2230 can be activated. When activated, near-infrared light source2230 can shine near-infrared light2236 throughdiffraction grating2232 and ontoenvironment2212. Subsequently, near-infrared camera2234 can capture near-infrared light2238 reflected fromenvironment2212 as an image ofT2 images2250. Near-infrared camera2224 can placeT2 images2250 ontoconnector2260 for use by to device(s) outside ofsystem2200. AsT1 images2240 are placed ontoconnector2260 at times based onT1 signal2214 andT2 images2250 are placed ontoconnector2260 at times based onT2 signal2216, T1 images and T2 images can be interleaved in time as interleaved T1/T2 images2262. As shown inFIG. 22, interleaved T1/T2 images2262 can be sent fromsystem2200, perhaps for use by device(s) outsidesystem2200.
  • In other embodiments not shown inFIG. 22, more than two near-infrared cameras can be used. In these embodiments, each of N near-infrared cameras, N>2, can be assigned an equal, periodic, time interval; e.g., a periodic time slot. During an assigned time slot, a corresponding light source to the near-infrared camera can be illuminated, image(s) ofenvironment2212 can be captured by the near-infrared camera and the image(s) transmitted. Timeslots can be allocated in round-robin fashion to each of the N near-infrared cameras as long as images ofenvironment2212 are to be captured and transmitted.
  • In some embodiments, near-infrared cameras2224,2234 are active continuously. When a signal for corresponding camera is not active; e.g.,T1 signal2214 forNIR camera2224, andT2 signal2216 for near-infrared camera2234, images generated by that camera can be considered to be invalid and so not be processed. To reduce any image processing applicable to interleaved T1/T2 images2262, a shutter synchronized with T1 and T2 signals can be used to block or allow light fromlight sources2220,2230 so a corresponding camera captures light from a light source associated with the corresponding camera; e.g., light fromlight source2220 is blocked by the shutter whenT1 signal2214 is not active and is allowed by the shutter whenT1 signal2214 is active.
  • Another technique to eliminate crosstalk involves is wavelength-division multiplexing where each near-infrared light source associated with a camera in a system of cameras has a different wavelength.
  • FIG. 23 is a block diagram of frequency-multiplexedsystem2300 of near-infrared cameras2324,2334 configured to capture images ofenvironment2310, in accordance with an example embodiment. λ1 light source2320 can be configured to emit light in a sub-range of the near-infrared range of frequencies, where the sub-range includes a wavelength λ1. Similarly, λ2light source2330 can be configured to emit light in a sub-range of the near-infrared range of frequencies, where the sub-range includes a wavelength λ2, where the sub-range including wavelength λ1 does not overlap the sub-range including wavelength λ2. For example, if λ1=820 nm and λ1=900 nm, the sub-range including λ1 can be 800-840 nm, and the sub-range including λ2 can be 880-920 nm. In some embodiments, the sub-range can include light of one frequency; e.g., using the previous example, the sub-range including λ1 can have one frequency of 820 nm and the sub-range including λ2 can have one frequency of 900 nm. Many other example values of λ1, λ2, and sub-ranges associated with λ1 and λ2 are possible as well.
  • λ1 light source2320 can emit λ1 light2326 that reachesenvironment2310.Environment2310 can reflect some or all of the light including frequency λ1 as λ1 light2328 to near-infrared camera2324. Subsequently, near-infrared camera2324 can capture nearλ1 light2328 as an image ofλ1 images2340. Near-infrared camera2324 can placeλ1 images2340 ontoconnector2360.
  • λ1 light source2323 can emit λ2 light2336 that reachesenvironment2310.Environment2310 can reflect some or all of the light including frequency λ2 as λ2 light2338 to near-infrared camera2334. Subsequently, near-infrared camera2334 can capture nearλ1 light2328 as an image ofλ1 images2350. Near-infrared camera2324 can placeλ2 images2350 ontoconnector2360.
  • As shown inFIG. 23, λ1 images and λ2 images can be sent fromsystem2300 as λ1 andλ2 images2362, perhaps for use by device(s) outsidesystem2200. A device receiving λ1 andλ2 images2362 can use filter(s) to ensure that images taken in light having a corresponding wavelength is detected and analyzed.
  • In other embodiments not shown inFIG. 23, more than two near-infrared cameras can be used. In these embodiments, each of N near-infrared cameras, N>2, can be assigned a non-overlapping frequency, and perhaps a non-overlapping sub-range of frequencies that includes the assigned frequency, within the near-infrared range of frequencies. Each camera can then have an associated light source configured to emit light with the assigned non-overlapping frequency (or emit light in the assigned sub-range of frequencies, if sub-ranges are assigned). The camera can capture light reflected fromenvironment2310 having the assigned non-overlapping frequency as an image, and send any captured images.
  • Metal Detecting Systems
  • A metal detector can work alongside a camera system, such ascamera1800,system2200, orsystem2300. For example,sensor system1860 discussed above in the context ofFIG. 18C includes both a camera and metal detector(s). An array of the metal detectors can be arranged around the camera system to allow simultaneous capture of 2D metal profiling data and images. The array of metal detectors can be arranged in a pattern; e.g., the array can be arranged in a cross pattern. In other scenarios, a metal detector or an array of metal detectors can be mounted on a vessel or near actuators, such as robotic arms, to map 2D or 3D metal profiles without using a visual or acoustical detector.
  • The metal profile, such as but not limited to a line scan, raster scan, or graph of magnetic flux over time, can be generated by an array of metal detectors moving across a target region while scanning. In some embodiments, the metal profile detector can utilize a phase array design where radiation pattern of the B field can be shifted by adjusting the phase difference between driving coils inside each metal sensor.
  • To reliably detect smaller metal items underwater; e.g., small munitions, a compact metal detector can be used. In some embodiments, the metal detector can have a relatively-small sensing resolution; e.g., a resolution of one square centimeter, and be capable of differentiating metal of different shapes and geometry using a fiber-optic sensor.
  • A metal detector can be based on a metal sensor having magnetostrictive materials. A magnetostrictive material is a material that changes shape or dimensions in the presence of a magnetic field. The metal detector operates by comparing a light signal sent through a loop of optical fiber coated with a magnetostrictive material with a light signal sent through an uncoated reference optical fiber. When the coated loop passes over para- or diamagnetic material, the magnetic field can induce strain in the fiber through magnetostriction of the coating. The induced strain can change the index of refraction within the coated fiber through the photoelastic effect, which changes the coated fiber's response relative to the reference. For example, the strain can causes a change in the length of the optical path. The metal detector can include an interferometer to measure the phase changes induced in the strained fiber relative to light passing through the reference coil. The presence of such phase changes indicates that metal has been detected.
  • Metal detectors using magnetostrictive materials can be configured to be relatively small in size. Further, magnetostrictive-based metal detectors can identify shapes or topographic features of metal targets while operating as a single metal detector or a metal detector in an array of metal detectors.
  • FIG. 24A showsstructure2400 of a metal sensor, in accordance with an example embodiment.Structure2400 includes a layer ofmagnetostrictive material2410a, alight carrying material2412, and another layer of magnetostrictive material2410b. For example, the light carrying material can be an optical fiber and the magnetostrictive material can be a ferromagnetic polymer.Structure2400 can be used when layers ofmagnetostrictive material2410aand2410bcoat light carrying material.
  • FIG. 24B is a diagram of interferometer-basedmetal detector2420, in accordance with an example embodiment.Laser2422 can send a light signal tooptical coupler2430, which is connected to both sensing coil2440 andreference coil2442. Inmetal detector2420, each ofcoils2440 and2442 is configured to act as a sensing arm in an interferometer. Both sensing coil2440 andreference coil2442 include a light pathway; e.g., one or more optical fibers in the coil. However, sensing coil2440 includes fiber coated with magnetostrictive material, whilereference coil2442 includes fiber that is not coated with magnetostrictive material; that is, coil2440 includesstructure2400 whilecoil2442 does not includestructure2400.
  • After light passes through each of sensing coil2440 andreference coil2442, the light is provided tooptical coupler2432 which can pass the light from sensing coil2440 todetector2452 and passes the light from reference coil2440 todetector2450.Light reaching detector2450 can be compared to light reachingdetector2452; e.g., to determine phase changes in light indicating differences in sensing coil2440 andreference coil2442. That is,metal detector2420 includes a Mach-Zehnder interferometer for detecting changes in light betweencoils2440 and2442 that represent the presence or absence of metal.
  • FIG. 24C is a diagram of interferometer-basedmetal detector2460, in accordance with an example embodiment.Laser2462 can send a light signal tooptical coupler2470, which is connected todetector2472,sensing coil2480, andreference coil2480. Bothsensing coil2480 andreference coil2482 include a light pathway; e.g., one or more optical fibers in the coil. Inmetal detector2460, each ofcoils2480 and2482 is configured to act as a sensing arm in an interferometer. However,sensing coil2480 includes fiber coated with magnetostrictive material, whilereference coil2482 includes fiber that is not coated with magnetostrictive material; that is,coil2480 includesstructure2400 whilecoil2482 does not includestructure2400.
  • Light passing throughsensing coil2480 can be provided tomirror2490, which can reflect the provided light back throughsensing coil2480 andcoupler2470 todetector2472. Similarly, light passing throughreference coil2482 can be provided tomirror2492, which can reflect the provided light back throughreference coil2482 andcoupler2470 todetector2472.Light reaching detector2472 from sensingcoil2480 can be compared to light reachingdetector2472 fromreference coil2482 e.g., to determine phase changes in light indicating differences insensing coil2480 andreference coil2482. That is,metal detector2460 includes a Michelson interferometer for detecting changes in light betweencoils2440 and2442 that represent the presence or absence of metal.
  • FIG. 25A is a diagram ofmetal detection system2500 that includesmultiple metal detectors2510a,2510b,2510c,2510d,2510e,2510f,2510g,2510h, and2510i, in accordance with an example embodiment. More specifically,metal detection system2500 includes ninemetal detectors2510a,2510b,2510c,2510d,2510e,2510f,2510g,2510h, and2510iarranged in a cross-shaped pattern. Each ofmetal detectors2510a,2510b,2510c,2510d,2510e,2510f,2510g,2510h, and2510ican be an interferometer-based metal detector, such asmetal detector2420 discussed above in more detail in the context ofFIG. 24B ormetal detector2460 discussed above in more detail in the context ofFIG. 24C.
  • In particular, each ofmetal detectors2510a,2510b,2510c,2510d,2510e,2510f,2510g,2510h, and2510ican utilize magnetostrictive material as part of the interferometer-based metal detector. When one or more metal detectors2510a-2510iare on an area having paramagnetic and/or diamagnetic material or move over an area having paramagnetic and/or diamagnetic material, the one or more metal detectors2510a-2510ican detect a perturbation due to presence of a paramagnetic and/or diamagnetic material in the field. Based on a sequence of magnitude of the perturbation collected by the array of sensors, we can reconstruct the metal profile of the area the sensors pass over. The metal profile can include graphs, diagrams, line scans, and/or raster scans discussed below.
  • Other patterns of metal detectors than cross-shaped patterns can be used in metal detection systems.FIG. 25B is a diagram ofmetal detection system2520 that includes multiple metal detectors, in accordance with an example embodiment. More specifically,metal detection system2520 includes ninemetal detectors2530a,2530b,2530c,2530d,2530e,2530f,2530g,2530h, and2530iarranged in an X-shaped pattern. Each ofmetal detectors2530a,2530b,2530c,2530d,2530e,2530f,2530g,2530h, and2530ican be an interferometer-based metal detector, such as metal detectors2510a-2510idiscussed above in the context ofFIG. 25A.
  • FIG. 25C is a diagram ofmetal detection system2540 that includes multiple metal detectors, in accordance with an example embodiment. More specifically,metal detection system2520 includes twenty-five metal detectors arranged in a grid-based pattern. Each of the twenty-five metal detectors, includingmetal detector2550, are shown inFIG. 25C as marked with a “MD”. Each metal detector inmetal detection system2540 can be an interferometer-based metal detector, such as metal detectors2510a-2510idiscussed above in the context ofFIG. 25A. Other patterns of metal detectors and other metal detectors can be used in metal detection systems as well.
  • FIG. 26A depicts a sensor system usingmetal detection system2500, in accordance with an example embodiment.Sensor system2600 can generate a metal profile, such as a line scan, by moving an array of metal detectors2510a-2510iinmetal detection system2500 alongscan lines2620 while the metal detectors are performing metal scans; e.g., searching for metallic objects belowsensor system2600.FIG. 26A shows metal detector(s) ofmetal detection system2500 performing metal scan2610 whilesensor system2600 is in motion along an upper-most scan line ofscan lines2620.
  • FIG. 26B depictssensor system2630 usingmetal detection system2500, in accordance with an example embodiment.Sensor system2630 can utilize a phase array design where radiation pattern of theB field2640 can be shifted by adjusting a phase difference between driving coils inside each metal detector2510a-2510iofmetal detector system2500 while metal detector2510a-2510iare performing metal scans. A raster scan can be generated by adding phase differences in metal scans between the cross-pattern of metal detectors inmetal detector system2500. To generate the raster scan,sensor system2640 can travel alongraster scan path2650, which is a zigzag shaped path as shown inFIG. 26B. The generated raster scan can be utilized as a metal profile for objects alongraster scan path2650.
  • In some embodiments,sensor system2600 and/orsensor system2630 can utilize more and/or different metal detection systems thanmetal detection system2500; e.g., one or moremetal detection systems2520 and/ormetal detection systems2540 can be utilized as well as, or rather thanmetal detection system2500 as shown inFIGS. 26A and 26B. Other techniques to generate metal profiles by using sensor systems equipped with metal detection systems are possible as well.
  • FIG. 27A isimage2710 ofexample metal objects2702,2704,2706, in accordance with an example embodiment.Image2710 shows an arrangement of metal objects, shown in left-to-right order, as rod2702 running along a left-most side of image2700,C clamp2704 centrally placed in image2700 androd2706 running along the left-most side of image2700.
  • FIG. 27B is agraph2720 of an output signal from an interferometer-based metal detector while scanning for metal objects in accordance with an example embodiment.Graph2720 shows magnetic flux density in Gauss versus time as observed by an interferometer-based metal detector that is part of an array of metal detectors in a metal detection system; e.g., metal detector2510aofmetal detection system2500. When the magnetic flux density reaches a peak, the metal detector has located a metal object.
  • Graph2710 also includesinset2730 with a copy ofimage2710 including rod2702,C Clamp2704, androd2706, in accordance with an example embodiment. In the scan performed to generategraph2720, the metal objects inimage2710 were scanned starting 0.2 after the beginning of the scan and the metal detector traveled in the direction of scan direction (SD)2732 with respect toinset2730.
  • Graph2720 shows data for an interferometer based metal detector as an upper/darker grey line ofgraph2720 and data for a Hall effect sensor as lower/lighter grey line ofgraph2720. The in interferometer based metal detector and the Hall effect sensor were at the same position inmetal detection system2710.Graph2720 shows that both sensors had similar responses; e.g., both the upper and lower lines ofgraph2720 have similar peaks and valleys during the scan.
  • Specifically, at a time about 0.2 seconds as indicated byarrow2740, the metal detector passed over rod2702. Then, at a time at about 0.25 seconds as indicated byarrow2742, the metal detector passed over a leftmost prong ofC Clamp2704 as shown ininset2730 i.e., the top of the “C”, and at a time at about 0.38 seconds as indicated byarrow2744, the metal detector passed over a rightmost prong ofC Clamp2704; i.e., the bottom of the “C”. Later, at a time of about 0.45 seconds as indicated byarrow2746, the metal detector passed overrod2706.
  • Graph2720 shows that the metal detector indicated a peak in magnetic flux, and thus detected metal, when passing over rod2702. The magnetic flux fell off from the peak of about 1200 Gauss at 0.2 seconds within approximately 0.01 seconds indicating that a narrow band of metal lie in the path of the metal detector while traveling alongscan direction2732. A similar peak is noted at about 0.45 seconds when the metal detector passed overrod2706. Peaks were reached at about 0.25 seconds when the leftmost prong ofC Clamp2704 was detected and at about 0.38 seconds when the rightmost prong ofC Clamp2704, but in between these two peaks, the metal detector dropped to a relatively-low Gauss level of 1000 Gauss (or less) at about 0.30 to 0.32 seconds, indicating that the metal detector did not detect metal during that interval; i.e., the metal detector did not detect the body ofC Clamp2704 according tograph2720.
  • FIG. 27C is a diagram2750 of metal objects determined based on the data inFIG. 27B, in accordance with an example embodiment. Diagram2750 ofFIG. 27C is a 2D reconstructed map of rod2702,C Clamp2704, androd2706 shown inimage2710 ofFIG. 27A using data from all metal detectors in a metal detection system while passing over the metal objects depicted inimage2710. Diagram2750 shows each ofrod2752,C clamp2754, androd2706 in the same relative positions as rod2702,C Clamp2704, androd2706 are shown inimage2710. Further, the general shape of rod2702 andC Clamp2704 ofimage2710 are respectively reflected asrod2752 andC Clamp2754 of diagram2750. Additionally,rod2756 in diagram2750 accurately depicts both the relative position and shape ofrod2706 shown inimage2710.
  • FIG. 28 illustrates ascenario2800 wherevessel2810 detects and retrievesunderwater objects2820,2822, in accordance with an example embodiment. Inscenario2800,vessel2810 is floating inriver2830 aboveriver bed2832. In other scenarios,vessel2810 can be used in an ocean, lake, pond, or other water environment thanriver2830.Vessel2810 has robot arms and hands acting asactuators2840 and2850. Eachactuator2840,2850 can be partially or completely controlled by a remote operator; e.g., a human operator aboardvessel2810 or located in another location. In some scenarios, the remote operator can adjust a level of automation ofactuator2840 and/or2850; e.g., to be manually operated, semi-autonomously operated, or be fully autonomous operated.
  • The remote operator can receive feedback about the operation of actuator2840 (and/or2850) and/or about the environment using sensor system2842 (and/or2852). Sensor system2842 (and/or2852) can include one or more depth-enabled cameras including visible-light camera(s) and near-infrared camera(s), metal detectors/metal detection systems, and light sources including visible-light light source(s) and near-infrared light source(s) perhaps configured to emit pseudo-random patterns of light. In particular, sensor system2842 (and/or2852) can be configured to provide depth images, where the depth images can be used to generate haptic feedback for three or more degrees of freedom in controlling actuator2840 (and/or2850).
  • Inscenario2800, the remote operator of actuator2840 (and/or2850) operates actuator2840 (and/or2850) in accord with a task list, where the task list is configured to control virtual fixtures associated with tasks of retrievingobject2820 fromriver bed2832 andobject2822 from underriver bed2832. Once an object is retrieved, the object can be placed inobject container2860; e.g., using a guidance fixture associated with the task list configured to guide actuator2840 (and/or2850) from a location of the object to a location ofobject container2860. Other virtual fixtures, such as but not limited to forbidden-region fixtures, can be associated with the task list.
  • Inscenario2800, another remote operator can use remotely-operatedvehicle2862 to provide additional information regarding theenvironment including objects2820,2822,actuators2840,2850,river bed2832, and/orriver2830. Remotely-operatedvehicle2862 can be attached to vessel281 viatether2864.Tether2864 can include power and/or communication cables to enable remotely-operatedvehicle2862 to be powered from and/or communicate withvessel2810. In other embodiments, remotely-operatedvehicle2862 can operate withouttether2864.
  • FIG. 29 is a flowchart ofmethod2900, in accordance with an embodiment.Method2900 can be carried out with a depth-enabled camera, such ascamera1800.Method2900 can start atblock2910, where a camera, such ascamera1800 discussed above in the context of at leastFIGS. 18A and 18B, can capture, within a predetermined interval of time, first light within a first frequency range of light in the underwater environment and second light within a second frequency range of light in the underwater environment. In some embodiments, the first frequency range of light can include a range of frequencies of light between 450 nanometers (nm) and 480 nm. Then, the second frequency range of light of frequencies can include a range of frequencies of light between 825 and 835 nm.
  • Atblock2920, the camera can generate a first image based on the first light and a second image based on the second light. The first frequency range of light can differ from the second frequency range of light.
  • Atblock2930, the camera can send at least the first image and the second image. In some embodiments, sending at least the first image and the second image from the camera can include: generating a predetermined number of frames per second, where each frame includes a first frame image based on light within the first frequency range received at a specified time and a second frame image based on light within the second frequency range captured at the specified time, and sending the predetermined number of frames per second from the camera. In particular of these embodiments, the predetermined number of frames per second can be a number N, with N>1. Then, the predetermined interval of time can be less than or equal to 1/N seconds.
  • In some embodiments, the camera can include a first light emitting diode (LED) and a second LED. Then,method2900 can further include: generating light within the first frequency range of light using the first LED and generating light within the second frequency range of light using the second LED. In particular of these embodiments, the camera further includes diffraction grating. Then, generating light within the second frequency range of light can include the scattering the light generated by the second LED into a pseudo-random pattern of light within the second frequency range using the diffraction grating. In more particular of these embodiments,method2900 can further include capturing the pseudo-random pattern of light within the second frequency range using the camera, and determining distortion within the second frequency range of light by comparing the captured pseudo-random pattern of light to an expected pseudo-random pattern of light.
  • In other embodiments, the camera can include a polarizer. Then,method2900 can further include reducing reflections in at least one image of the first image and the second image using the polarizer. In still other embodiments, the camera can include an optical adaptor. Then,method2900 can further include separating light received at the device into the first light and the second light using the optical adaptor. In yet other embodiments, the first image and second image are configured to be utilized for generating haptic feedback.
  • The above detailed description describes various features and functions of the disclosed systems, devices, and methods with reference to the accompanying figures. In the figures, similar symbols typically identify similar components, unless context dictates otherwise. The illustrative embodiments described in the detailed description, figures, and claims are not meant to be limiting. Other embodiments can be utilized, and other changes can be made, without departing from the spirit or scope of the subject matter presented herein. It will be readily understood that the aspects of the present disclosure, as generally described herein, and illustrated in the figures, can be arranged, substituted, combined, separated, and designed in a wide variety of different configurations, all of which are explicitly contemplated herein.
  • With respect to any or all of the ladder diagrams, scenarios, and flow charts in the figures and as discussed herein, each block and/or communication may represent a processing of information and/or a transmission of information in accordance with example embodiments. Alternative embodiments are included within the scope of these example embodiments. In these alternative embodiments, for example, functions described as blocks, transmissions, communications, requests, responses, and/or messages may be executed out of order from that shown or discussed, including substantially concurrent or in reverse order, depending on the functionality involved. Further, more or fewer blocks and/or functions may be used with any of the ladder diagrams, scenarios, and flow charts discussed herein, and these ladder diagrams, scenarios, and flow charts may be combined with one another, in part or in whole.
  • A block that represents a processing of information may correspond to circuitry that can be configured to perform the specific logical functions of a herein-described method or technique. Alternatively or additionally, a block that represents a processing of information may correspond to a module, a segment, or a portion of program code (including related data). The program code may include one or more instructions executable by a processor for implementing specific logical functions or actions in the method or technique. The program code and/or related data may be stored on any type of computer readable medium such as a storage device including a disk or hard drive or other storage medium.
  • The computer readable medium may also include physical and/or non-transitory computer readable media such as computer-readable media that stores data for short periods of time like register memory, processor cache, and random access memory (RAM). The computer readable media may also include physical and/or non-transitory computer readable media that stores program code and/or data for longer periods of time, such as secondary or persistent long term storage, like read only memory (ROM), optical or magnetic disks, compact-disc read only memory (CD-ROM), for example. The computer readable media may also be any other volatile or non-volatile storage systems. A computer readable medium may be considered a computer readable storage medium, for example, or a tangible storage device.
  • The terms physical and/or tangible computer-readable medium and physical and/or tangible computer-readable media refer to any physical and/or tangible medium that can be configured to store instructions for execution by a processor, processing unit and/or computing device. Such a medium or media can take many forms, including but not limited to, non-volatile media and volatile media. Non-volatile media includes, for example, read only memory (ROM), flash memory, magnetic-disk memory, optical-disk memory, removable-disk memory, magnetic-tape memory, hard drive devices, compact disc ROMs (CD-ROMs), direct video disc ROMs (DVD-ROMs), computer diskettes, and/or paper cards. Volatile media include dynamic memory, such as main memory, cache memory, and/or random access memory (RAM). Many other types of tangible computer-readable media are possible as well. As such, the herein-described data storage can comprise and/or be one or more physical and/or tangible computer-readable media.
  • Moreover, a block that represents one or more information transmissions may correspond to information transmissions between software and/or hardware modules in the same physical device. However, other information transmissions may be between software modules and/or hardware modules in different physical devices.
  • While various aspects and embodiments have been disclosed herein, other aspects and embodiments will be apparent to those skilled in the art. The various aspects and embodiments disclosed herein are for purposes of illustration and are not intended to be limiting, with the true scope and spirit being indicated by the following claims.

Claims (20)

1. A device configured for operation in an underwater environment, comprising:
a camera configured to capture, within a predetermined interval of time, first light within a first frequency range of light in the underwater environment and second light within a second frequency range of light in the underwater environment, wherein the camera is configured to generate a first image based on the first light and a second image based on the second light, wherein the first frequency range of light differs from the second frequency range of light, and wherein the camera comprises a communication interface configured at least to send at least the first image and the second image and to receive one or more commands based on haptic feedback with the haptic feedback generated based on the first image and the second image.
2. The device ofclaim 1, wherein the camera comprises a first light emitting diode (LED) configured to generate light within the first frequency range of light and a second LED configured to generate light within the second frequency range of light.
3. The device ofclaim 2, wherein the camera further comprises a diffraction grating configured to scatter light from the second LED into a pseudo-random pattern of light within the second frequency range.
4. The device ofclaim 3, wherein the second light comprises the pseudo-random pattern of light within the second frequency range, and wherein the device is configured to determine distortion within the second frequency range of light by comparing the pseudo-random pattern of light captured within the second light to an expected pseudo-random pattern of light.
5. The device ofclaim 1, wherein the camera comprises a polarizer configured to reduce reflections at least one image of the first image and the second image.
6. The device ofclaim 1, further comprising an optical adaptor configured to separate light received at the device into the first light and the second light.
7. The device ofclaim 1, wherein the camera is configured to generate a predetermined number of frames per second, wherein each frame comprises a first frame image based on received light within the first frequency range at a specified time and a second frame image based on received light within the second frequency range at the specified time, and wherein the communication interface is configured to transmit the predetermined number of frames per second.
8. The device ofclaim 7, wherein the predetermined number of frames is a number N greater than one, and wherein the predetermined interval of time is less than or equal to 1/N seconds.
9. The device ofclaim 1, wherein the camera is configured to be waterproof.
10. The device ofclaim 1, wherein the haptic feedback is based on visual information of the first image and on depth information determined from the second image.
11. The device ofclaim 1, further comprising an actuator configured to be controlled by the one or more commands.
12. A method, comprising:
capturing, within a predetermined interval of time, first light within a first frequency range of light in an underwater environment and second light within a second frequency range of light in the underwater environment using a camera;
generating a first image based on the first light and a second image based on the second light using the camera, wherein the first frequency range of light differs from the second frequency range of light; and
sending at least the first image and the second image from the camera.
13. The method ofclaim 12, wherein the camera comprises a first light emitting diode (LED) and a second LED, and wherein the method further comprises:
generating light within the first frequency range of light using the first LED; and
generating light within the second frequency range of light using the second LED.
14. The method ofclaim 13, wherein the camera further comprises a diffraction grating, and wherein generating light within the second frequency range of light comprises scattering the light generated by the second LED into a pseudo-random pattern of light within the second frequency range using the diffraction grating.
15. The method ofclaim 14, further comprising:
capturing the pseudo-random pattern of light within the second frequency range using the camera; and
determining distortion within the second frequency range of light by comparing the captured pseudo-random pattern of light to an expected pseudo-random pattern of light.
16. The method ofclaim 12, wherein the camera comprises a polarizer, and wherein the method further comprises:
reducing reflections in at least one image of the first image and the second image using the polarizer.
17. The method ofclaim 12, wherein the camera comprises an optical adaptor, and wherein the method further comprises:
separating light received at the device into the first light and the second light using the optical adaptor.
18. The method ofclaim 12, wherein sending at least the first image and the second image from the camera comprises:
generating a predetermined number of frames per second, wherein each frame comprises a first frame image based on light within the first frequency range captured at a specified time and a second frame image based on light within the second frequency range captured at the specified time; and
sending the predetermined number of frames per second from the camera.
19. The method ofclaim 18, wherein the predetermined number of frames is a number N greater than one, and wherein the predetermined interval of time is less than or equal to 1/N seconds.
20. The method ofclaim 12, wherein the first image and second image are configured to be utilized for generating haptic feedback.
US14/164,1152013-01-242014-01-24Haptically-Enabled Co-Robotics for Underwater TasksAbandonedUS20140320629A1 (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
US14/164,115US20140320629A1 (en)2013-01-242014-01-24Haptically-Enabled Co-Robotics for Underwater Tasks

Applications Claiming Priority (4)

Application NumberPriority DateFiling DateTitle
US201361756132P2013-01-242013-01-24
US201361764908P2013-02-142013-02-14
US201361764921P2013-02-142013-02-14
US14/164,115US20140320629A1 (en)2013-01-242014-01-24Haptically-Enabled Co-Robotics for Underwater Tasks

Publications (1)

Publication NumberPublication Date
US20140320629A1true US20140320629A1 (en)2014-10-30

Family

ID=51788815

Family Applications (4)

Application NumberTitlePriority DateFiling Date
US14/164,115AbandonedUS20140320629A1 (en)2013-01-242014-01-24Haptically-Enabled Co-Robotics for Underwater Tasks
US14/164,111AbandonedUS20140320392A1 (en)2013-01-242014-01-24Virtual Fixtures for Improved Performance in Human/Autonomous Manipulation Tasks
US14/164,114Expired - Fee RelatedUS9477307B2 (en)2013-01-242014-01-24Methods and systems for six degree-of-freedom haptic interaction with streaming point data
US15/291,060Expired - Fee RelatedUS9753542B2 (en)2013-01-242016-10-11Methods and systems for six-degree-of-freedom haptic interaction with streaming point data

Family Applications After (3)

Application NumberTitlePriority DateFiling Date
US14/164,111AbandonedUS20140320392A1 (en)2013-01-242014-01-24Virtual Fixtures for Improved Performance in Human/Autonomous Manipulation Tasks
US14/164,114Expired - Fee RelatedUS9477307B2 (en)2013-01-242014-01-24Methods and systems for six degree-of-freedom haptic interaction with streaming point data
US15/291,060Expired - Fee RelatedUS9753542B2 (en)2013-01-242016-10-11Methods and systems for six-degree-of-freedom haptic interaction with streaming point data

Country Status (1)

CountryLink
US (4)US20140320629A1 (en)

Cited By (28)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20140300702A1 (en)*2013-03-152014-10-09Tagir SaydkhuzhinSystems and Methods for 3D Photorealistic Automated Modeling
US20150145960A1 (en)*2011-11-042015-05-28Empire Technology Development LlcIr signal capture for images
US20150306767A1 (en)*2014-04-242015-10-29Toyota Jidosha Kabushiki KaishaMotion limiting device and motion limiting method
US20150314440A1 (en)*2014-04-302015-11-05Coleman P. ParkerRobotic Control System Using Virtual Reality Input
US20160189391A1 (en)*2014-02-262016-06-30Apeiros, LlcMobile, wearable, automated target tracking system
US9471142B2 (en)2011-06-152016-10-18The University Of WashingtonMethods and systems for haptic rendering and creating virtual fixtures from point clouds
US9477307B2 (en)2013-01-242016-10-25The University Of WashingtonMethods and systems for six degree-of-freedom haptic interaction with streaming point data
US20170048494A1 (en)*2014-04-242017-02-16Cathx Research LtdUnderwater surveys
US9592608B1 (en)*2014-12-152017-03-14X Development LlcMethods and systems for providing feedback during teach mode
US9919422B1 (en)2016-01-062018-03-20X Development LlcMethods and systems to provide mechanical feedback during movement of a robotic system
US20180231973A1 (en)*2017-02-162018-08-16Wal-Mart Stores, Inc.System and Methods for a Virtual Reality Showroom with Autonomous Storage and Retrieval
US20180234624A1 (en)*2017-02-152018-08-16Samsung Electronics Co., Ltd.Electronic device and method for determining underwater shooting
WO2019014245A1 (en)*2017-07-102019-01-173D at Depth, Inc.Underwater optical positioning systems and methods
US10226869B2 (en)2014-03-032019-03-12University Of WashingtonHaptic virtual fixture tools
US10698112B2 (en)2017-05-042020-06-303D at Depth, Inc.Systems and methods for monitoring underwater structures
CN112131921A (en)*2019-06-252020-12-25海盛科技有限公司 Stereo vision-based biological automatic measurement system and its measurement method
US20210171283A1 (en)*2017-11-172021-06-10Ocado Innovation LimitedControl device and method for a robot system
US11278369B2 (en)*2016-04-282022-03-22Sony CorporationControl device, control method, and surgical system
US20220319031A1 (en)*2021-03-312022-10-06Auris Health, Inc.Vision-based 6dof camera pose estimation in bronchoscopy
US20230003516A1 (en)*2019-11-152023-01-05Lufthansa Technik AgBorescope with pattern projection
US20230210613A1 (en)*2020-06-122023-07-06Covidien LpSurgical robotic system with motion integration
CN117289796A (en)*2023-09-222023-12-26中山大学High-interaction mixed reality system and method for complex equipment based on haptic glove
US20240025049A1 (en)*2022-07-252024-01-25Shanghai Flexiv Robotics Technology Co., Ltd.Robot teleoperation method, robot and storage medium
US20240286279A1 (en)*2020-04-242024-08-29Deutsches Zentrum für luft-und Raumfajrt e.V.Method for controlling a robot, and system
US20240302900A1 (en)*2021-06-252024-09-12Keio UniversityOperational data management system, operational data management method, and storage medium
US12165263B1 (en)2022-12-132024-12-10Astrovirtual, Inc.Web browser derived content including real-time visualizations in a three-dimensional gaming environment
US20240430554A1 (en)*2023-06-222024-12-26Integrotech Sp. Z O.O.Device for analysis and recording of an image, especially in an explosion hazard zone
US12347575B2 (en)2020-09-252025-07-013D at Depth, Inc.Systems and methods for laser inspection and measurements

Families Citing this family (50)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
AU2014239979B2 (en)2013-03-152017-06-22Aurora Operations, Inc.Methods, systems, and apparatus for multi-sensory stereo vision for robotics
TWI526980B (en)*2013-10-162016-03-21聯詠科技股份有限公司Non-overlap data transmission method for liquid crystal display and related circuit
US9619029B2 (en)*2013-11-142017-04-11Immersion CorporationHaptic trigger control system
US9164587B2 (en)2013-11-142015-10-20Immersion CorporationHaptic spatialization system
US10489361B2 (en)2014-01-272019-11-26Camelot Uk Bidco LimitedSystem and methods for cleansing automated robotic traffic from sets of usage logs
WO2015116056A1 (en)*2014-01-292015-08-06Hewlett-Packard Development Company, L.P.Force feedback
US10048703B1 (en)*2014-06-062018-08-14The United States Of America As Represented By The Secretary Of The NavyForce feedback pressure cuff systems and methods
US9387588B1 (en)2014-08-252016-07-12Google Inc.Handling gait disturbances with asynchronous timing
US9618937B1 (en)2014-08-252017-04-11Google Inc.Slip detection using robotic limbs
US11586208B2 (en)*2014-10-242023-02-21Clearpath Robotics Inc.Systems and methods for executing a task with an unmanned vehicle
US10185396B2 (en)2014-11-122019-01-22Immersion CorporationHaptic trigger modification system
US9499218B1 (en)2014-12-302016-11-22Google Inc.Mechanically-timed footsteps for a robotic device
DE102015114742B4 (en)*2015-09-032017-06-14Imk Automotive Gmbh Kinematics for an intracorporeally driven instrument of an endoscope
DE102015220495A1 (en)*2015-10-212017-04-27Kuka Roboter Gmbh Protective field adaptation of a manipulator system
GB2543777B (en)*2015-10-272018-07-25Imagination Tech LtdSystems and methods for processing images of objects
JP6625421B2 (en)2015-12-112019-12-25シスメックス株式会社 Medical robot system, data analysis device, and medical robot monitoring method
CN105653769B (en)*2015-12-242018-09-28电子科技大学The multidisciplinary reliability design optimization method of mechanical arm under time-varying condition of uncertainty
US9925667B1 (en)2016-01-252018-03-27Boston Dynamics, Inc.Continuous slip recovery
US10077007B2 (en)*2016-03-142018-09-18Uber Technologies, Inc.Sidepod stereo camera system for an autonomous vehicle
WO2017187223A1 (en)*2016-04-252017-11-02Telefonaktiebolaget Lm Ericsson (Publ)Method and apparatus for automating physical equipment replacement and maintenance
CN105824250B (en)*2016-05-142018-10-19大连理工大学 Bionic arm control system based on cerebellum model and modeling method of cerebellum model
US10410365B2 (en)*2016-06-022019-09-10Verily Life Sciences LlcSystem and method for 3D scene reconstruction with dual complementary pattern illumination
EP3502841B1 (en)*2016-08-182023-07-26Sony Group CorporationInformation processing device, information processing system and information processing method
US10712561B2 (en)*2016-11-042020-07-14Microsoft Technology Licensing, LlcInterference mitigation via adaptive depth imaging
CN106625661B (en)*2016-12-092019-08-09北京邮电大学 A Construction Method of Adaptive Virtual Fixture Based on OSG
US11202682B2 (en)*2016-12-162021-12-21Mako Surgical Corp.Techniques for modifying tool operation in a surgical robotic system based on comparing actual and commanded states of the tool relative to a surgical site
US10249096B2 (en)2017-05-172019-04-02International Business Machines CorporationMixing virtual image data and physical image data
CN107391929B (en)*2017-07-212020-07-28北京粒创科技有限公司Virtual platform system based on user behavior data
CN107507267A (en)*2017-07-282017-12-22电子科技大学Human body back three-dimensional reconstruction method
US11027432B2 (en)2017-09-062021-06-08Stryker CorporationTechniques for controlling position of an end effector of a robotic device relative to a virtual constraint
EP3462411B1 (en)*2017-09-272025-02-19Arkite NVConfiguration tool and method for a quality control system
US20190105562A1 (en)*2017-10-112019-04-11Immersion CorporationHaptic effects with multiple peripheral devices
US10967862B2 (en)2017-11-072021-04-06Uatc, LlcRoad anomaly detection for autonomous vehicle
KR102787616B1 (en)2017-12-182025-03-28돌비 인터네셔널 에이비Method and system for handling local transitions between listening positions in a virtual reality environment
CN118824260A (en)2018-04-112024-10-22杜比国际公司 Method, device and system for 6DOF audio rendering and data representation and bitstream structure for 6DOF audio rendering
US10678264B2 (en)2018-10-102020-06-09Midea Group Co., Ltd.Method and system for providing remote robotic control
US10816994B2 (en)2018-10-102020-10-27Midea Group Co., Ltd.Method and system for providing remote robotic control
EP3915269B1 (en)2019-01-242025-05-14InterDigital VC Holdings, Inc.System and method for adaptive spatial content streaming with multiple levels of detail and degrees of freedom
JP7198950B2 (en)*2019-07-122023-01-04ニューラリンク コーポレーション Optical coherence tomography for robotic neurosurgery
CN110989605B (en)*2019-12-132020-09-18哈尔滨工业大学Three-body intelligent system architecture and detection robot
SG10201913005YA (en)*2019-12-232020-09-29Sensetime Int Pte LtdMethod, apparatus, and system for recognizing target object
CN112016202B (en)*2020-08-272022-01-04湖南三一工业职业技术学院 A Reliability Allocation Method for Industrial Robots
WO2022051726A1 (en)*2020-09-072022-03-10The Board Of Trustees Of The Leland Stanford Junior UniversityCompact paired parallel architecture for high-fidelity haptic applications
EP4210860A4 (en)2020-09-082025-05-21MacDougall, Frederick William Coalification and Carbon Sequestration Using Deep Ocean Hydrothermal Vents
US11794893B2 (en)2020-09-082023-10-24Frederick William MacDougallTransportation system for transporting organic payloads
US12076099B2 (en)2021-07-062024-09-03Verb Surgical Inc.Projection operator for inverse kinematics of a surgical robot for low degree of freedom tools
US12023117B2 (en)*2021-07-062024-07-02Verb Surgical Inc.Projection of user interface pose command to reduced degree of freedom space for a surgical robot
US11919161B2 (en)*2021-10-152024-03-05Fanuc CorporationGrasp generation for machine tending
WO2023119657A1 (en)*2021-12-242023-06-29日本電信電話株式会社Contact target estimation device, method, and program
WO2025058992A1 (en)*2023-09-122025-03-20University Of WashingtonWireless microrobot and operating techniques

Citations (7)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20080103639A1 (en)*2006-10-252008-05-01The Boeing CompanySystems and Methods for Haptics-Enabled Teleoperation of Vehicles and Other Devices
US7599071B2 (en)*2005-04-062009-10-06Dimensional Photonics International, Inc.Determining positional error of an optical component using structured light patterns
US8049892B2 (en)*2008-01-222011-11-01Honeywell International Inc.Apparatus and method for camera-based color measurements
WO2012174406A1 (en)*2011-06-152012-12-20University Of WashingtonMethods and systems for haptic rendering and creating virtual fixtures from point clouds
US8508646B2 (en)*2008-12-222013-08-13Apple Inc.Camera with internal polarizing filter
US20130274596A1 (en)*2012-04-162013-10-17Children's National Medical CenterDual-mode stereo imaging system for tracking and control in surgical and interventional procedures
US9001190B2 (en)*2011-07-052015-04-07Microsoft Technology Licensing, LlcComputer vision system and method using a depth sensor

Family Cites Families (34)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US6597347B1 (en)1991-11-262003-07-22Itu Research Inc.Methods and apparatus for providing touch-sensitive input in multiple degrees of freedom
US6963792B1 (en)1992-01-212005-11-08Sri InternationalSurgical method
EP0653078B1 (en)1992-08-011998-11-18Siemens AktiengesellschaftMethod and process control system for controlling, monitoring and regulating in particular complex industrial processes, such as for example in a nuclear power station
US5629594A (en)1992-12-021997-05-13Cybernet Systems CorporationForce feedback system
WO1995020788A1 (en)1994-01-271995-08-03Exos, Inc.Intelligent remote multimode sense and display system utilizing haptic information compression
US6084587A (en)1996-08-022000-07-04Sensable Technologies, Inc.Method and apparatus for generating and interfacing with a haptic virtual reality environment
US6191796B1 (en)1998-01-212001-02-20Sensable Technologies, Inc.Method and apparatus for generating and interfacing with rigid and deformable surfaces in a haptic virtual reality environment
US6421048B1 (en)1998-07-172002-07-16Sensable Technologies, Inc.Systems and methods for interacting with virtual objects in a haptic virtual reality environment
US6704694B1 (en)1998-10-162004-03-09Massachusetts Institute Of TechnologyRay based interaction system
US7084867B1 (en)1999-04-022006-08-01Massachusetts Institute Of TechnologyHaptic interface system for collision detection and applications therefore
US7084869B2 (en)2000-03-312006-08-01Massachusetts Institute Of TechnologyMethods and apparatus for detecting and correcting penetration between objects
US8010180B2 (en)2002-03-062011-08-30Mako Surgical Corp.Haptic guidance system and method
AU2003257309A1 (en)2002-08-132004-02-25Microbotics CorporationMicrosurgical robot system
JP4174825B2 (en)2003-05-232008-11-05株式会社テクノリンク Biological stimulator
US20050148432A1 (en)2003-11-032005-07-07Carmein David E.E.Combined omni-directional treadmill and electronic perception technology
DE102004043514A1 (en)2004-09-082006-03-09Sick Ag Method and device for controlling a safety-related function of a machine
US8398541B2 (en)2006-06-062013-03-19Intuitive Surgical Operations, Inc.Interactive user interfaces for robotic minimally invasive surgical systems
US7689320B2 (en)*2005-12-202010-03-30Intuitive Surgical Operations, Inc.Robotic surgical system with joint motion controller adapted to reduce instrument tip vibrations
US8150142B2 (en)2007-04-022012-04-03Prime Sense Ltd.Depth mapping using projected patterns
US8493496B2 (en)2007-04-022013-07-23Primesense Ltd.Depth mapping using projected patterns
WO2009045827A2 (en)2007-09-302009-04-09Intuitive Surgical, Inc.Methods and systems for tool locating and tool tracking robotic instruments in robotic surgical systems
KR101652535B1 (en)2008-06-182016-08-30오블롱 인더스트리즈, 인크Gesture-based control system for vehicle interfaces
US9439736B2 (en)2009-07-222016-09-13St. Jude Medical, Atrial Fibrillation Division, Inc.System and method for controlling a remote medical device guidance system in three-dimensions using gestures
US20110066406A1 (en)2009-09-152011-03-17Chung Yuan Christian UniversityMethod for Generating Real-Time Haptic Response Information for a Haptic Simulating Device
US8392342B2 (en)*2009-11-182013-03-05Empire Technology Development LlcMethod and apparatus for predicting movement of a tool in each of four dimensions and generating feedback during surgical events using a 4D virtual real-time space
US9396545B2 (en)*2010-06-102016-07-19Autodesk, Inc.Segmentation of ground-based laser scanning points from urban environment
US20140043329A1 (en)*2011-03-212014-02-13Peng WangMethod of augmented makeover with 3d face modeling and landmark alignment
EP2752019B1 (en)2011-08-312019-12-11Google LLCMethod and system for providing efficient and accurate estimates of tv viewership ratings
KR101850027B1 (en)*2011-12-082018-04-24한국전자통신연구원Real-time 3-dimension actual environment reconstruction apparatus and method
US9639156B2 (en)2011-12-292017-05-02Mako Surgical Corp.Systems and methods for selectively activating haptic guide zones
US9183676B2 (en)*2012-04-272015-11-10Microsoft Technology Licensing, LlcDisplaying a collision between real and virtual objects
HK1211820A1 (en)*2012-08-242016-06-03University Of HoustonRobotic device and systems for image-guided and robot-assisted surgery
US20140320629A1 (en)2013-01-242014-10-30University Of Washington Through Its Center For CommericializationHaptically-Enabled Co-Robotics for Underwater Tasks
EP3114677B1 (en)2014-03-032020-08-05University of WashingtonHaptic virtual fixture tools

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US7599071B2 (en)*2005-04-062009-10-06Dimensional Photonics International, Inc.Determining positional error of an optical component using structured light patterns
US20080103639A1 (en)*2006-10-252008-05-01The Boeing CompanySystems and Methods for Haptics-Enabled Teleoperation of Vehicles and Other Devices
US8049892B2 (en)*2008-01-222011-11-01Honeywell International Inc.Apparatus and method for camera-based color measurements
US8508646B2 (en)*2008-12-222013-08-13Apple Inc.Camera with internal polarizing filter
WO2012174406A1 (en)*2011-06-152012-12-20University Of WashingtonMethods and systems for haptic rendering and creating virtual fixtures from point clouds
US9001190B2 (en)*2011-07-052015-04-07Microsoft Technology Licensing, LlcComputer vision system and method using a depth sensor
US20130274596A1 (en)*2012-04-162013-10-17Children's National Medical CenterDual-mode stereo imaging system for tracking and control in surgical and interventional procedures

Cited By (53)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US9471142B2 (en)2011-06-152016-10-18The University Of WashingtonMethods and systems for haptic rendering and creating virtual fixtures from point clouds
US20150145960A1 (en)*2011-11-042015-05-28Empire Technology Development LlcIr signal capture for images
US9398288B2 (en)*2011-11-042016-07-19Empire Technology Development LlcIR signal capture for images
US9753542B2 (en)2013-01-242017-09-05University Of Washington Through Its Center For CommercializationMethods and systems for six-degree-of-freedom haptic interaction with streaming point data
US9477307B2 (en)2013-01-242016-10-25The University Of WashingtonMethods and systems for six degree-of-freedom haptic interaction with streaming point data
US20140300702A1 (en)*2013-03-152014-10-09Tagir SaydkhuzhinSystems and Methods for 3D Photorealistic Automated Modeling
US9495759B2 (en)*2014-02-262016-11-15Apeiros, LlcMobile, wearable, automated target tracking system
US20160189391A1 (en)*2014-02-262016-06-30Apeiros, LlcMobile, wearable, automated target tracking system
US10226869B2 (en)2014-03-032019-03-12University Of WashingtonHaptic virtual fixture tools
US9469031B2 (en)*2014-04-242016-10-18Toyota Jidosha Kabushiki KaishaMotion limiting device and motion limiting method
US20170048494A1 (en)*2014-04-242017-02-16Cathx Research LtdUnderwater surveys
US20150306767A1 (en)*2014-04-242015-10-29Toyota Jidosha Kabushiki KaishaMotion limiting device and motion limiting method
AU2015250746B2 (en)*2014-04-242020-02-20Cathx Research LtdUnderwater surveys
US9579799B2 (en)*2014-04-302017-02-28Coleman P. ParkerRobotic control system using virtual reality input
US20150314440A1 (en)*2014-04-302015-11-05Coleman P. ParkerRobotic Control System Using Virtual Reality Input
US9592608B1 (en)*2014-12-152017-03-14X Development LlcMethods and systems for providing feedback during teach mode
US9919416B1 (en)*2014-12-152018-03-20X Development LlcMethods and systems for providing feedback during teach mode
US9919422B1 (en)2016-01-062018-03-20X Development LlcMethods and systems to provide mechanical feedback during movement of a robotic system
US11278369B2 (en)*2016-04-282022-03-22Sony CorporationControl device, control method, and surgical system
US20180234624A1 (en)*2017-02-152018-08-16Samsung Electronics Co., Ltd.Electronic device and method for determining underwater shooting
CN108427533A (en)*2017-02-152018-08-21三星电子株式会社The method of electronic equipment and environment for determining electronic equipment
US11042240B2 (en)*2017-02-152021-06-22Samsung Electronics Co., LtdElectronic device and method for determining underwater shooting
US20180231973A1 (en)*2017-02-162018-08-16Wal-Mart Stores, Inc.System and Methods for a Virtual Reality Showroom with Autonomous Storage and Retrieval
WO2018151908A1 (en)*2017-02-162018-08-23Walmart Apollo, LlcSystems and methods for a virtual reality showroom with autonomous storage and retrieval
US12019159B2 (en)2017-05-042024-06-253D at Depth, Inc.Systems and methods for monitoring underwater structures
US11249193B2 (en)2017-05-042022-02-153D at Depth, Inc.Systems and methods for monitoring underwater structures
US10698112B2 (en)2017-05-042020-06-303D at Depth, Inc.Systems and methods for monitoring underwater structures
US10871567B2 (en)2017-07-102020-12-223D at Depth, Inc.Underwater optical positioning systems and methods
WO2019014245A1 (en)*2017-07-102019-01-173D at Depth, Inc.Underwater optical positioning systems and methods
US11125875B2 (en)2017-07-102021-09-213D at Depth, Inc.Underwater optical metrology system
US10545233B1 (en)2017-07-102020-01-283D at Depth, Inc.Underwater optical metrology system
US10502829B2 (en)2017-07-102019-12-103D at Depth, Inc.Underwater optical metrology system
US12169240B2 (en)2017-07-102024-12-173D at Depth, Inc.Underwater optical positioning systems and methods
US12146951B2 (en)*2017-07-102024-11-193D at Depth, Inc.Underwater optical metrology system
US11774586B2 (en)2017-07-102023-10-033D at Depth, Inc.Underwater optical metrology system
US20230393271A1 (en)*2017-07-102023-12-073D at Depth, Inc.Underwater optical metrology system
US12421034B2 (en)*2017-11-172025-09-23Ocado Innovation LimitedControl device and method for a robot system
US20210171283A1 (en)*2017-11-172021-06-10Ocado Innovation LimitedControl device and method for a robot system
US11787631B2 (en)*2017-11-172023-10-17Ocado Innovation LimitedControl device and method for a robot system
CN112131921A (en)*2019-06-252020-12-25海盛科技有限公司 Stereo vision-based biological automatic measurement system and its measurement method
US20230003516A1 (en)*2019-11-152023-01-05Lufthansa Technik AgBorescope with pattern projection
US11619486B2 (en)*2019-11-152023-04-04Lufthansa Technik AgBorescope with pattern projection
US20240286279A1 (en)*2020-04-242024-08-29Deutsches Zentrum für luft-und Raumfajrt e.V.Method for controlling a robot, and system
US20230210613A1 (en)*2020-06-122023-07-06Covidien LpSurgical robotic system with motion integration
US12347575B2 (en)2020-09-252025-07-013D at Depth, Inc.Systems and methods for laser inspection and measurements
US12087007B2 (en)*2021-03-312024-09-10Auris Health, Inc.Vision-based 6DOF camera pose estimation in bronchoscopy
US20220319031A1 (en)*2021-03-312022-10-06Auris Health, Inc.Vision-based 6dof camera pose estimation in bronchoscopy
US20240302900A1 (en)*2021-06-252024-09-12Keio UniversityOperational data management system, operational data management method, and storage medium
US20240025049A1 (en)*2022-07-252024-01-25Shanghai Flexiv Robotics Technology Co., Ltd.Robot teleoperation method, robot and storage medium
EP4563295A4 (en)*2022-07-252025-10-01Shanghai Flexiv Robotics Tech Co Ltd REMOTE OPERATION METHOD FOR ROBOTS AND ROBOT AND STORAGE MEDIUM
US12165263B1 (en)2022-12-132024-12-10Astrovirtual, Inc.Web browser derived content including real-time visualizations in a three-dimensional gaming environment
US20240430554A1 (en)*2023-06-222024-12-26Integrotech Sp. Z O.O.Device for analysis and recording of an image, especially in an explosion hazard zone
CN117289796A (en)*2023-09-222023-12-26中山大学High-interaction mixed reality system and method for complex equipment based on haptic glove

Also Published As

Publication numberPublication date
US9753542B2 (en)2017-09-05
US9477307B2 (en)2016-10-25
US20140320489A1 (en)2014-10-30
US20170024014A1 (en)2017-01-26
US20140320392A1 (en)2014-10-30

Similar Documents

PublicationPublication DateTitle
US9753542B2 (en)Methods and systems for six-degree-of-freedom haptic interaction with streaming point data
US10226869B2 (en)Haptic virtual fixture tools
US10394327B2 (en)Integration of auxiliary sensors with point cloud-based haptic rendering and virtual fixtures
EP2721463B1 (en)Methods and systems for haptic rendering and creating virtual fixtures from point clouds
Brantner et al.Controlling Ocean One: Human–robot collaboration for deep‐sea manipulation
Delmerico et al.Spatial computing and intuitive interaction: Bringing mixed reality and robotics together
LichiardopolA survey on teleoperation
Pan et al.Fast‐Tracker 2.0: Improving autonomy of aerial tracking with active vision and human location regression
Abdullah et al.Ego-to-exo: Interfacing third person visuals from egocentric views in real-time for improved rov teleoperation
Chavez et al.Robust gesture-based communication for underwater human-robot interaction in the context of search and rescue diver missions
Phung et al.A shared autonomy system for precise and efficient remote underwater manipulation
Abdullah et al.Human-machine interfaces for subsea telerobotics: From soda-straw to natural language interactions
Aldhaheri et al.Underwater Human-Robot and Human-Swarm Interaction: A Review and Perspective
IslamEye on the back: augmented visuals for improved ROV teleoperation in deep water surveillance and inspection
Wu et al.Towards a framework for human factors in underwater robotics
OmaraliExploring robot teleoperation in virtual reality
BrantnerHuman-Robot Collaboration in Challenging Environments
Vilela et al.UAV Control Using Eye Gestures: Exploring the Skies Through Your Eyes
Speers et al.Diver-based control of a tethered unmanned underwater vehicle
US20240300629A1 (en)Shared Autonomy for Remote Collaboration
Sayers et al.A manipulator work package for teleoperation from unmanned untethered vehicles-current feasibility and future applications
XiaVR-Haptic Sensory Augmentation and Body Motion Control for Subsea Robot Teleoperation
LiEnhancing safety in telerobotic surgery via haptic feedback
Brown et al.ROV Teleoperation in the Presence of Cross‐Currents Using Soft Haptics
BillingsVisual Methods Towards Autonomous Underwater Manipulation

Legal Events

DateCodeTitleDescription
ASAssignment

Owner name:UNIVERSITY OF WASHINGTON THROUGH ITS CENTER FOR CO

Free format text:ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CHIZECK, HOWARD JAY;RYDEN, FREDRIK;STEWART, ANDREW;AND OTHERS;SIGNING DATES FROM 20140205 TO 20140226;REEL/FRAME:032349/0622

ASAssignment

Owner name:NATIONAL SCIENCE FOUNDATION, VIRGINIA

Free format text:CONFIRMATORY LICENSE;ASSIGNOR:UNIVERSITY OF WASHINGTON;REEL/FRAME:035156/0575

Effective date:20140225

STCBInformation on status: application discontinuation

Free format text:ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION


[8]ページ先頭

©2009-2025 Movatter.jp