Movatterモバイル変換


[0]ホーム

URL:


US9218643B2 - Method and system for registering images - Google Patents

Method and system for registering images
Download PDF

Info

Publication number
US9218643B2
US9218643B2US14/117,009US201214117009AUS9218643B2US 9218643 B2US9218643 B2US 9218643B2US 201214117009 AUS201214117009 AUS 201214117009AUS 9218643 B2US9218643 B2US 9218643B2
Authority
US
United States
Prior art keywords
image
excision
moving
tissue
update vectors
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US14/117,009
Other versions
US20140079338A1 (en
Inventor
Jeffrey H. Siewerdsen
Sajendra Nithiananthan
Daniel J. Mirota
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Johns Hopkins University
Original Assignee
Johns Hopkins University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Johns Hopkins UniversityfiledCriticalJohns Hopkins University
Priority to US14/117,009priorityCriticalpatent/US9218643B2/en
Assigned to THE JOHNS HOPKINS UNIVERSITYreassignmentTHE JOHNS HOPKINS UNIVERSITYASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS).Assignors: MIROTA, DANIEL J., NITHIANNANTHAN, SAJENDRA, SIEWERDSEN, JEFFREY H.
Assigned to THE JOHNS HOPKINS UNIVERSITYreassignmentTHE JOHNS HOPKINS UNIVERSITYASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS).Assignors: SIEWERDSEN, JEFFREY H., MIROTA, DANIEL J, NITHIANNANTHAN, SAJENDRA
Publication of US20140079338A1publicationCriticalpatent/US20140079338A1/en
Application grantedgrantedCritical
Publication of US9218643B2publicationCriticalpatent/US9218643B2/en
Assigned to NATIONAL INSTITUTES OF HEALTH (NIH), U.S. DEPT. OF HEALTH AND HUMAN SERVICES (DHHS), U.S. GOVERNMENTreassignmentNATIONAL INSTITUTES OF HEALTH (NIH), U.S. DEPT. OF HEALTH AND HUMAN SERVICES (DHHS), U.S. GOVERNMENTCONFIRMATORY LICENSE (SEE DOCUMENT FOR DETAILS).Assignors: JOHNS HOPKINS UNIVERSITY
Activelegal-statusCriticalCurrent
Adjusted expirationlegal-statusCritical

Links

Images

Classifications

Definitions

Landscapes

Abstract

A system for registering images includes an image registration unit. The image registration unit is configured to receive first image data for a first image in an N-dimensional space, receive second image data for a second image in the N-dimensional space, calculate a field of update vectors that maps the first image into a moving image, and map the first image into the moving image using the field of update vectors such that the moving image more closely matches the second image. The field of update vectors includes a plurality of N+M dimensional update vectors, each update vector having N spatial components and M extra components. N is a number greater than zero, and M is a number greater than zero. The M extra components of the plurality of update vectors identify portions of the first image that are assigned external values during the mapping the first image into the moving image.

Description

CROSS-REFERENCE OF RELATED APPLICATION
This is a national stage application under 35 U.S.C. §371 of PCT/US2012/037797 filed May 14, 2012, the entire contents of which are incorporated herein by reference and this application claims priority to U.S. Provisional Application No. 61/485,580 filed May 12, 2011, the entire contents of which are hereby incorporated by reference.
This invention was made with Government support of Grant No. R01-CA-127444, awarded by the Department of Health and Human Services, The National Institutes of Health (NIH). The U.S. Government has certain rights in this invention.
BACKGROUND
1. Field of Invention
The field of the currently claimed embodiments of this invention relates to systems and methods for registering images, and more particularly to systems and methods for registering images in an N dimensional space using N+M dimensions.
2. Discussion of Related Art
Ongoing advances in cone-beam computed tomography (CBCT) enable high-quality 3D imaging for image-guided interventions, including image-guided radiation therapy (IGRT) and image-guided surgery (IGS). The ability to acquire up-to-date images in the course of intervention offers the potential to overcome the limitations of conventional image guidance systems that operate in the context of preoperative data only and fail to account for anatomical change occurring during therapy. Such systems have entered broad application in IGRT,1cardiovascular interventions,2and high-precision IGS including orthopaedic/spine surgery3,4and head and neck/skull base surgery.5-8
Key to many of these applications is the ability to register preoperative images (e.g., preoperative CT along with registered MR or PET images) and planning data to intraoperative CBCT. Such registration algorithms must be sufficiently fast and well integrated so as not to impede surgical workflow and provide sufficient geometric accuracy for the particular interventional task. Rigid registration is often insufficient due to variation in patient positioning between preoperative and intraoperative setups as well as anatomical deformations occurring during intervention. In head and neck/skull base surgery (the focus of examples below), despite the largely rigid anatomical context, rigid registration alone fails to account for independent (piecewise-rigid) motion of the neck, jaw, and skull as well as soft-tissue deformations occurring during the procedure—e.g., displacement of sinus contents, herniation of the orbital wall/lamina papyracea, and deformation of the tongue and oropharynx.
These challenges motivated previous work in developing a variant of the Demons registration method9-13well suited to CBCT-guided procedures.14The method includes: i.) a basic morphological pyramid providing registration within ˜20 s; ii.) a “smart” convergence criterion that automatically advances each level of the pyramid to achieve sub-voxel (˜0.5 mm) registration accuracy and eliminate extraneous iterations;15and iii.) an intensity matching step concurrent with the iterative registration process to provide robustness against image intensity (voxel value) mismatch in CT-to-CBCT or CBCT-to-CBCT registration.16
In addition to the basic challenges of tissue deformation in image-guided procedures is a novel and largely unaddressed problem: what if the differences between the preoperative (“moving”) image and the intraoperative (“fixed”) image involve not only tissue deformation but also the physical removal of mass? In IGRT, an example of a missing tissue problem is weight loss over the course of multiple radiotherapy fractions. In IGS, the problem can be more explicit and includes the physical resection of tissues on surgical approach (e.g., excision of ethmoid air cells on approach to the sphenoid sinus) and removal of the surgical target itself (e.g., drillout of a bony lesion). The behavior of deformable registration algorithms in the presence of missing tissue is an important consideration, since straightforward application of conventional registration approaches may lead to spurious distortion in regions of mismatch—which are often the regions of primary interest (e.g., in proximity to a partially resected surgical target). Registration in the presence of large surgical excisions presents new technical challenges compared to registration in the presence of deformations alone. Similarly, registration in the presence of structures added to the fixed image (e.g., surgical devices) presents a significant challenge. Accordingly, there remains a need for improved systems and methods for registering images.
SUMMARY
A system for registering images according to an embodiment of the current invention includes an image registration unit. The image registration unit is configured to receive first image data for a first image in an N-dimensional space, receive second image data for a second image in the N-dimensional space, calculate a field of update vectors that maps the first image into a moving image, and map the first image into the moving image using the field of update vectors such that the moving image more closely matches the second image. The field of update vectors includes a plurality of N+M dimensional update vectors, each update vector having N spatial components and M extra components. N is a number greater than zero, and M is a number greater than zero. The M extra components of the plurality of update vectors identify portions of the first image that are assigned external values during the mapping the first image into the moving image.
A method of registering images according to an embodiment of the current invention includes receiving first image data for a first image in an N-dimensional space, receiving second image data for a second image in the N-dimensional space, calculating a field of update vectors that maps the first image into a moving image, and mapping the first image into the moving image using the field of update vectors such that the moving image more closely matches the second image. The field of update vectors includes a plurality of N+M dimensional update vectors, each update vector having N spatial components and M extra components. N is an integer greater than zero, and M is an integer greater than zero. The M extra components of the plurality of update vectors identify portions of the first image that are assigned external values during the mapping the first image into the moving image.
A computer-readable medium according to an embodiment of the current invention includes computer-executable code for registering images. The computer-executable code includes instructions that, when executed by the computer, causes the computer to receive first image data for a first image in an N-dimensional space, receive second image data for a second image in the N-dimensional space, calculate a field of update vectors that maps the first image into a moving image, and map the first image into the moving image using the field of update vectors such that the moving image more closely matches the second image. The field of update vectors includes a plurality of N+M dimensional update vectors, each update vector having N spatial components and M extra components. N is an integer greater than zero, and M is an integer greater than zero. The M extra components of the plurality of update vectors identify portions of the first image that are assigned external values during the mapping the first image into the moving image.
BRIEF DESCRIPTION OF THE DRAWINGS
Further objectives and advantages will become apparent from a consideration of the description, drawings, and examples.
FIG. 1 is a schematic illustration of a system and method for registering images according to an embodiment of the current invention. It provides a flow chart illustrating a single iteration of the registration process (a variant of the Demons algorithm, referred to herein as the extra-dimensional Demons (XDD) algorithm) in which segmentation of the excised volume is integrated within the Demons registration framework.
FIGS. 2A-2B provide a schematic illustration of XDD for (a) 2D and (b) 3D image registration according to some embodiments of the current invention. Deformations are represented by in-plane/in-volume vectors (dark shading), and voxels identified within a region of excision (light shading) are subject to out-of-plane/out-of-volume “deformation” (i.e., excision).
FIGS. 3A-3B provide (a) Illustration of tissue and air membership functions used in segmenting regions of excision. The histogram corresponds to a CBCT image of the phantom inFIGS. 5A-5B. (b) The resultant probability map [i.e., joint probability given by the product of membership functions in (a)] as a function of the fixed and moving image intensities.
FIGS. 4A-4D provide illustrations of two example probability maps (Pplan) associated with anticipated regions of tissue excision. (a,b) Example of plan (shading) for drillout of the vidian canal overlaid on axial and sagittal views of a CBCT image. (c,d) Example large plan overlay.
FIGS. 5A-5B show sagittal CBCT images of a deformable phantom (Playfoam™ beads in PlayDoh™ medium) before and after deformation and excision. The excision area (target volume) is highlighted dark shading (manually segmented), and the area immediately outside the excision (adjacent normal tissue) is highlighted in light shading (defined by automatic dilation of the target volume).
FIGS. 6A-6D provide Comparison of conventional Demons and XDD registration in a deformable phantom containing real deformations and a ˜1 cm3excision of material emulating a tissue-in-air excision scenario. Coronal slices from the CBCT volume are shown. (a) Registered image (and zoomed-in region) resulting from the conventional Demons approach, with (b) NCC map (across the zoomed-in region) computed using a 10×10×10 voxel sliding window across the region of interest. (c,d) The same, for the XDD registration approach. Whereas (a-b) exhibit unrealistic distortion and reduction in NCC, (c-d) demonstrates a fairly accurate ejection of voxels within the region of excision and maintenance of NCC (with reduction within the air void likely due to quantum noise).
FIG. 7 shows evolution in the number of voxels excised from the 3D volume in XDD registration. Four levels of the morphological pyramid are evident (DS8, DS4, DS2, and DS1). The horizontal dashed line in each level marks the true number of voxels within the excision volume in the fixed image. The open circles mark the cumulative number of voxels ejected from the volume, whereas the small dots mark the (instantaneous) number of voxels ejected in a given iteration. For the nominal XDD parameters described in the text, a gradual, conservative convergence on the true excision volume is observed (without overshoot).
FIG. 8 shows dependence of deformable registration on the size of excision. A single coronal plane from the 3D volume is shown in each case. (a) Fixed images featuring an excision varying in size from ˜1 cm3(the real excision inFIGS. 5-7) to 2.5-17 cm3(simulated excisions). Registered images and NCC map for the conventional Demons approach are shown in (b,c,d), and those for the XDD approach are shown in (e, f, g).
FIG. 9 shows dependence of deformable registration on the size of excision. While the performance of the conventional Demons algorithm declines sharply with excision size, NMI is maintained to a fairly high degree for the XDD approach. The first point on each curve (excision volume equal to 0 cm3) corresponds to the deformation-only case.
FIG. 10 shows deformable registration in the presence of tissue excision about the clivus. (a) Rigid registration of the preoperative image to the fixed image image shown in (d). (b) Conventional Demons registration exhibits spurious distortion and failure to account for missing tissue. Arrows highlight features of note in the text. (c) XDD registration, demonstrating a close match to the fixed image shown in (d).
FIG. 11 shows deformable registration in the presence of tissue excision along the vidian nerve canal. (a) Rigid registration of the preoperative image to the fixed image shown in (d). (b) Conventional Demons registration exhibits spurious distortion and failure to account for missing tissue. Arrows highlight features of note in the text. (c) XDD registration, demonstrating a close match to the fixed image shown in (d).
FIG. 12 shows deformable registration in the presence of tissue excision throughout the left ethmoid sinus. (a) Rigid registration of the preoperative image to the fixed image image shown in (d). (b) Conventional Demons registration exhibits spurious distortion and failure to account for missing tissue. Arrows highlight features of note in the text. (c) XDD registration, demonstrating a close match to the fixed image shown in (d).
FIG. 13 shows registration performance in cadaver studies emulating CBCT-guided skull base surgery involving drill-out tasks of the clivus, vidian canal, and ethmoid sinuses. (a) NMI for rigid, conventional Demons, and XDD registration. (b) Accuracy in and about the region of excision in the form of “operating curves”—i.e., fraction of target tissue correctly removed (“sensitivity”) plotted versus fraction of adjacent normal correctly preserved (“specificity”).
DETAILED DESCRIPTION
Some embodiments of the current invention are discussed in detail below. In describing embodiments, specific terminology is employed for the sake of clarity. However, the invention is not intended to be limited to the specific terminology so selected. A person skilled in the relevant art will recognize that other equivalent components can be employed and other methods developed without departing from the broad concepts of the current invention. All references cited anywhere in this specification, including the Background and Detailed Description sections, are incorporated by reference as if each had been individually incorporated.
Existing deformable registration methods generally fail to account for material excised or introduced between image acquisitions and typically simply “move” voxels within the images with no ability to account for tissue that is removed (or introduced) between scans. Materials of interest can include, for example, tissue (bone, muscle, fat, tumor, etc.) removed between scans or devices (e.g., needles, packing, staples, etc.) introduced between scans. However, the broad concepts of the current invention are not limited to these particular examples. We have developed an approach in which an extra dimension is added during the registration process to act as a sink for voxels removed between scans. An analogous method would include an extra dimension to serve as a source for voxels introduced between scans. A series of cadaveric images acquired using a prototype mobile C-arm for cone-beam CT (CBCT) was used to model tissue deformation and excision occurring during a surgical procedure according to an embodiment of the current invention, and the ability of deformable registration to correctly account for anatomical changes under these conditions was investigated. Using a previously developed deformable registration algorithm we identified the difficulties that traditional registration algorithms encounter when faced with excised tissue and have developed an algorithm better suited for use in intraoperative image-guided procedures in order that image registration can be performed while accurately accounting for tissue excision while avoiding erroneous deformations around the excision. The addition of extra dimension(s) to the registration problem (e.g., treating a 3D image as a 4D image, with the fourth dimension representing a sink for voxels “ejected” from the image) is shown to resolve spurious distortions suffered by conventional registration approaches.
According to an embodiment of the current invention, we hypothesized that a modified Demons algorithm that explicitly identifies and models tissue excision in the course of registration can provide superior registration performance compared to straightforward application of the conventional Demons approach in images featuring tissue excision. Specifically, an “extra-dimensional” Demons method (denoted XDD) is provided that adds a fourth dimension to the usual 3D registration problem into which voxels corresponding to tissues identified as missing in the fixed image may be ejected from the moving image. Although the focus of description below is on 3D medical images, the approach is general and may include 2D images (e.g., photographs, maps, etc.) and non-medical images.
As a starting point, we delineate a spectrum of scenarios involving missing tissue in surgical guidance: i.) tissue-in-air excision tasks [where tissues of interest include bone (e.g., osteotomy, ethmoidectomy, vidian canal drill-out, and clival drillout) and/or soft tissues (superficial fat/muscle resection, glossectomy, lung nodule resection, mucocele resection, and suction of fluid]); and ii.) soft tissue-in-tissue resection tasks (e.g., resection of a brain tumor, liver metastasis, etc. from within the surrounding medium of normal soft-tissue). In the first, the resected volume results in an air void, whereas in the second, surrounding soft tissue in-fills the resected volume. Initial development of the XDD concept detailed below according to an embodiment of the current invention focuses on the first scenario, in which tissue excision leaves an air volume in place of the missing tissue, although deformations of the surrounding normal tissue may have occurred. The tissue-in-air scenario corresponds to a broad spectrum of surgeries, including head and neck/skull base surgery.
Previous work has identified that for accurate deformable registration in the presence of excisions, the excision area must be identified and accounted for accurate registration,17and identification of “missing data” regions may be performed concurrently with the registration process.18For example, Risholm et al.19,20used a Demons variant for MR-guided brain surgery and recognized that erroneous deformations were calculated within areas of resection, and these errors were found to degrade the overall registration solution. They developed a novel variant of the Demons algorithm for registration of intraoperative MRI brain images in which excisions were identified by a level-set segmentation method, and the Gaussian smoothing step of the Demons algorithm was replaced with a filter designed to accommodate errors caused by retraction of brain tissue and resection of tumor.
Some embodiments of the current invention are distinct in methodology (the addition of an extra dimension in the registration problem), motivation (to reduce the effect of spurious distortions induced by missing tissue, while maintaining the speed and simplicity of the Demons algorithm), and application (CBCT-guided head and neck/skull base surgery). The conventional and XDD forms of the Demons algorithm are tested in phantom and simulation to illustrate and quantify the behavior of deformable registration in the presence of missing tissue in some examples below. Further experimental validation involved a series of cadavers undergoing ethmoidectomy, vidian canal drillout, and clival drillout corresponding to endoscopic skull base surgery. The performance of conventional Demons and XDD was quantified in terms of image-based metrics [e.g., normalized cross-correlation (NCC) and normalized mutual information (NMI) between deformed and fixed image volumes], target registration error (TRE), and the accuracy of voxels identified as either “missing” (and therefore ejected by XDD) or normal (and therefore preserved).
FIG. 1 is a schematic illustration of a system and method for registering images according to an embodiment of the current invention. The system can include an image registration unit that is configured to perform the image registration as illustrated inFIG. 1. The image registration unit can be a computer or portion of a computer, for example, and can be integrated into other equipment. For example, the image registration unit can be integrated into an image-guided intervention system, such as an image-guided radiation therapy (IGRT) or image-guided surgery (IGS) system. For example, it may be integrated into CT, CBCT or magnetic resonance systems. However, the systems and methods according to the current invention are not limited to these examples. Medical imaging applications are one area where embodiments of the current invention can be utilized. However, any system or method in which two or more images are being registered, whether they are two-dimensional or three-dimensional images, could be a potential application of embodiments of the current invention. For example, some embodiments of the current invention can be directed to, but are not limited to, photography, microscopy, satellite imagery, etc.
The image registration unit according to some embodiments of the current invention is configured to receive first image data I0(x) for a first image in an N-dimensional space, receive second image data I1(x) for a second image in the N-dimensional space, calculate a field of update vectors
Figure US09218643-20151222-P00001
(x)M:N+Mthat maps the first image into a moving image, and map the first image into the moving image using the field of update vectors such that the moving image more closely matches the second image. In other words, the first image is modified using the update vectors such that the resultant modified image is a better match to the second image than the original first image was.FIG. 1 is an example of N=3 and M=1, but the general concepts of the current invention are not limited to this example. The field of update vectors
Figure US09218643-20151222-P00001
(x)M:N+Minclude a plurality of N+M dimensional update vectors, each update vector having N spatial components and M extra components. N is an integer greater than zero and M is also an integer greater than zero. The M extra components of the plurality of update vectors identify portions of the first image that are assigned external values during the mapping the first image into the moving image.
The process of calculating update vectors and modifying the original or last modified image and mapping into another moving image can be repeated a plurality of times to iteratively improve matching of successive moving images to the second image. The plurality of times can be a predetermined number. In other words, the loop inFIG. 1 would terminate after the specified number of iterations have been performed. In another embodiment, the plurality of times can be determined based on a measure of change between successive iterations. Additional functions can be performed on the moving image such as filtering, e.g., smoothening, as is indicated inFIG. 1.
In some embodiments of the current invention, the portions of the first image that are assigned external values during the mapping into the moving image correspond to at least one of removal or addition of image portions to the first image relative to the second image. In some embodiments, the first image includes an image of an object and the second image includes an image of the object in which a section of the object is absent. In this embodiment, the portions of said first image that are assigned external values during the mapping into the moving image correspond to the section of the object that is absent in the second image. In some embodiments, the assigned external values during the mapping into the moving image correspond to an object introduced into the second image that was absent from the first image. For example, a surgical tool may be present in the second image but not in the first. This is just one illustrative example and not intended to limit the broad concepts of the current invention.
In some embodiments of the current invention, the M extra components of the update vectors are assigned values based on a probability that an external value is to be assigned during the mapping the first image into the moving image. In some embodiments, the probability can include a weighting factor to weight some of the plurality of update vectors more strongly for assigning the external value than other of the plurality of update vectors.
In some embodiments of the current invention, N=3 and M=1 as is illustrated in the example ofFIG. 1. In some embodiments, the first image can include an image of tissue and the second image can include the image of the tissue that has a portion excised. In this case, the portions of the first image that are assigned external values during the mapping into the moving image are assigned background image values corresponding to absence of tissue to correspond to the portion excised. In some of the examples below, and in the example ofFIG. 1, such introduced values are referred to as “air”.
The following examples will help further describe some concepts of the current invention as well as describe some particular applications. However, the broad concepts of the current invention are not limited to these particular examples.
EXAMPLES
Registration Methods
Conventional Demons Registration
A previously reported variant of the symmetric-force Demons algorithm was employed as a reference and basis of comparison.9,12,14The method has been validated in phantom, cadaver, and clinical studies to provide geometric accuracy to within approximately the voxel size (˜0.5 mm) for typical deformations encountered in CBCT-guided head and neck surgery. The Demons registration framework consists of a series of iterative steps. First, an update vector field is calculated based on currently overlapping voxel intensities. Second, the update field is added to the existing solution (initialized either by a zero field or by rigid registration), and the resulting deformation vector field is regularized by smoothing with a Gaussian kernel. Finally, the smoothed deformation field is used to update the moving image, and the process is repeated with the updated version of the moving image.
The update field,
Figure US09218643-20151222-P00001
, is calculated based on the image intensities of the moving and fixed images, denoted I0and I1respectively, at each overlapping position in the images:
u(x)=2[I0(x)-I1(x)][I0(x)+I1(x)]K[I0(x)-I1(x)]2+I0(x)+I1(x)2(1)
where the normalization factor K is the reciprocal of the squared voxel size (units 1/mm2).11The update field is then added to the existing solution, and the result is convolved with a Gaussian kernel such that the deformation field,
Figure US09218643-20151222-P00002
, at iteration n is given by:
Figure US09218643-20151222-P00002
n=Gσ*(
Figure US09218643-20151222-P00001
n-1+
Figure US09218643-20151222-P00001
n),  (2)
where * indicates 3D convolution, and the width of the Gaussian kernel σ was fixed to 1 voxel for all experiments described below. The deformation field is then used to update the moving image, and the registration process is repeated iteratively.
We consider the “moving” image, I0, as the image obtained at an earlier time-point that is to be registered to a more up-to-date image. Similarly, the “fixed” image, I1, is the up-to-date image to which the moving image is to be matched. For example, the moving image could be a preoperative CT, and the fixed image an intraoperative CBCT obtained at the time of surgery, allowing registration of preoperative imaging and planning data to the intraoperative scene. Alternatively, the moving image could be a CBCT acquired at the beginning of a procedure (into which CT and planning data have been registered as in the previous sentence), and the fixed image a CBCT acquired at a later point in the procedure. In this manner, registration continually moves image and planning data “forward” to the most up-to-date geometric context that matches the state of the patient at any point in the procedure.
Registration is carried out according to a multi-resolution morphological pyramid to improve registration speed and robustness against local minima. The images are binned and downsampled by factors of 8, 4, 2, and 1 (denoted DS8, DS4, DS2, and DS1, respectively), and registration is performed first at the coarsest level, with the result at each level initializing registration in the subsequent level at finer detail. In the examples described below, the number of registration iterations was set at (15, 25, 25, and 15) for (DS8, DS4, DS2, and DS1, respectively) based on observation of the convergence rate. Previous work15demonstrated an optimized convergence criterion based on the difference between subsequent deformation fields (as opposed to a fixed number of iterations per level); however, the current examples fixed the number of iterations for all cases to maintain consistency across experiments and study the effect of the “extra-dimensional” component specifically. Previous work16also demonstrated a Demons variant with an iterative intensity matching process that overcomes errors associated with image intensity mismatch between the fixed and moving images; however, since all cases considered below involved CBCT-to-CBCT registration of images acquired on the same, calibrated CBCT imaging system, the iterative intensity match process was not needed.
Extra-Dimensional Demons (XDD) Registration
Extra Dimension
A novel concept in XDD is to account for tissue excision by an increase in the dimensionality of the solution from three dimensions to include a fourth dimension into which voxels may be moved if identified as belonging to an excision (or conversely from which air may enter the image in the region of the excision). In conventional Demons registration, the update and deformation fields consist of vectors of the same dimensionality as the image space. Therefore, registration of two 3D image volumes involves a 3D deformation field made up of 3D vectors. We denote the 3D volume domain as R3with coordinates (x,y,z) or equivalently (x1,x2,x3). Conventional deformable registration therefore operates entirely within R3. The framework is extended according to an embodiment of the current invention so that the solution for a 3D registration is a 3D deformation field made up of 4D vectors. As illustrated inFIG. 2, this effectively adds a fourth dimension (for 3D image registration) into which voxels may be “ejected” (or equivalently, from which voxels with image intensity corresponding to air enter the volume). As detailed below, motion along the fourth component is only considered if a voxel satisfies specific criteria identifying it as part of an excision. Within the XDD framework, therefore, the vectors comprising the fields given by Eqs. (1) and (2) are 4-vectors, with the first three components representing deformation in (x1,x2,x3), and the fourth component representing excision.FIG. 2 illustrates the method in (a) a 2D image registration (more easily visualized within the constraints of 2D display) wherein excised regions are ejected from R2(x1,x2) to a R3hyperplane, and (b) a 3D image registration in which excised regions are ejected from R3 (x1,x2,x3) to a R4hypercube.
The main challenge in such an approach is assigning the value of the fourth component of the deformation field so that only those voxels truly representative of excised tissue are removed during the registration process. Since the gradient between R3and R4is intrinsically steep, the entire 3D image would tend to move to R4without constraint if the standard Demons force was applied in all dimensions. We therefore first identify the region likely to be associated with excision automatically by segmentation performed simultaneous to the iterative registration process. By performing segmentation simultaneously with the iterative registration process (as opposed to using a single initial segmentation), the process allows progressively more refined estimates of the excision volume as other tissue is deformed and improves robustness to segmentation error. This allows a fairly simple segmentation method detailed below—intensity thresholding of tissue to be ejected versus air (i.e., the material remaining in the voxel for the tissue-in-air “drill-out” scenario). It also allows the possibility of recovery from segmentation error by bringing voxels back to R3if erroneously ejected in previous iterations.
The method according to this embodiment is illustrated inFIG. 1, showing the segmentation and 4D force calculation modifications to the conventional Demons framework. At each registration iteration shown inFIG. 1, the estimate of voxels to be excised is based on the currently registered set of images. Segmentation of voxels possibly to be “ejected” employs a probability (membership function) computed from the image intensities. In subsequent registration steps, the magnitude of the “out-of-volume” component of the update force vector field is calculated proportional to the excision probability. The segmentation and extra-dimensional force calculation are detailed in the following sections.
Identifying Excised Voxels
Initial implementation of XDD used a fairly simple segmentation method to identify candidate voxels for “ejection” based on an air-tissue intensity threshold appropriate to the tissue-in-air excision scenario (e.g., bone drill-out task). Given an intensity threshold T, two probabilistic membership functions (i.e., the probability that a voxel is associated with tissue or air) are computed at every location in the images as follows:
Pmovingtissue(x)=sigm(I0(x),α,T)  (3)
Pfixedair(x)=sigm(I1(x),−α,T)  (4)
where the sigmoid function is:
sigm(I,α,T)=11+-α(I-T)(5)
The membership function Pmovingtissueis the probability that a particular voxel in the moving image corresponds to tissue (not air), and Pfixedairis the probability that a particular voxel in the fixed image corresponds to air (not tissue). The parameter α controls the “steepness” of the sigmoid, and the probability Pexcisionof a given voxel belonging to excised tissue is given by the joint probability of Eqs. (3) and (4):
Pexcision(x)=Pmovingtissue(xPfixedair(x)  (6)
Simply interpreted, Eq. (6) states that if a given voxel in the moving image is likely tissue, and that same voxel in the fixed image is likely air, then that voxel likely corresponds to “missing tissue” (i.e., tissue removed from the volume between the moving and fixed images), and thus it is assigned a higher probability of being excised. This interpretation is strictly valid only if images are perfectly registered and would not be expected to hold if images are misaligned—thus the motivation for incorporating the segmentation within the iterative registration process as shown inFIG. 1 to allow improved segmentation in each iteration.FIG. 3 shows example tissue and air membership functions along with the resulting excision probability “map” (i.e., Pexcisioncomputed as a function of the fixed and moving image intensity). The example calculation corresponds to the deformable phantom used in studies detailed below, with T=0.003 mm−1and α=1000.
However, note that the sigmoid functions are just one example of a way of segmenting tissue. Other functions, including a simple scalar threshold value for example, and other, more sophisticated methods of image segmentation can be used in other embodiments. For example, segmentation can be based not only on the image value but also on parameters of shape, connectedness, etc. Segmentation is a very large and sophisticated field, and the sigmoids are just one example.
Incorporation of a Surgical Plan
The product of probabilities in Eq. (6) treats all areas of the image as equally likely to experience an excision; however, this is an oversimplification as there is almost always a volume of interest (i.e., the region of surgical approach and the volume about the surgical target) which is known a priori to be more likely to be excised. Direct implementation of Eq. (6) was found to be somewhat prone to false positives (i.e., voxels falsely identified as “excised” and ejected from the volume) arising from image noise, artifact, or gross misregistration. To mitigate such effects, and to use available prior knowledge where available, a probabilistic mask (Pplan, referred to as the surgical “plan”) was included in the joint probability. The mask acts as a weighting function that increases the probability of excision about the site of anticipated approach and excision and decreases it at locations distant from the excision:
Pexcision(x)=Pmovingtissue(xPfixedair(xPplan(x)  (7)
The mask function was defined simply as a long-range Gaussian cloud with peak value 1.0 centered roughly on the expected excision site and falling off with distance according to a free parameter (σplan) corresponding to the estimated size of the excision, thus constraining excisions to a region local to the expected surgical site. As shown inFIGS. 4A-4D for the cadaver experiments described below, the Gaussian mask can be coarsely tuned depending on the extent of anticipated excision. For example, as inFIG. 4A, the surgical plan can be placed narrowly (σplan=10 mm in all directions) about the vidian canal (a structure drilled out and followed on approach to the skull base) and the clivus (the bone volume at the skull base that would be drilled out if infiltrated with tumor). Alternatively, as inFIGS. 4C-4D, the surgical plan can be placed broadly (σplan=15 mm in (LR,SI) directions and 30 mm in the AP direction) to encompass the ethmoid sinuses. In practice, the Pplanmultiplier in Pexcisionwas found to reduce erroneous “salt-and-pepper” excisions far from the actual excision—particularly in areas of image noise and artifact.
Extra-Dimensional Deformation Calculation
Tissue excision is modelled in the registration process by way of an additional (fourth) component of the deformation field vectors, denoted
Figure US09218643-20151222-P00002
4, whereas normal deformations (i.e., movement within R3) are described by the in-volume components
Figure US09218643-20151222-P00002
1:3. When calculating the update field, the conventional Demons algorithm force is computed for the in-volume components as in Eq. (1), with the resulting three-vector reduced in proportion to (1−Pexcision):
Figure US09218643-20151222-P00001
(x)1:3=(1−Pexcision(x))·
Figure US09218643-20151222-P00001
(x)  (8)
In this way, regions far from the excision site (i.e., where Pexcision→0) experience deformations approximating those in the conventional Demons approach.
The extra-dimensional (“out-of-volume”) component of the update field,
Figure US09218643-20151222-P00001
4, is computed based on the joint probabilities Pmovingtissue, Pfixedair, and Pplancalculated in the previous segmentation step as:
Figure US09218643-20151222-P00001
(x)4=Pexcision(xkforce·avox  (9)
such that the fourth component of the update field is a function of the excision probability for each voxel. A (optional) proportionality constant kforceallowed the magnitude of the out-of-volume vector component to be increased or decreased to eject voxels farther or nearer into the R4hypercube at each iteration. The parameter was fixed to a value of kforce=1.0 (dimensionless) in all experiments reported below. The term avoxis the voxel size at the current level of the morphological pyramid. Addition of the 4D update field to the current estimate of the 4D deformation field, followed by 4D Gaussian smoothing, was performed as in conventional Demons registration and with the same smoothing parameters (but with increased dimensionality).
Image Update
Through the process described above, voxels that are identified as excised see the value of the fourth component of the deformation field increase with each iteration. As shown inFIG. 1, it is during the image update step that excisions (if any) are introduced into the moving image. As in conventional Demons registration, in-volume deformations are handled by linear interpolation of surrounding intensity values. Deformations represented by the fourth component of the deformation field (i.e., out-of-volume motion) can analogously be considered as a nearest-neighbor interpolation in which the extra-dimensional space is treated as an air-filled void for the tissue-in-air excision scenario considered here. Therefore, voxels for which the fourth component of the deformation field is large enough (i.e., far enough into the R4hypercube) are assigned an air intensity value upon update of the moving image. The displacement beyond which excision is considered to have occurred constitutes a “horizon” in the fourth dimension, the threshold for which is written as avox*khorizon. The paramater khorizonmay be freely adjusted to modify the “gravity” of the R3space—i.e., the ease with which voxels may escape the volume. Following a sensitivity analysis demonstrating relative invariance of the resulting deformation across a range in khorizon(0.5-1.5), a value of khorizon=1.0 was chosen for all experiments reported below. The implication: a voxel exhibiting a displacement with
Figure US09218643-20151222-P00002
4>avoxis subject to ejection on the current iteration.
Multi-Resolution Implementation
As with conventional Demons, XDD was implemented in a multi-resolution morphological pyramid with the same downsampling factors as described above (DS8, DS4, DS2, and DS1). For multi-scale XDD, the magnitude of the fourth component of the deformation field,
Figure US09218643-20151222-P00002
4, was “reset” to zero between levels of the morphological pyramid to reduce spurious ejection in the first iteration of the new level. A variety of multi-resolution XDD scenarios was examined in which the extra-dimensional aspect of the registration was turned on or off at various levels of the hierarchical pyramid—e.g., using conventional Demons at the DS8 level to provide coarse alignment and turning “on” XDD at the finer scales in DS4, DS2, and DS1.
Experimental Methods
Experiments were carried out to characterize the behavior of both the conventional Demons and the XDD embodiment of the current invention in the presence of excised tissue in the fixed image. Three main experiments were performed. First, a custom deformable phantom was imaged in CBCT, before and after deformation and with a volume of material excised between the “preoperative” (moving) image and the “intraoperative” (fixed) image. Test cases involved deformation-only, excision-only, and deformation+excision. Second, simulation studies were conducted using images of the same phantom in which simulated excisions of varying size were imparted. Such phantom experiments allowed variation and sensitivity analysis of the parameters intrinsic to XDD. Finally, cadaver experiments were performed to quantify and validate XDD registration performance in data featuring realistic anatomy and surgical excisions in the context of CBCT-guided endoscopic skull base surgery.
Phantom and Simulation Studies
As illustrated inFIGS. 5A-5B, a simple deformable (and excisable) phantom was constructed from a mixture of two materials—PlayFoam™ (Educational Insights, Lake Forest Ill.) sculpting beads embedded in a medium of PlayDoh™ (Hasbro, Pawtucket R.I.)—to form a roughly cylindrical shape ˜6 cm in diameter, ˜10 cm in height. This simple model provided high contrast between internal features and the background medium (˜400 HU) which could be both deformed and excised. The visibility of individual PlayFoam™ beads (each ˜1-2 mm diameter) allowed visual assessment of deformations and spurious distortion that would not be evident in a uniform phantom. As a further aid to quantifying registration accuracy, a collection of 3.2 mm diameter acrylic spheres were mixed within the phantom to provide uniquely identifiable “target” points in analysis of TRE.
The phantom was manually deformed (without excision) to emulate a deformation-only case, with CBCT images acquired before (I0) and after (Idef) the deformation. Subsequently, as shown inFIG. 5B, an irregular volume (˜1 cm3) was excised from the midsection of the phantom using a curette and tweezers (without further deforming the phantom), and a third CBCT image (Idef+exc) was acquired after the excision. The resulting three CBCT image pairs (I0−Idef, Idef−Idef+exc, and I0−Idef+exc) therefore emulate cases of deformation-only, excision-only, and deformation+excision. Registration was performed with both the conventional Demons and XDD methods in each case. Registration accuracy was quantified globally (i.e., across the entire image) and locally (i.e., within the immediate area of the excision) as described below.
The CBCT images acquired in the phantom study were then extended to a simulation study to investigate the effect of excision size on registration accuracy and the robustness of XDD registration in the presence of larger excisions. The simulated data were formed from the Idef+excimage in which the excision volume was manually delineated (shown in dark shading inFIGS. 5A-5B). Simulated images were formed in which the excision volume was increased in size by dilating the excision volume with kernels of increasing size, thus expanding the excision to arbitrary size by digitally “blanking” the dilated region to the image intensity of air (0 mm−1). Simulated excision volumes ranging in size from 2.5 cm3to 17 cm3were investigated. Deformable registration was performed to match the I0image to each of the simulated Idef+excimages, and registration accuracy was evaluated as described below.
For both the real phantom and simulation studies, the segmentation parameters shown inFIGS. 3A-3B were held fixed, giving a threshold near the air boundary and with relatively sharp transition from air to tissue. Other parameters (e.g., kforceand khorizon) were varied in sensitivity analysis, but none showed dramatic variation in the resulting deformation over the range investigated and the nominal values mentioned above were held fixed.
Registration accuracy was evaluated qualitatively and quantitatively. Visually, the quality of registration was assessed by examining the extent of distortion in the registered image—e.g., apparent shearing of the circular PlayFoam™ beads. Quantitatively, the registration accuracy was assessed in terms of the TRE measured as the distance between the centroids of the implanted acrylic spheres in the fixed and moving images before and after registration. In addition, the registration accuracy was assessed in terms of image-based metrics, NCC and NMI, computed across the entire image (global assessment) and in a smaller region about the excision. As shown inFIGS. 5A-5B, NMI was measured in the combined regions of excision (dark, central shading) and adjacent “normal” tissue (surrounding, lighter shading). The area immediately about the excision arguably requires the most accurate registration (e.g., to assess proximity of the drill-out to nearby critical anatomy, such as the carotid arteries in a clival drill-out), and it was hypothesized that this region would be most susceptible to spurious distortion in the conventional Demons approach.
Cadaver Study
Finally, cadaver experiments were performed using three fixed human head specimens undergoing endoscopic skull base surgery guided by C-arm CBCT.21-23A fellowship-trained neurosurgeon and ENT surgeon performed drill-out tasks pertinent to trans-sphenoid skull base surgery in each head, and CBCT images were acquired before and after intervention. Three drill-out tasks were evaluated (in reverse order): i.) drill-out of the clivus, the bone volume posterior to the sphenoid sinus bounded laterally by the carotid canals and posteriorly by the brain; ii.) drill-out of the vidian nerve canal, the bilateral corridor that provides an important landmark on approach to the clivus; and iii.) excision of the anterior and posterior ethmoid air cells, the bony architecture bilateral to the nasal septum attaching the sinus turbinates and lamina papyracea.
For each cadaver excision task, registration was performed from the pre-excision to post-excision CBCT image. For XDD registration, a surgical plan (mask function) as shown inFIGS. 4A-4D was used to mitigate erroneous excision at large distances from the anticipated site of intervention. The segmentation threshold (T) was selected automatically based on the Otsu method24—specifically, T equal to half the Otsu threshold, generally giving a value near the edge of the air threshold. Setting the sigmoid parameter to α=0.01 maintained approximately the same “steepness” of the threshold functions as in phantom experiments while accounting for the different intensity range in CBCT images reconstructed on an arbitrary 0-4000 scale (as opposed to an attenuation coefficient scale used in the phantom study).
Analysis of registration performance was carried out in a manner similar to that described above for the phantom experiments. The true excision area was manually segmented in the postoperative CBCT image. The area immediately surrounding each excision was defined as the adjacent normal tissue volume in which measures of registration quality were assessed. In addition to the image-based metrics of NCC and NMI computed across the entire image and within local subregions, the accuracy of XDD in correctly ejecting voxels within the excision volume (while preserving voxels in surrounding, un-excised normal tissue) was measured in terms of the excision “sensitivity” (i.e., fraction of excision volume correctly ejected, analogous to true-positive fraction) and “specificity” [i.e., fraction of surrounding normal tissue properly preserved, analogous to (one minus) false-positive fraction].
Results
Phantom and Simulation Studies
Basic Comparison of XDD and Conventional Demons.
FIGS. 6A-6D demonstrate image distortion imparted by the conventional Demons algorithm in the presence of missing tissue, evident as unrealistic warping of features within and adjacent to the region of excision. Registration accuracy at large distances (>˜1 cm) from the boundary of the excision is largely unaffected and exhibits the expected (sub-voxel) precision demonstrated in previous work. The NCC map inFIG. 6B quantifies the failure in registration within and around the region of excision, where the conventional Demons approach appears to force tissue “inward” toward the boundary of the excision, but does not do so in a way that provides an accurate match to the fixed image. The performance of the XDD approach is shown inFIGS. 6C-6D, showing XDD registration results to be largely free of spurious distortion. The identification of excised tissue and modeled ejection of voxels within the iterative registration process provides a close match to the fixed image right up to the excision boundary and in the adjacent “normal” tissue, while maintaining the geometric registration accuracy of the conventional Demons approach in other regions of the image. The NCC map quantifies a higher degree of agreement between the deformed and fixed image, with a decrease in NCC in the central void of the excision likely due to a mismatch to quantum noise in air.
The locality of the excision effect involved in XDD and the maintenance of registration accuracy throughout the rest of the image is evident in the images ofFIGS. 6A-6D, the NMI, and the TRE computed from the entire registered image. Specifically, the NMI computed over the entire image volume was identical (NMI=1.17) for conventional Demons and XDD registration, compared to NMI=1.09 for rigid registration (initialization). Conversely, in the area containing and immediately surrounding the excision, the results indicate degraded registration performance for conventional Demons registration (NMI=1.10) compared to the same region in XDD (NMI=1.16). Similarly, the TRE calculated from 6 acrylic “target” spheres embedded throughout the bulk of the phantom was (0.40±0.15) mm for conventional Demons and (0.40±0.15) mm for XDD, compared to (1.00±0.70) mm for rigid registration. These results verify that XDD maintains a high level of registration accuracy as demonstrated by the conventional Demons approach throughout the image volume (without excision), while improving performance and reducing or eliminating distortion in the region local to the excision.
The evolution of the conventional Demons registration process across multiple levels of the morphological pyramid and iterations has been previously reported,15typified by a fairly monotonic increase in NCC (or other metric) with each iteration (with some variability in the first few iterations of each new level of the pyramid). The evolution of XDD was anticipated to be similar in terms of image metrics, but the approach introduces an additional question regarding the behavior of the excision effect—i.e., at which level of the pyramid (and at which iterations) are voxels “ejected” from the volume. As shown inFIG. 7, simultaneous 3D deformable registration and 4D “excision” exhibits a fairly monotonic evolution at each level of the pyramid. The cumulative excision gradually approaches the “true” excision volume at each level and converges fairly conservatively (i.e., does not overshoot) for the parameters employed. Within a given level of the pyramid, the largest number of voxels is ejected in the first few iterations, followed by a sharp decrease and nearly constant rate of ejection. The scenario illustrated is the nominal case in which XDD was “on” for all levels of the registration pyramid. Variations on this nominal scheme were tested, viz, “off” in the first (DS8) level, and “on” in the three subsequent levels (DS4, DS2, and DS1); however, it was generally found that using XDD at all levels of the morphological pyramid improved registration accuracy and reduced erroneous deformations that were difficult to correct in subsequent levels.
Registration Performance as a Function of Excision Size
FIG. 8 shows the dependence of registration accuracy on the size of the excised volume for both the conventional Demons and XDD approaches. The ˜1 cm3excision is the same as the real data considered in the previous section, and the larger excisions were simulated by dilation as described above. Some degradation in registration quality is evident for the larger excision volumes—visible as image distortion and quantified by the reduction in NMI for each case. The conventional Demons method suffers major distortion and gross misrepresentation of tissue within and around the excision, similar to that inFIG. 6 but amplified steeply for larger excisions. XDD registration on the other hand exhibited a high degree of robustness to excision across the full range of sizes investigated, with only a slight degradation of registration accuracy for the largest excisions. The results are quantified inFIG. 9, where NMI for the conventional Demons approach declines sharply with excision size, but XDD is fairly robust.
Cadaver Study
The results for realistic skull base surgery drill-out tasks conducted in cadaveric specimens are summarized inFIG. 10 (clivus),FIG. 11 (vidian canal), andFIG. 12 (ethmoid sinuses). Overall, the results confirm the findings of the phantom and simulation studies: the conventional Demons approach suffers distortions in the presence of excisions and a quantifiable reduction in NMI and other metrics. The XDD approach accounts for missing tissue and maintains overall geometric accuracy in the deformable registration. A few notable features are highlighted by arrows inFIGS. 10-12. For the clival drill-out (FIG. 10), conventional Demons causes the anterior clival wall to collapse onto the posterior wall, and bony structure lateral to the excision site to collapse unrealistically. XDD yields an image closely matching the fixed image in terms of both the excision site and adjacent anatomy. For the vidian canal drill-out (FIG. 11), conventional Demons collapses and distorts the bony architecture of the middle cranial fossa and pterygopalatine plate to yield a false bony protrusion within the sinus space. XDD properly reproduces the excision and exhibits little or no distortion of surrounding architecture. For the ethmoid air cell excision (FIG. 12), conventional Demons induces spurious deformations and thinning of air cell walls and even introducing an apparent (false) air cell in the excision site. Errors in the deformable registration are evident in soft tissues medial to the excision site. Again, XDD performs comparably well, even partially accounting for differences in fluid filling of the sinuses between the preoperative and intraoperative images.
These qualitative observations are confirmed in the quantitative figures of merit shown inFIG. 13. The NMI demonstrates an increase in registration accuracy for the XDD approach in each case, and the metrics of excision sensitivity and specificity show that the excision areas are correctly modeled to a level of 90% or greater. XDD demonstrates improvement in sensitivity (i.e., fraction of target volume accurately removed) in each case (from ˜60% for conventional Demons to >90% for XDD). We had hypothesized a possible reduction in “specificity” (i.e., fraction of adjacent normal tissue correctly preserved) for XDD due to false-positive ejection of normal tissue voxels; however, XDD demonstrated improved specificity in 2 out of 3 of the surgical tasks investigated (a slight reduction in specificity observed for the ethmoidectomy case). Small areas of failure for the XDD registration are evident in some cases at the edges of the excision area due to slight under-estimation of the excision volume. As in the phantom and simulation studies, the cadaver results suggest the greatest improvement in registration accuracy for XDD over conventional Demons registration for the larger excision volumes.
Discussion and Conclusions
The results presented in these examples indicate that conventional Demons registration is prone to erroneous distortion of registered images in the presence of excisions, and registration accuracy degrades in the area local to the site of the excision, which is the area where increased registration accuracy may be most needed. Explicit modeling of tissue excision implemented in the XDD registration approach resolved such effects to a significant degree. XDD largely removed excision-induced distortion about the site of the excision while maintaining the same level of registration accuracy as conventional Demons registration in regions far from the excision site.
Initial implementation of the XDD method involved simultaneous registration and excision segmentation using a simple thresholds of air-tissue intensity interfaces. Such an approach has advantages of simplicity and computational efficiency, but may be more difficult to optimize for more complex resections (e.g., tissue-in-tissue resection) as well as weighing tradeoffs between false-positive ejection (i.e., normal tissue erroneously ejected from the volume) versus true-negative deformation (i.e., normal tissue properly deformed within three-space). Incorporation of a simple surgical plan (e.g., a large Gaussian cloud marking an additional probability of excision) reduced distal, erroneous “salt-and-pepper” excision and makes reasonable use of prior information without additional computational burden.
Analogous to the missing-tissue problem investigated above is the question: what if the fixed and moving images differ in the physical addition of tissue (or other material)? In IGS, for example, the fixed image (intraoperative CBCT) may contain an interventional device (e.g., needle) not present in the moving (preoperative CT) image. The addition of an extra dimension from which material (i.e., signal intensity) may be introduced to the moving image at voxel locations identified as regions of mismatch according to another embodiment of the current invention.
Accurate account of tissue excision is an important aspect of deformable registration in image-guided surgery. Initial implementation of a Demons variant modified to include extra dimensions in the deformation field provided an accurate means of ejecting voxels from the moving image while maintaining the overall (sub-voxel) accuracy of the conventional Demons approach. Application to preclinical (cadaver) studies of CBCT-guided head and neck/skull base surgery demonstrated a major improvement in registration accuracy under conditions of realistic excision tasks.
REFERENCES
    • 1. Dawson L A, Jaffray D A. Advances in image-guided radiation therapy. J. Clin. Oncol. 2007; 25(8):938-946.
    • 2. Lauritsch G, Boese J, Wigstrom L, Kemeth H, Fahrig R. Towards cardiac C-arm computed tomography. IEEE Trans Med Imaging. 2006; 25(7):922-934.
    • 3. Schafer S, Nithiananthan S, Mirota D J, et al. Mobile C-arm cone-beam CT for guidance of spine surgery: image quality, radiation dose, and integration with interventional guidance. Med Phys. 2011; 38(8):4563-4574.
    • 4. Santos E R G, Ledonio C G, Castro C A, Truong W H, Sembrano J N. The accuracy of intraoperative O-arm images for the assessment of pedicle screw postion. Spine. 2012; 37(2):E119-125.
    • 5. Daly M J, Siewerdsen J H, Moseley D J, Jaffray D A, Irish J C. Intraoperative cone-beam CT for guidance of head and neck surgery: Assessment of dose and image quality using a C-arm prototype. Med Phys. 2006; 33(10):3767-80.
    • 6. Bachar G, Siewerdsen J H, Daly M J, Jaffray D A, Irish J C. Image quality and localization accuracy in C-arm tomosynthesis-guided head and neck surgery. Med Phys. 2007; 34(12):4664-4677.
    • 7. Bachar G, Barker E, Nithiananthan S, et al. Three-dimensional tomosynthesis and cone-beam computed tomography: an experimental study for fast, low-dose intraoperative imaging technology for guidance of sinus and skull base surgery. Laryngoscope. 2009; 119(3):434-441.
    • 8. Balachandran R, Schurzig D, Fitzpatrick J M, Labadie R F. Evaluation of portable CT scanners for otologic image-guided surgery. Int J Comput Assist Radiol Surg. 2012; 7(2):315-321.
    • 9. Thirion J. Fast non-rigid matching of 3D medical images. France: INRIA; 1995. Available at: http://hal.inria.fr/inria-00077268/en/. Accessed Feb. 12, 2009.
    • 10. Thirion J P. Image matching as a diffusion process: an analogy with Maxwell's demons. Med Image Anal. 1998; 2(3):243-260.
    • 11. Pennec X, Cachier P, Ayache N. Understanding the “Demon's Algorithm”: 3D Non-rigid Registration by Gradient Descent. In: Proceedings of the Second International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer-Verlag; 1999:597-605. Available at: http://portal.acm.org/citation.cfm?id=709761. Accessed Apr. 26, 2009.
    • 12. Wang H, Dong L, O'Daniel J, et al. Validation of an accelerated “demons” algorithm for deformable image registration in radiation therapy. Phys Med Biol. 2005; 50(12):2887-2905.
    • 13. Vercauteren T, Pennec X, Perchant A, Ayache N. Diffeomorphic demons: Efficient non-parametric image registration. NeuroImage. 2009; 45(1, Supplement 1):S61-S72.
    • 14. Nithiananthan S, Brock K K, Irish J C, Siewerdsen J H. Deformable registration for intra-operative cone-beam CT guidance of head and neck surgery. In: Engineering in Medicine and Biology Society, 2008. EMBS 2008. 30th Annual International Conference of the IEEE.; 2008:3634-3637.
    • 15. Nithiananthan S, Brock K K, Daly M J, et al. Demons deformable registration for CBCT-guided procedures in the head and neck: Convergence and accuracy. Med. Phys. 2009; 36(10):4755-4764.
    • 16. Nithiananthan S, Schafer S, Uneri A, et al. Demons Deformable Registration of CT and Cone-Beam C T Using an Iterative Intensity Matching Approach. Medical Physics. 2011.
    • 17. Miga M I, Roberts D W, Kennedy F E, et al. Modeling of retraction and resection for intraoperative updating of images. Neurosurgery. 2001; 49(1):75-84; discussion 84-5.
    • 18. Periaswamy S, Farid H. Medical image registration with partial data. Med Image Anal. 2006; 10(3):452-464.
    • 19. Risholm P, Samsett E, Talos I-F, Wells W. A non-rigid registration framework that accommodates resection and retraction. Inf Process Med Imaging. 2009; 21:447-458.
    • 20. Risholm P, Samset E, Wells III W. Validation of a nonrigid registration framework that accommodates tissue resection. In: San Diego, Calif., USA; 2010:762319-762319-11. Available at: http://link.aip.org/link/PSISDG/v7623/i1/p762319/s1&Agg=doi. Accessed Jan. 24, 2011.
    • 21. Jaffray D A, Siewerdsen J H. Cone-beam computed tomography with a flat-panel imager: initial performance characterization. Med Phys. 2000; 27(6):1311-1323.
    • 22. Chan Y, Siewerdsen J H, Rafferty M A, et al. Cone-beam computed tomography on a mobile C-arm: novel intraoperative imaging technology for guidance of head and neck surgery. J Otolaryngol Head Neck Surg. 2008; 37(1):81-90.
    • 23. Siewerdsen J H, Chan Y, Rafferty M A, et al. Cone-beam CT with a flat-panel detector on a mobile C-arm: preclinical investigation in image-guided surgery of the head and neck. In: Medical Imaging 2005: Visualization, Image-Guided Procedures, and Display. Vol 5744. San Diego, Calif., USA: SPIE; 2005:789-797. Available at: http://link.aip.org/link/?PSI/5744/789/1. Accessed Feb. 10, 2009.
    • 24. Sezgin M, Sankur B. Survey over image thresholding techniques and quantitative performance evaluation. J. Electron. Imaging. 2004; 13(1):146-168.
The embodiments illustrated and discussed in this specification are intended only to teach those skilled in the art how to make and use the invention. In describing embodiments of the invention, specific terminology is employed for the sake of clarity. However, the invention is not intended to be limited to the specific terminology so selected. The above-described embodiments of the invention may be modified or varied, without departing from the scope of the invention, as appreciated by those skilled in the art in light of the above teachings. It is therefore to be understood that, within the scope of the claims and their equivalents, the invention may be practiced otherwise than as specifically described.

Claims (42)

We claim:
1. A system for registering images comprising an image registration unit, said image registration unit comprising non-transitory executable code that, when executed by the registration unit, causes the registration unit to:
receive first image data for a first image in an N-dimensional space;
receive second image data for a second image in said N-dimensional space;
calculate a field of update vectors that maps said first image into a moving image; and
map said first image into said moving image using said field of update vectors such that said moving image more closely matches said second image,
wherein said field of update vectors comprises a plurality of N+M dimensional update vectors, each update vector having N spatial components and M extra components,
wherein N is a number greater than zero,
wherein M is a number greater than zero, and
wherein said M extra components of said plurality of update vectors identify portions of said first image that are assigned external values during said mapping said first image into said moving image.
2. A system according toclaim 1, wherein said calculating and said mapping are repeated a plurality of times by said registration unit to iteratively improve matching of successive moving images to said second image.
3. A system according toclaim 2, wherein said plurality of times is a predetermined number.
4. A system medium according toclaim 2, wherein said plurality of times is determined based on a measure of change between successive iterations.
5. A system according toclaim 1, wherein said portions of said first image that are assigned external values during said mapping into said moving image correspond to at least one of removal or addition of image portions to said first image relative to said second image.
6. A system according toclaim 5, wherein said first image comprises an image of an object and said second image comprises an image of said object in which a section of said object is absent,
wherein said portions of said first image that are assigned external values during said mapping into said moving image correspond to said section of said object that is absent in said second image.
7. A system according toclaim 5, wherein said assigned external values during said mapping into said moving image correspond to an object introduced into said second image that was absent from said first image.
8. A system according toclaim 6, wherein N=3 and M=1,
wherein said first image comprises an image of tissue,
wherein said second image comprises said image of said tissue that has a portion excised, and
wherein said portions of said first image that are assigned external values during said mapping into said moving image are assigned background image values corresponding to absence of tissue to correspond to said portion excised.
9. A system according toclaim 1, wherein said M extra components of said update vectors are assigned values based on a probability that an external value is to be assigned during said mapping said first image into said moving image.
10. A system according toclaim 9, wherein said probability comprises a weighting factor to weight some of said plurality of update vectors more strongly for assigning said external value than other of said plurality of update vectors.
11. A system according toclaim 8, wherein said M=1 extra component of said update vectors utilize an overlap probability function given by

Pexcision(x)=Pmovingtissue(x)·Pfixedair(x)
wherein

Pmovingtissue(x)=sigm(I0(x),α,T)

Pfixedair(x)=sigm(I1(x),=α,T)
wherein
sigm(I,α,T)=11+-α(I-T)
wherein I0(x) is image intensity in said first image at x, I1(x) is image intensity in said second image at x, and
wherein α and T are selectable parameters.
12. A system according toclaim 11, wherein said update vectors are given by three spatial components as

{right arrow over (u)}(x)1:3=(1−Pexcision(x))·{right arrow over (u)}(x)
and fourth extra component as

{right arrow over (u)}(x)4=Pexcision(x)·kforce·avox,
wherein
u(x)=2[I0(x)-I1(x)][I0(x)+I1(x)]K[I0(x)-I1(x)]2+I0(x)+I1(x)2,
 and
wherein avoxis a selectable voxel size, K is a normalization factor, and kforceand is a selectable parameter.
13. A system according toclaim 8, wherein said M=1 extra component of said update vectors utilize an overlap probability function given by

Pexcision(x)=Pmovingtissue(x)·Pplan(x)
wherein

Pmoviongtissue(x)=sigm(x),α,T)

Pfixedair(x)=sigm(I1(x),=α,T)
wherein
sigm(I,α,T)=11+-α(I-T)
wherein I0(x) is image intensity in said first image at x, I1(x) is image intensity in said second image at x,
wherein α and T are selectable parameters, and Pplan(x) is a long-range Gaussian cloud with peak value 1.0 centered substantially on an expected excision site.
14. A system according toclaim 13, wherein said update vectors are given by three spatial components as

{right arrow over (u)}(x)1:3=(1−Pexcision(x))·{right arrow over (u)}(x)
and fourth extra component as

{right arrow over (u)}(x)4=Pexcision(x)·kforce·avox,
wherein
u(x)=2[I0(x)-I1(x)][I0(x)+I1(x)]K[I0(x)-I1(x)]2+I0(x)+I1(x)2,
 and
wherein avoxis a selectable voxel size, K is a normalization factor, and kforceand is a selectable parameter.
15. A method of registering images using an image registration unit, said image registration unit comprising non-transitory executable code that, when executed by the registration unit, causes the registration unit to perform the method, comprising;
receiving first image data for a first image in an N-dimensional space;
receiving second image data for a second image in said N-dimensional space;
calculating a field of update vectors that maps said first image into a moving image; and
mapping said first image into said moving image using said field of update vectors such that said moving image more closely matches said second image,
wherein said field of update vectors comprises a plurality of N+M dimensional update vectors, each update vector having N spatial components and M extra components,
wherein N is an integer greater than zero,
wherein M is an integer greater than zero, and
wherein said M extra components of said plurality of update vectors identify portions of said first image that are assigned external values during said mapping said first image into said moving image.
16. A method of registering images according toclaim 15, wherein said calculating and said mapping are repeated a plurality of times to iteratively improve matching of successive moving images to said second image.
17. A method of registering images according toclaim 16, wherein said plurality of times is a predetermined number.
18. A method of registering images according toclaim 16, wherein said plurality of times is determined based on a measure of change between successive iterations.
19. A method of registering images according toclaim 15, wherein said portions of said first image that are assigned external values during said mapping into said moving image correspond to at least one of removal or addition of image portions to said first image relative to said second image.
20. A method of registering images according toclaim 19, wherein said first image comprises an image of an object and said second image comprises an image of said object in which a section of said object is absent,
wherein said portions of said first image that are assigned external values during said mapping into said moving image correspond to said section of said object that is absent in said second image.
21. A method of registering images according toclaim 19, wherein said assigned external values during said mapping into said moving image correspond to an object introduced into said second image that was absent from said first image.
22. A method of registering images according toclaim 20, wherein N=3 and M=1,
wherein said first image comprises an image of tissue,
wherein said second image comprises said image of said tissue that has a portion excised, and
wherein said portions of said first image that are assigned external values during said mapping into said moving image are assigned background image values corresponding to absence of tissue to correspond to said portion excised.
23. A method of registering images according toclaim 15, wherein said M extra components of said update vectors are assigned values based on a probability that an external value is to be assigned during said mapping said first image into said moving image.
24. A method of registering images according toclaim 23, wherein said probability comprises a weighting factor to weight some of said plurality of update vectors more strongly for assigning said external value than other of said plurality of update vectors.
25. A method of registering images according toclaim 22, wherein said M=1 extra component of said update vectors utilize an overlap probability function given by

Pexcision(x)=Pmovingtissue(x)·Pfixedair(x)
wherein

Pmovingtissue(x)=sigm(I0(x),α,T)

Pfixedair(x)=sigm(I1(x),−α,T
wherein
sigm(I,α,T)=11+-α(I-T)
wherein I0(x) is image intensity in said first image at x, I1(x) is image intensity in said second image at x, and
wherein α and T are selectable parameters.
26. A method of registering images according toclaim 25, wherein said update vectors are given by three spatial components as

{right arrow over (u)}(x)1:3=(1−Pexcision(x))·{right arrow over (u)}(x)
and fourth extra component as

{right arrow over (u)}(x)4=Pexcision(x)·kforce·vox,
wherein
u(x)=2[I0(x)-I1(x)][I0(x)+I1(x)]K[I0(x)-I1(x)]2+I0(x)+I1(x)2,
 and
wherein avoxis a selectable voxel size, and kforceand K are selectable parameters.
27. A method of registering images according toclaim 22, wherein said M=1 extra component of said update vectors utilize an overlap probability function given by

Pexcision(x)=Pmovingtissue(x)·Pplan(x)
wherein

Pmovingtissue(x)=sigm(I0(x), α,T)

Pfixedair(x)sigm(I1(x),−α,T)
wherein
sigm(I,α,T)=11+-α(I-T)
wherein I0(x) is image intensity in said first image at x, I1(x) is image intensity in said second image at x,
wherein α and T are selectable parameters, and
Pplan(x) is a long-range Gaussian cloud with peak value 1.0 centered substantially on an expected excision site.
28. A method of registering images according toclaim 27, wherein said update vectors are given by three spatial components as

{right arrow over (u)}(x)1:3=(1−Pexcision(x))·{right arrow over (u)}(x)
and fourth extra component as

{right arrow over (u)}(x)4=Pexcision(x)·kforce·avox
wherein
u(x)=2[I0(x)-I1(x)][I0(x)+I1(x)]K[I0(x)-I1(x)]2+I0(x)+I1(x)2,
 and
wherein avoxis a selectable voxel size, and kforceand K are selectable parameters.
29. A computer-readable medium comprising non-transitory computer-executable code for registering images, said non-transitory computer-executable code comprising instructions that, when retrieved from a non-transitory computer-readable storage medium and executed by a computer, causes said computer to:
receive first image data for a first image in an N-dimensional space;
receive second image data for a second image in said N-dimensional space;
calculate a field of update vectors that maps said first image into a moving image; and
map said first image into said moving image using said field of update vectors such that said moving image more closely matches said second image,
wherein said field of update vectors comprises a plurality of N+M dimensional update vectors, each update vector having N spatial components and M extra components,
wherein N is an integer greater than zero,
wherein M is an integer greater than zero, and
wherein said M extra components of said plurality of update vectors identify portions of said first image that are assigned external values during said mapping said first image into said moving image.
30. A computer-readable medium according toclaim 29, wherein said calculating and said mapping are repeated a plurality of times by said computer to iteratively improve matching of successive moving images to said second image.
31. A computer-readable medium according toclaim 30, wherein said plurality of times is a predetermined number.
32. A computer-readable medium according toclaim 30, wherein said plurality of times is determined based on a measure of change between successive iterations.
33. A computer-readable medium according toclaim 29, wherein said portions of said first image that are assigned external values during said mapping into said moving image correspond to at least one of removal or addition of image portions to said first image relative to said second image.
34. A computer-readable medium according toclaim 33, wherein said first image comprises an image of an object and said second image comprises an image of said object in which a section of said object is absent,
wherein said portions of said first image that are assigned external values during said mapping into said moving image correspond to said section of said object that is absent in said second image.
35. A computer-readable medium according toclaim 33, wherein said assigned external values during said mapping into said moving image correspond to an object introduced into said second image that was absent from said first image.
36. A computer-readable medium according toclaim 34, wherein N=3 and M=1,
wherein said first image comprises an image of tissue,
wherein said second image comprises said image of said tissue that has a portion excised, and
wherein said portions of said first image that are assigned external values during said mapping into said moving image are assigned background image values corresponding to absence of tissue to correspond to said portion excised.
37. A computer-readable medium according toclaim 29, wherein said M extra components of said update vectors are assigned values based on a probability that an external value is to be assigned during said mapping said first image into said moving image.
38. A computer-readable medium according toclaim 37, wherein said probability comprises a weighting factor to weight some of said plurality of update vectors more strongly for assigning said external value than other of said plurality of update vectors.
39. A computer-readable medium according toclaim 36, wherein said M=1 extra component of said update vectors utilize an overlap probability function given by

Pexcision(x)=Pmovingtissue(x)·Pfixedair(x)
wherein

Pmovingtissue(x)=sigm(I0(x),α, T)

Pfixedair(x)=sigm(I1(x),−α, T)
wherein
sigm(I,α,T)=11+-α(I-T)
wherein I0(x) is image intensity in said first image at x, I1(x) is image intensity in said second image at x, and
wherein α and T are selectable parameters.
40. A computer-readable medium according toclaim 39, wherein said update vectors are given by three spatial components as

{right arrow over (u)}(x)1:3=(1−Pexcision(x))·{right arrow over (u)}(x)
and fourth extra component as

{right arrow over (u)}(x)4=Pexercision(x)·kforce·avox
wherein
u(x)=2[I0(x)-I1(x)][I0(x)+I1(x)]K[I0(x)-I1(x)]2+I0(x)+I1(x)2,
 and
wherein avoxis a selectable voxel size, and kforceand K are selectable parameters.
41. A computer-readable medium according toclaim 36, wherein said M=1 extra component of said update vectors utilize an overlap probability function given by

Pexcision(x)=Pmovingtissue(x)·Pfixedair(x)·Pplan(x)
wherein

Pmovingtissue(x)=sigm(I0(x),α,T)

Pfixedair(x)=sigm(I1(x),−α,T)
wherein
sigm(I,α,T)=11+-α(I-T)
wherein I0(x) is image intensity in said first image at x, I1(x) is image intensity in said second image at x,
wherein a and T are selectable parameters, and
Pplan(x) is a long-range Gaussian cloud with peak value 1.0 centered substantially on an expected excision site.
42. A computer-readable medium according toclaim 41, wherein said update vectors are given by three spatial components as

{right arrow over (u)}(x)1:3=(1−Pexcision(X))·{right arrow over (U)}(x)
and fourth extra component as

{right arrow over (u)}(x)4=Pexcision(x)·kforce·avox,
wherein
u(x)=2[I0(x)-I1(x)][I0(x)+I1(x)]K[I0(x)-I1(x)]2+I0(x)+I1(x)2,
 and
wherein avoxis a selectable voxel size, and kforceand K are selectable parameters.
US14/117,0092011-05-122012-05-14Method and system for registering imagesActive2032-09-06US9218643B2 (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
US14/117,009US9218643B2 (en)2011-05-122012-05-14Method and system for registering images

Applications Claiming Priority (3)

Application NumberPriority DateFiling DateTitle
US201161485580P2011-05-122011-05-12
PCT/US2012/037797WO2012155136A2 (en)2011-05-122012-05-14Method and system for registering images
US14/117,009US9218643B2 (en)2011-05-122012-05-14Method and system for registering images

Publications (2)

Publication NumberPublication Date
US20140079338A1 US20140079338A1 (en)2014-03-20
US9218643B2true US9218643B2 (en)2015-12-22

Family

ID=47140054

Family Applications (1)

Application NumberTitlePriority DateFiling Date
US14/117,009Active2032-09-06US9218643B2 (en)2011-05-122012-05-14Method and system for registering images

Country Status (2)

CountryLink
US (1)US9218643B2 (en)
WO (1)WO2012155136A2 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20190213743A1 (en)*2014-08-222019-07-11Analogic CorporationUpdating Reference Imaging Data with Update 2D and/or 3D Imaging Data
US20250054098A1 (en)*2023-08-082025-02-13City University Of Hong KongPreserving Tumor Volumes for Unsupervised Medical Image Registration

Families Citing this family (22)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
WO2012103949A1 (en)*2011-02-032012-08-09Brainlab AgRetrospective mri image distortion correction
CN102959584B (en)*2011-12-212015-03-25中国科学院自动化研究所Function magnetic resonance image registration method
CN104969265A (en)*2012-11-232015-10-07卡丹医学成像股份有限公司 Method and system for displaying a transition between a first rendering projection and a second rendering projection to a user
CN103854276B (en)*2012-12-042018-02-09东芝医疗系统株式会社Image registration and segmenting device and method, and medical image equipment
JP6505078B2 (en)*2013-03-292019-04-24コーニンクレッカ フィリップス エヌ ヴェKoninklijke Philips N.V. Image registration
CN103236059B (en)*2013-04-252016-06-15深圳先进技术研究院Differomorphism demons method for registering images and system based on MODAL TRANSFORMATION OF A
US9449384B2 (en)2014-04-302016-09-20Mitsubishi Electric Research Laboratories, Inc.Method for registering deformable images using random Markov fields
CN104091337B (en)*2014-07-112017-07-14北京工业大学A kind of deformation medical image registration method based on PCA and differomorphism Demons
CN104346786B (en)*2014-11-062017-05-17南京信息工程大学Image denoising algorithm based on Demons algorithm
CN104933716A (en)*2015-06-162015-09-23山东大学(威海)Non-rigid registration method applied to medical image
US10453080B2 (en)*2016-01-272019-10-22International Business Machines CorporationOptimizing registration fields with user engagement score
US10729396B2 (en)2016-08-312020-08-04International Business Machines CorporationTracking anatomical findings within medical images
US20180060535A1 (en)*2016-08-312018-03-01International Business Machines CorporationUpdating probabilities of conditions based on annotations on medical images
US10460488B2 (en)*2016-09-062019-10-29International Business Machines CorporationSpine labeling automation
EP3600112A4 (en)*2017-03-222020-10-28Intuitive Surgical Operations Inc.Systems and methods for intelligently seeding registration
US11175132B2 (en)*2017-08-112021-11-16Zoox, Inc.Sensor perturbation
US10783699B2 (en)*2018-02-162020-09-22AI Analysis, Inc.Sub-voxel refinement of anatomical models
GB2569803B (en)*2017-12-222021-11-24Novarum Dx LtdAnalysis of a captured image to determine a test outcome
CN110930440B (en)*2019-12-092023-06-27Oppo广东移动通信有限公司 Image alignment method, device, storage medium and electronic equipment
CN112837392B (en)*2021-04-082023-01-03上海联影医疗科技股份有限公司CT image generation method, device, system, electronic device and storage medium
CN113223061B (en)*2021-05-312024-04-05温州医科大学Elastic registration method and device for medical image
CN115578335B (en)*2022-09-292023-05-05西安理工大学 Vocal cord leukoplakia image classification method based on multi-scale feature extraction

Citations (12)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20020191696A1 (en)*2001-06-152002-12-19Shankar MoniMethods and systems for reducing requantization-originated generational error in predictive video streams using motion compensation
US6731810B1 (en)*1998-12-242004-05-04Hudson Soft Co., Ltd.Method and apparatus for coding moving image and medium for recording program of coding moving image
US20050063468A1 (en)*2003-09-242005-03-24Kddi CorporationMotion vector detecting apparatus
US20060004275A1 (en)*2004-06-302006-01-05Vija A HSystems and methods for localized image registration and fusion
US7023450B1 (en)*1999-09-292006-04-04Thomson LicensingData processing method and apparatus for a display device
US20080009724A1 (en)*2006-05-162008-01-10Medison Co., Ltd.Ultrasound system for fusing an ultrasound image and an external medical image
US20080080788A1 (en)*2006-10-032008-04-03Janne NordSpatially variant image deformation
US20080123927A1 (en)*2006-11-162008-05-29Vanderbilt UniversityApparatus and methods of compensating for organ deformation, registration of internal structures to images, and applications of same
US20090087124A1 (en)*2007-09-282009-04-02Varian Medical Systems FinlandRadiation systems and methods using deformable image registration
US20090304252A1 (en)*2008-06-052009-12-10Dong Gyu HyunNon-Rigid Registration Between CT Images And Ultrasound Images
US7756359B1 (en)*2006-10-032010-07-13Varian Medical Systems International AgUse of local or global quality of match to estimate cone beam CT reconstruction artifact correction weight in image registration
US20100226440A1 (en)*2009-03-052010-09-09Fujitsu LimitedImage encoding device, image encoding control method, and program

Patent Citations (14)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US6731810B1 (en)*1998-12-242004-05-04Hudson Soft Co., Ltd.Method and apparatus for coding moving image and medium for recording program of coding moving image
US7023450B1 (en)*1999-09-292006-04-04Thomson LicensingData processing method and apparatus for a display device
US20020191696A1 (en)*2001-06-152002-12-19Shankar MoniMethods and systems for reducing requantization-originated generational error in predictive video streams using motion compensation
US20050063468A1 (en)*2003-09-242005-03-24Kddi CorporationMotion vector detecting apparatus
US20060004275A1 (en)*2004-06-302006-01-05Vija A HSystems and methods for localized image registration and fusion
US20080009724A1 (en)*2006-05-162008-01-10Medison Co., Ltd.Ultrasound system for fusing an ultrasound image and an external medical image
US20080080788A1 (en)*2006-10-032008-04-03Janne NordSpatially variant image deformation
US7756359B1 (en)*2006-10-032010-07-13Varian Medical Systems International AgUse of local or global quality of match to estimate cone beam CT reconstruction artifact correction weight in image registration
US20080123927A1 (en)*2006-11-162008-05-29Vanderbilt UniversityApparatus and methods of compensating for organ deformation, registration of internal structures to images, and applications of same
US20090087124A1 (en)*2007-09-282009-04-02Varian Medical Systems FinlandRadiation systems and methods using deformable image registration
US20090304252A1 (en)*2008-06-052009-12-10Dong Gyu HyunNon-Rigid Registration Between CT Images And Ultrasound Images
US8369597B2 (en)*2008-06-052013-02-05Medison Co., Ltd.Non-rigid registration between CT images and ultrasound images
US20100226440A1 (en)*2009-03-052010-09-09Fujitsu LimitedImage encoding device, image encoding control method, and program
US8295353B2 (en)*2009-03-052012-10-23Fujitsu LimitedImage encoding device, image encoding control method, and program

Non-Patent Citations (25)

* Cited by examiner, † Cited by third party
Title
Bachar et al., Image quality and localization accuracy in C-arm tomosynthesis-guided head and neck surgery. Med Phys. 2007;34(12):4664-4677.
Bachar et al., Three-dimensional tomosynthesis and cone-beam computed tomography: an experimental study for fast, low-dose intraoperative imaging technology for guidance of sinus and skull base surgery. Laryngoscope. 2009;119(3):434-441.
Balachandran et al., Evaluation of portable CT scanners for otologic image-guided surgery. Int J Comput Assist Radiol Surg. 2012;7(2):315-321.
Chan et al., Cone-beam computed tomography on a mobile C-arm: novel intraoperative imaging technology for guidance of head and neck surgery. J Otolaryngol Head Neck Surg. 2008;37(1):81-90.
Daly et al., Intraoperative cone-beam CT for guidance of head and neck surgery: Assessment of dose and image quality using a C-arm prototype. Med Phys. 2006;33(10):3767-80.
Dawson et al., Advances in image-guided radiation therapy. J. Clin. Oncol. 2007;25(8):938-946.
International Search Report and Written Opinion of PCT/US2012/037797.
Jaffray et al., Cone-beam computed tomography with a flat-panel imager: initial performance characterization. Med Phys. 2000;27(6):1311-1323.
Lauritsch et al., Towards cardiac C-arm computed tomography. IEEE Trans Med Imaging. 2006;25(7):922-934.
Miga et al., Modeling of retraction and resection for intraoperative updating of images. Neurosurgery. 2001;49(1):75-84; discussion 84-5.
Nithiananthan et al., Deformable registration for intra-operative cone-beam CT guidance of head and neck surgery. In: Engineering in Medicine and Biology Society, 2008. EMBS 2008. 30th Annual International Conference of the IEEE.; 2008:3634-3637.
Nithiananthan et al., Demons deformable registration for CBCT-guided procedures in the head and neck: Convergence and accuracy. Med. Phys. 2009;36(10):4755-4764.
Nithiananthan et al., Demons Deformable Registration of CT and Cone-Beam CT Using an Iterative Intensity Matching Approach. Medical Physics. 2011.
Pennec et al., Understanding the "Demon's Algorithm": 3D Non-rigid Registration by Gradient Descent. In: Proceedings of the Second International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer-Verlag; 1999:597-605. Available at: http://portal.acm.org/citation.cfm?id=709761. Accessed Apr. 26,2009.
Periaswamy et al., Medical image registration with partial data. Med Image Anal. 2006;10(3):452-464.
Risholm et al., A non-rigid registration framework that accommodates resection and retraction. Inf Process Med Imaging. 2009;21:447458.
Risholm et al., Validation of a nonrigid registration framework that accommodates tissue resection. In: San Diego, California, USA; 2010:762319-762319-11. Available at:; http://link.aip .org/link/P SISDG/v7623/il/p762319/sl&Agg=doi. Accessed Jan. 24,2011.
Santos et al., The accuracy of intraoperative 0-arm images for the assessment of pedicle screw postion. Spine. 2012;37(2):E119-125.
Schafer et al., Mobile C-arm cone-beam CT for guidance of spine surgery: image quality, radiation dose, and integration with interventional guidance. Med Phys. 2011;38(8):4563-4574.
Sezgin et al., Survey over image thresholding techniques and quantitative performance evaluation. J. Electron. Imaging. 2004;13(1):146-168.
Siewerdsen et al., Cone-beam CT with a flat-panel detector on a mobile C-arm: preclinical investigation in image-guided surgery of the head and neck. In: Medical Imaging 2005: Visualization, Image-Guided Procedures, and Display.vol. 5744. San Diego, CA, USA: SPIE; 2005:789-797. Available at: http ://link.aip org/link/?P SI/5744/789/1 . Accessed Feb. 10, 2009.
Thirion J. Fast non-rigid matching of 3D medical images. France: INRIA; 1995. Available at: http://hal.inrialeinria-00077268/en/. Accessed Feb. 12,2009.
Thirion JP. Image matching as a diffusion process: an analogy with Maxwell's demons. Med Image Anal. 1998;2(3):243-260.
Vercauteren et al., Diffeomorphic demons: Efficient non-parametric image registration. Neurolmage. 2009;45(1, Supplement 1):561-572.
Wang et al., Validation of an accelerated "demons" algorithm for deformable image registration in radiation therapy. Phys Med Biol. 2005;50(12):2887-2905.

Cited By (3)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20190213743A1 (en)*2014-08-222019-07-11Analogic CorporationUpdating Reference Imaging Data with Update 2D and/or 3D Imaging Data
US10650537B2 (en)*2014-08-222020-05-12B-K Medical ApsUpdating reference imaging data with update 2D and/or 3D imaging data
US20250054098A1 (en)*2023-08-082025-02-13City University Of Hong KongPreserving Tumor Volumes for Unsupervised Medical Image Registration

Also Published As

Publication numberPublication date
US20140079338A1 (en)2014-03-20
WO2012155136A3 (en)2013-03-21
WO2012155136A2 (en)2012-11-15

Similar Documents

PublicationPublication DateTitle
US9218643B2 (en)Method and system for registering images
Nithiananthan et al.Extra‐dimensional Demons: a method for incorporating missing tissue in deformable image registration
Maken et al.2D-to-3D: a review for computational 3D image reconstruction from X-ray images
US11657518B2 (en)Method for deformable 3D-2D registration using multiple locally rigid registrations
US8358818B2 (en)Apparatus and methods of compensating for organ deformation, registration of internal structures to images, and applications of same
JP5520378B2 (en) Apparatus and method for aligning two medical images
Lee et al.Breast lesion co-localisation between X-ray and MR images using finite element modelling
De Silva et al.Registration of MRI to intraoperative radiographs for target localization in spinal interventions
CN107624192A (en) Systems and methods for surgical guidance and intraoperative pathology of tissue differentiation via endoscopy
Fang et al.3D shape reconstruction of lumbar vertebra from two X-ray images and a CT model
Gueziri et al.Toward real-time rigid registration of intra-operative ultrasound with preoperative CT images for lumbar spinal fusion surgery
Castro-Mateos et al.3D segmentation of annulus fibrosus and nucleus pulposus from T2-weighted magnetic resonance images
Reaungamornrat et al.Deformable image registration with local rigidity constraints for cone-beam CT-guided spine surgery
Melinska et al.Statistical, morphometric, anatomical shape model (atlas) of calcaneus
Xie et al.Statistical model-based segmentation of the proximal femur in digital antero-posterior (AP) pelvic radiographs
EP3391332B1 (en)Determination of registration accuracy
Zollei2D-3D rigid-body registration of X-ray Fluoroscopy and CT images
US9286688B2 (en)Automatic segmentation of articulated structures
Han et al.A momentum-based diffeomorphic demons framework for deformable MR-CT image registration
Garcia Guevara et al.Biomechanics-based graph matching for augmented CT-CBCT
Zheng et al.HipMatch: an object-oriented cross-platform program for accurate determination of cup orientation using 2D–3D registration of single standard X-ray radiograph and a CT volume
Nithiananthan et al.Incorporating tissue excision in deformable image registration: a modified demons algorithm for cone-beam CT-guided surgery
EP4012650A1 (en)Segmentation of anatomical structure in image
Steger et al.Personalized articulated atlas with a dynamic adaptation strategy for bone segmentation in CT or CT/MR head and neck images
Thompson et al.Use of a CT statistical deformation model for multi-modal pelvic bone segmentation

Legal Events

DateCodeTitleDescription
ASAssignment

Owner name:THE JOHNS HOPKINS UNIVERSITY, DISTRICT OF COLUMBIA

Free format text:ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:NITHIANNANTHAN, SAJENDRA;SIEWERDSEN, JEFFREY H.;MIROTA, DANIEL J.;SIGNING DATES FROM 20120814 TO 20120822;REEL/FRAME:028955/0023

ASAssignment

Owner name:THE JOHNS HOPKINS UNIVERSITY, MARYLAND

Free format text:ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SIEWERDSEN, JEFFREY H.;NITHIANNANTHAN, SAJENDRA;MIROTA, DANIEL J;SIGNING DATES FROM 20120814 TO 20140114;REEL/FRAME:032039/0296

FEPPFee payment procedure

Free format text:PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

STCFInformation on status: patent grant

Free format text:PATENTED CASE

ASAssignment

Owner name:NATIONAL INSTITUTES OF HEALTH (NIH), U.S. DEPT. OF

Free format text:CONFIRMATORY LICENSE;ASSIGNOR:JOHNS HOPKINS UNIVERSITY;REEL/FRAME:039205/0965

Effective date:20160504

MAFPMaintenance fee payment

Free format text:PAYMENT OF MAINTENANCE FEE, 4TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2551); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

Year of fee payment:4

MAFPMaintenance fee payment

Free format text:PAYMENT OF MAINTENANCE FEE, 8TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2552); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

Year of fee payment:8


[8]ページ先頭

©2009-2025 Movatter.jp