DYNAMIC TRACKING OF MOVING TARGETS
Treatment targets such as tumors or lesions, located within an anatomical region that undergoes motion (which may be periodic with cycle P), are dynamically tracked. A 4D mathematical model is established for the non-rigid motion and deformation of the anatomical region, from a set of CT or other 3D images. The 4D mathematical model relates the 3D locations of part(s) of the anatomical region with the targets being tracked, as a function of the position in time within P. Using fiducial-less non-rigid image registration between pre-operative DRRs and intra-operative x-ray images, the absolute position of the target and/or other part(s) of the anatomical region is determined. The cycle P is determined using motion sensors such as surface markers. The radiation beams are delivered using: 1) the results of non-rigid image registration; 2) the 4D model; and 3) the position in time within P.
Latest Accuray, Inc. Patents:
- FUNDAMENTALS OF SEMI-CONDUCTOR-BASED PHOTON COUNTING DETECTORS
- Multimodal radiation apparatus and methods
- Sparse background measurement and correction for improving imaging
- Multimodal radiation apparatus and methods
- METHODS FOR SATURATION CORRECTION AND DYNAMIC GAIN CONFIGURATION AND APPARATUSES FOR PERFORMING THE SAME
This application is a continuation of application Ser. No. 10/957,467, filed Sep. 30, 2004, which is hereby incorporated by reference.
BACKGROUNDIn some medical applications, it may be necessary to dynamically track targets that move with time. For example, in radiosurgery it may be desirable to dynamically track tumors and/or lesions in the human body that move with respiration and/or heartbeat. In radiosurgery, accurate trajectories of the radiation beams through the patient anatomy to the lesion or tumor being treated can be critical, in order to achieve the radiation dose distribution that was computed during treatment planning time. For regions of the human anatomy that move, for example due to breathing or heartbeat, it is important to take such motions into consideration, when computing the effect of the motion on the treatment plan being generated. Dynamic tracking may also be useful in applications other than radio surgery in which parts of the anatomy move, due to breathing, heartbeat, or any other type of motion.
Fiducial markers have been used in the past, in order to track moving regions of the anatomy. Fiducials-based tracking can be difficult for a patient, for a number of reasons. For example, high accuracy tends to be achieved by using bone-implanted fiducial markers, but less invasive techniques such as skin-attached markers or anatomical positions tend to be less accurate. Implantation of fiducials into a patient is generally painful and difficult, especially for the C-spine, the implantation process for which may frequently lead to clinical complications.
In some methods that use gating to handle anatomical motion, dynamic tracking may be achieved by establishing a relationship between internally implanted fiducials, and externally placed markers that are tracked in real time. These methods do not take into account the non-rigid motions and deformations of the surrounding anatomy, as a function of the motion cycle.
A method and system that address these deficiencies are thus desirable. In particular, it is desirable to provide a reliable and efficient method and system for dynamically tracking moving targets.
SUMMARYA method and system are presented for dynamically tracking one or more targets within an anatomical region so that desired doses of therapeutic radiation can be delivered to the targets, while the anatomical region is undergoing motion. The targets may be tumors or lesions, for example. The anatomical region may include one or more reference structures, in addition to the targets. The locations of the targets within the moving anatomical region are determined relative to the reference structures, while taking into account the deformation that the anatomical region undergoes during its motion. The deformation of the anatomical region includes non-rigid deformations, as well as rigid deformations.
In one embodiment, the motion of the anatomical region may be periodic. Examples of such periodic motion may include, for example, respiration or heartbeat. The periodic motion may be characterized by a periodic cycle P. In other embodiments, the motion of the anatomical region may be aperiodic motion, in which the cycle P may change over time, or for which no cycle P is defined. Throughout this patent application, the terms “periodic motion” and “periodic motion with cycle P” include periodic motions having a time-varying period P(t). In other words, the terms “periodic motion” and “periodic motion with cycle P” should be understood as being descriptive of, and referring to, the changing nature of the motions in the human anatomy.
In one embodiment, dynamic tracking of a periodic motion with cycle P is accomplished by constructing a 4D (four dimensional) mathematical model that describes the non-rigid deformation and non-rigid motion of the anatomical region, as a function of the relative position in time within the cycle P. The 4D mathematical model may be constructed by generating a plurality of images (including but not limited to CT images) Ij (j=1, . . . , p) of the anatomical region, where each image is taken at one of a succession of time points tj (j=1, . . . p) within the cycle P. The continuous non-rigid deformation of the anatomical region, as a function of the cycle P, is mathematically modeled by morphing a CT image, acquired at one instant in time, into another CT image acquired at a subsequent instant in time. The 4D mathematical model relates the 3D (three dimensional) locations of the skeletal structures with the 3D locations of the targets, as a function of the instant in the cycle P (which may be, for example, a breathing cycle or a heartbeat cycle).
In one embodiment, the 4D mathematical model of the non-rigid deformation is used by a treatment plan generator to generate a radiation dose distribution that prescribes a desired amount of the therapeutic radiation to be delivered, in real time, to the targets within the moving anatomical region. The radiation dose distribution accounts for the non-rigid deformation of the moving anatomical region during its periodic motion.
The cycle P for the periodic motion may be determined for example by dynamically tracking motion sensors (such as surface markers or LEDs (light emitting diodes), by way of example) attached to the skin or exterior surface of the anatomical region. The relative position within the motion cycle P of any desired time point, including the time point tj at which the j-th image Ij was taken, can be determined in this way.
In this patent, the term “real time” refers to a time scale that is substantially simultaneous to the actual radiation treatment and delivery. For example, tracking that occurs at a speed of at least several Hz or higher falls within the time scale described as “real time” in this patent. In this patent, the term “near real time” refers to a time scale that is slower than the time scale referred to as “real time,” for example by about one or more orders of magnitude. For example, events occurring at a speed of less than about once a second falls within the time scale described as “near real time” in this patent.
In one embodiment, the near real time locations of the reference structures are determined by performing fiducial-less tracking of the reference structures. A set of DRRs are generated from the CT image data, using the same beam projection geometry as is used to acquire the live x-ray images. The live x-ray images, acquired intra-operatively in near real time, are registered with the DRRs (digitally reconstructed radiographs). In one embodiment, a non-rigid image registration algorithm may be used, in which the non-rigid transformation parameters are derived from a 3D full motion field. The full motion field may be constructed by estimating many local motion vectors. In one embodiment, multi-level block matching may be used in the non-rigid image registration algorithm, in conjunction with a similarity measure based on pattern intensity. In one embodiment, hierarchical mesh motion estimation may be performed.
In one embodiment, the locations of the targets can be determined in real time, by correlating the known near real time locations of the reference structures, which were determined using non-rigid image registration, with the target locations that are being sought, using the 4D mathematical model that describes how the locations of the targets are related to the locations of the reference structures, as a function of the relative position in time within the cycle P. The information from the motion sensors can be used to determine the relative position within the cycle P of any desired time point.
In one embodiment, radiation treatment is delivered to the real time locations of the targets, in accordance with the radiation dose distribution generated by the treatment planning generator. The radiation beam delivery can be synchronized to the instant in breathing cycle determined in the treatment planning.
A number of techniques are described for dynamically tracking tumors/lesions in the human body that is undergoing motion. The methods are of principal use in radiosurgery, but may also be useful in other applications where it may be necessary to dynamically track parts of the anatomy that move, for example because of respiration or heartbeat.
Some portions of the detailed description that follows are presented in terms of algorithms and symbolic representations of operations on data bits within a computer memory. These algorithmic descriptions and representations are the means used by those skilled in the data processing arts to most effectively convey the substance of their work to others skilled in the art. An algorithm is conceived to be a self-consistent sequence of acts leading to a desired result. These acts require physical manipulations of physical quantities. Usually, though not necessarily, these quantities may take the form of electrical or magnetic signals capable of being stored, transferred, combined, compared, and otherwise manipulated. It has proven convenient at times, principally for reasons of common usage, to refer to these signals as bits, values, elements, symbols, characters, terms, numbers, or the like.
It should be borne in mind, however, that all of these and similar terms are to be associated with the appropriate physical quantities and are merely convenient labels applied to these quantities. Unless specifically stated otherwise, it is appreciated that throughout the descriptions below, discussions that utilize terms such as “processing” or “computing” or “calculating” or “determining” or “displaying” or the like, refer to the actions and processes of a computer system, or a similar electronic computing device. The computer system manipulates and transforms data, represented as physical or electronic quantities within the computer system's registers and memories, into other data similarly represented as physical quantities within the computer system's memories or registers, or within other such information storage, transmission or display devices.
The methods and techniques described below can be implemented by an apparatus for performing the operations discussed below. Such an apparatus may be specially constructed for the required purposes, or it may comprise a general purpose computer, selectively activated or reconfigured by a computer program stored in the computer. Such a computer program may be stored in a computer readable storage medium, such as, but not limited to, any type of disk including floppy disks, optical disks, CD-ROMs, and magnetic-optical disks, read-only memories (ROMs), random access memories (RAMs), EPROMs, EEPROMs, magnetic or optical cards, or any type of media suitable for storing electronic instructions. Each such computer readable storage medium may be coupled to a computer system bus.
The algorithms and displays presented herein are not inherently related to any particular computer or other apparatus. Various general purpose systems may be used with programs that are designed in accordance with the teachings below, or it may prove convenient to construct a more specialized apparatus to perform the requisite methods and techniques. For example, any of the methods described below can be implemented in hard-wired circuitry, or by programming a general purpose processor, or by any combination of hardware and software. One of skill in the art will appreciate that the methods and techniques described below can be practiced with a wide variety of computer system configurations, including hand-held devices, multiprocessor systems, microprocessor-based or programmable consumer electronics, network PCs, minicomputers, mainframe computers, and the like. The methods and techniques described below can also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network.
The methods and systems described below may be implemented using computer software. Computer software may be referred to using different terms, for example a program, a procedure, or an application. If written in a programming language conforming to a recognized standard, sequences of instructions designed to implement these methods and systems can be compiled for execution on a variety of hardware platforms and for interface to a variety of operating systems. Also, these methods and systems are not described with reference to any particular programming language. It will be appreciated that a variety of programming languages may be used to implement the teachings of the invention as described herein. Furthermore, it is common in the art to speak of software, in one form or another, as taking an action or causing a result. Such expressions are merely a shorthand way of saying that the execution of the software by a computer causes one or more processors in the computer to perform an action or produce a result.
As mentioned earlier, throughout this description the term “periodic motion with cycle P” should be understood as including periodic motions in which the cycle P of the periodic motion, as well as the amplitude and waveform of the motion, change with time, In other words, the term “periodic motion” or “periodic motion with cycle P” should be understood as also referring to the changing nature of the motions in the human anatomy.
As mentioned earlier, the anatomical region may undergo a deformation (which may be a non-rigid deformation) during its motion. While in the embodiment illustrated in
In overview, in the embodiment illustrated in
As a first step, a set of CT images Ij (j=1, . . . p) are acquired in step 105, each CT image taken at one of a succession of time points tj (j=1, . . . p) within the cycle P. In step 110, the cycle of the periodic motion (e.g., the respiratory cycle or the heartbeat cycle) is established, for example by dynamic tracking of markers or other sensors attached to the skin of the anatomical region.
In step 120, a 4D (3D+time) mathematical model is constructed from these CT scans and from the motion cycle information obtained from the sensors. The mathematical model describes a non-rigid motion and deformation of the anatomical region as it undergoes its periodic motion (e.g. respiration), as a function of the instant in the motion cycle P. The 4D mathematical model relates the locations of the targets to the locations of the reference structures, as a function of the relative position in time within the cycle P, within the periodically moving anatomical region. More generally, this 4D model may describe the 3D+time relationship between a part or parts of the anatomy, and a target to be tracked for radiosurgery, as a function of the instant in time in the breathing cycle.
Next, in steps 130, 135, and 140 the absolute position of a part or parts of the anatomy is determined, using x-ray imaging and 2D-to-3D image registration. In these steps, the locations of the reference structures and/or the targets are determined in near real time, by fiducial-less tracking of the reference structures and/or targets using a non-rigid image registration algorithm.
As explained earlier, in this patent the term “real time” refers to a time scale that is substantially simultaneous to the actual radiation treatment and delivery. In just one exemplary embodiment, intra-operative real-time tracking of a target may be implemented using optical markers which track at about a 30 Hz rate, and using a model which updates its predictions at about a 80 Hz rate. This is intended to be an illustrative example only, and real time tracking may occur at a wide range of different tracking speeds, generally higher than about 1 Hz. The term “near real time” refers to a time scale that is slower, for example by about one or more orders of magnitude, than the time scale described by the term “real time.” As an example, the time scale for acquiring x-ray images, which may range from about a fraction of a second to about several seconds, will be described as “near real time.”
A set of DRRs of the anatomical region are generated, in step 130, from the CT images acquired in step 105. In step 135, live x-ray projection images of the anatomical region are acquired. In step 140, the DRRs of the reference structures are registered with the near real time x-ray images of the reference structures.
Step 150 describes 4D treatment planning, in which a radiation dose distribution is computed that results from continuous beam delivery through the non-rigidly moving and deforming anatomical region. In this step, the radiation beam trajectories are determined, using the knowledge (obtained in step 140) of the absolute locations of the reference structures, and using the 4D model that relates the reference structures to the instant in breathing cycle (as determined using information from the sensors), and to the targets whose locations are being tracked.
In step 160, the target locations are determined. In one embodiment, the 4D mathematical model may be used, together with the absolute positions of the skeletal structures as determined by fiducial-less tracking, and the information obtained from the sensors regarding the motion cycle, to determine the target locations. In an alternative embodiment, the target locations may be determined by 2D/3D non-rigid image registration, during which the DRRs (generated during the treatment planning stage) are registered onto near real time x-ray images. Finally, in step 170, radiation is delivered to the targets in accordance with the radiation dose distribution generated through 4D treatment planning.
In the illustrated embodiment, the CT images are taken at time points tk, tl, and tp, respectively. The epochs or time points within the breathing cycle P are preferably chosen to substantially encompass the overall dynamic range of the periodic motion. For example, in one embodiment, the time points may include: a time point t1 corresponding to a trough of the cycle P; a time point tp corresponding to a peak of the cycle P; and a third time point tk disposed at an intermediate location between the peak and the trough of the cycle P. In other embodiments, the time points selected for taking the CT images may include more than the three time points t1, tp, and tk described above.
From this set of CT studies, a 4D mathematical model is constructed that morphs the CT image acquired at one instant or time point in the motion cycle into another CT image acquired at a subsequent instant or time point in the motion cycle, providing a model for the continuous non-rigid deformation of the anatomy as a function of the motion cycle. In image processing, it is well known in the art to morph one image into another image, and to describe this in terms of a mathematical model. Any standard software and/or algorithms that are known and may be commercially available can be used.
In one embodiment, the 4D mathematical model constructed from the set of CT images shown in
In order to compute the desired radiation dose distributions, the beam trajectories are initially defined with respect to a nominal patient coordinate system. In one embodiment, the nominal patient coordinate system may be chosen to orient with respect to one of the several CT images illustrated in
In one embodiment, the 4D mathematical model relates the 3D locations of one or more reference structures with the 3D locations of the target, as a function of the instant in the motion cycle. In the 4D model, one or more of the selected reference structures may be stationary with respect to the motion cycle, while others of the selected reference structures may undergo non-rigid motion with respect to the motion cycle.
In the embodiment illustrated in
The 3D location and orientation of the multiple skeletal structures enable vectors to be drawn that point from each skeletal structure to the tumor or lesion. A model that describes the 3D motion of the tumor/lesion as a function of the breathing cycle, with reference to a rigid skeletal structure (such as the skeletal structure indicated in
Once the relationship between the locations of the reference structures and the locations of the target is determined by the 4D model, as a function of the point in time within the periodic motion cycle, the absolute position of the reference structures is determined in near real time. In one embodiment, this is accomplished by 1) generating DRRs of the reference structures from the 3D CT images, which were shown in
As a first step in the 2D/3D non-rigid image registration process, a library of DRRs is generated for the projection geometry that will be used in the acquisition of live images at the time of treatment delivery. A pair of live (or “near real time”) x-ray images of the patient is acquired during treatment delivery, using the same projection geometry as used for generating the DRRs. The CT images used to generate the DRRs will correspond to one of the epochs in the motion cycle, typically the same one with respect to which beams are assigned to during treatment planning. The DRRs may be generated for several translations and orientations of the CT image, in order to cover the typical range of patient movements during treatment.
The acquired images are registered with the DRRs using a feature recognition algorithm that tracks the reference structures. The image registration algorithm (described in detail below) may be repeated for multiple reference structures, to give the position and orientation of each structure with respect to the corresponding structures in the CT study (from which DRRs are made). In one embodiment, the difference in imaging characteristics of the tumor or lesion or a nearby anatomical region may be enhanced further by using high-sensitivity x-ray detectors.
In one embodiment, the location of the tumor/lesion can be derived from the locations of the reference structures, determined by the non-rigid registration process. The location and orientation of multiple reference structures tracked using the fiducial-less algorithm are used to interpolate the location of the tumor/lesion, using their corresponding geometric relationships, as learned from the CT studies and the 4D model. In another embodiment, fiducial-less tracking is performed for the targets (e.g. tumors/lesions) themselves. If the target being tracked is sufficiently different in x-ray imaging characteristics relative to the surrounding tissue, the target itself can be directly tracked using the 2D-to-3D registration technique described below.
A non-rigid 2D-to-3D image registration technique that accounts for non-rigid deformations of the anatomy, and which uses anatomical reference structures instead of fiducials, is described below. While the non-rigid image registration algorithm described below is described in the context of skeletal structures, and in particular skeletal structures in the spinal region of the human anatomy, it should be understood that reference structures other than spinal skeletal structures may also be used with this non-rigid image registration algorithm.
The non-rigid image registration technique is also described in the following five co-owned patent applications: 1) U.S. patent application Ser. No. 10/880,486, characterized by attorney docket number ACCL-136, entitled “Fiducial-less Tracking With Non-Rigid Image Registration”; 2) U.S. patent application Ser. No. 10/881,612, characterized by attorney docket number ACCL-137, entitled “Motion Field Generation For Non-rigid Image Registration”; 3) U.S. patent application Ser. No. 10/881,209, characterized by attorney docket number ACCL-146, entitled “ROI Selection In Image Registration”; 4) U.S. patent application Ser. No. 10/881,208, characterized by attorney docket number ACCL-147, entitled “Image Enhancement Method and System For Fiducial-less Tracking of Treatment Targets”; 5) U.S. patent application Ser. No. 10/881,206, characterized by attorney docket number ACCL-150, entitled “DRR Generation Using A Non-Linear Attenuation Model.” All five patent applications, owned are incorporated by reference herein in their entireties.
As a first step, 2D DRRs may be generated from pre-operative 3D scan data representative of a CT image Ij, in step 402. In one embodiment, the images for which non-rigid image registration is performed (i.e., DRRs and x-ray images) are discretized images each characterized by an array of pixels, each pixel having an associated pixel value representative of the intensity of the image at a surface unit area corresponding to the pixel.
In one embodiment, an improved DRR generation process can be implemented in step 402 to bring out the skeletal reference structures, which are usually not easily visible in the images, or even may be hidden. In step 402, the CT scan data are modified based on a non-linear attenuation model that emphasizes the skeletal structures and thus improves the quality of the DRRs. In step 403 in flowchart 400, an image enhancement technique may also be implemented for the DRRs. In this step, a top-hat filter is used to bring out the skeletal structures in the DRRs generated in step 402.
In the illustrated embodiment, image registration is performed in a selected region of interest (ROI) within the enhanced DRR, in order to improve efficiency. Accordingly, an ROI is defined in the DRR, in step 404, after enhancement of the DRRs. The ROI selection process performed in step 404 is based on image entropy, and is fully automatic so as not to require user interaction. Intra-operative 2D x-ray projection images are then generated, in near real time, in step 410. Image enhancement is performed on the x-ray images, in step 415, using a top-hat filter by analogy to step 403.
Non-rigid image registration is then performed between the enhanced x-ray images and the enhanced DRRs, within the ROI. In particular, a similarity measure is used to compare the pixel intensities in the x-ray images and the DRR images, in order to determine any change in the position and/or orientation and/or physiological deformation of the patient. In steps 420-450, a non-rigid deformation that describes real patient movement and body deformation is defined. To define the non-rigid deformation, a full motion field is constructed that is composed of many local motion fields, i.e. a plurality of locally estimated motion vectors. To estimate local motion at a given point of interest within the ROI, a similarity measure based on pattern intensity is used to compare pixel intensities.
A full motion field that is composed of many local motions may describe any desired non-rigid deformation. Further, a full motion field derived in this manner can account for non-rigid motions (translations and/or rotations) of the object, in addition to non-rigid deformations, between different image acquisitions of the object. In order to efficiently compute the local motion vectors at any point of interest within the ROI, hierarchical mesh motion estimation and multi-level block matching (performed in conjunction with an intensity-based similarity measure) are performed. These methods allow for a fast computation of the image registration algorithm 400. A smoothness constraint is imposed to reconstruct the motion field at mesh nodes in which mismatching occurred. The non-rigid transformation parameters for the non-rigid image registration are then computed from the full motion field.
In the embodiment illustrated in
In the final steps, shown as step 455 and step 460 in
The quality of DRR images relies on proper attenuation modeling, as well as a proper interpolation scheme for interpolation the CT numbers. In one embodiment, in step 402 (in the flowchart shown in
C(x,y,z)=a C0(x,y,z)ebC
where C(x,y,z) represents the modified CT number of a 3D CT voxel located at a point (x,y,z); a and b represent weighting coefficients;
and C0(x,y,z) represents the un-modified CT number, based on a linear attenuation model, of a 3D CT voxel having a location (x,y,z).
The interpolation scheme used in one embodiment to improve the quality of DRRs is bi-linear interpolation. In this embodiment, bi-linear interpolation is performed in step 402, to integrate the CT numbers along the CT voxels that are encountered by each cast ray. In one embodiment, the bi-linear interpolation is followed by a 1-D polynomial interpolation over three voxel slices, for each voxel of interest. The three voxel slices include the voxel slice containing the voxel of interest, plus each adjacent voxel slice.
In some embodiments, fiducial-less tracking relies on skeletal reference structures (e.g. vertebral structures) that are usually not easily visible, or may even be hidden in the DRRs and in the x-ray projection images. Because such skeletal structures have to be registered, both the DRR and the x-ray images have to be enhanced to bring out the details of the vertebral structures and improve their visibility. In one embodiment, therefore, image enhancement is undertaken for both the DRRs and the x-ray projection images. In most thoracic and lumbar cases, the skeletal structures are not easily visible or even hidden in DRR and X-ray images. For these cases therefore, enhancement of the DRR and the x-ray images is necessary, in order to make registration at all possible. In cervical cases, the skeletal structures of spine are well visible in both the DRR and the x-ray images, but the details of the structures are still not clear. Accordingly, in cervical cases, the DRR and the x-ray images should be enhanced to improve the registration.
In the embodiment illustrated in
In one embodiment, the top-hat filter is designed by using a weighted combination of image opening and closing with a certain structural element. The top hat filter operator is defined mathematically as follows:
where fe represents the enhanced image, resulting from the application of the top hat filter operator to each pixel in the original image;
f represents the original image;
w and b represent weighting coefficients,
γB(f) represents a structural element for the opening of the original image f, and
φB(f) represents a structural element for the closing of the original image f.
In expression (2) above, WTH(f)=f−γB(f) is called a white top-hat filter, whereas BTH(f)=φB(f)−f is called a black top-hat filter. The structural elements γB(f) and φB(f) are masks that are used to perform the basic morphological operation. The sizes of the structural elements vary slightly for cervical, thoracic, and lumbar applications. The empirical values are determined experimentally. The weighting coefficients w and b are determined adaptively by the amplitudes of WTH(f) and BTH(f), respectively. Empirically, the values of the weighting coefficients w and b have been found to be about 1 each (w=1, b=1), for a cervical case in which less tissue is present. In the lumbar case, in which more tissue is present, the values of w and b have been found to be greater than about 2 each (w>2, b>2). In the lumbar case, the weighting process brings out the skeletal structures to a greater degree, compared with the cervical case.
In one embodiment, image registration is conducted only in a certain region of interest (ROI) defined in the DRR. The ROI contains the treatment target (e.g. a tumor or lesion). In one embodiment, image entropy is specifically defined, in step 404 in
The Shannon entropy, known from conventional communication theory, is commonly used as a measure of information in signal and image processing. It is defined as H=−Σinpi log pi where H represents the average information supplied by a set of n symbols whose probabilities are given by p1, p2, . . . , pn. When applied to the pixels of each image (as enhanced in steps 403 or 415 in
In one embodiment, the Shannon entropy H is modified, based on the fact that the skeletal structures occur in bright areas. In this embodiment, a modified Shannon entropy is used for each image, which is defined as follows:
H=−ΣIIp(I)log p(I), (3)
where again I is the image intensity level, and p(I) is the probability of the image intensity value I occurring within the ROI. In step 404 (shown in
Restricting the image registration process to within a ROI has several advantages. One advantage is that such a restriction can speed up the registration process, since the registration needs to be performed only for the ROI. For example, the similarity measure needs only be computed for the ROI, and block matching need only be performed within the ROI. Further, the registration process is more accurate when limited to an area within the ROI. The more limited the region in which registration is conducted, the less likely it is that structures within the ROI would have moved relative to each other between the time of the pre-operative CT scans and the time of the medical treatment.
Based on the improved and enhanced DRRs (generated in steps 402 and 403 in
A non-rigid image registration allows the inherent local anatomical variations that exist between different image acquisitions to be accounted for, in contrast to a rigid image registration which does not allow the overcoming of such variations. Non-rigid registration defines a deformation field that provides a translation or mapping for every pixel in the image. In one embodiment, a full motion field, composed of many local motion vectors or fields, is computed in order to derive the non-rigid deformation field.
In order to estimate local motion fields, in one embodiment, a multi-level block matching method is used in conjunction with a similarity measure based on pattern intensity. This approach allows the local motion to be rapidly and accurately estimated in most parts of the ROI. Multi-level block matching, which allows for computational efficiency, is described in conjunction with a rigid registration algorithm, in a commonly owned application, U.S. Ser. No. 10/652,786 (the “'786 application”), incorporated by reference in its entirety. A similarity measure based on pattern intensity, used in conjunction with a registration algorithm based on rigid transformations, i.e. the “FAST 6D algorithm” developed by Accuray, Inc. for use with the Cyberknife radiosurgery system, is described in full in commonly owned applications, U.S. Ser. Nos. 10/652,786 (the '786 application“), 10/652,717 (the “'717 application”), and 10/652,785 (the “'785 application”), which are all incorporated by reference in their entireties. In the present patent, the pattern intensity based similarity measure and the multi-level block matching method are used in conjunction with a registration algorithm based on a non-rigid (rather than a rigid) transformation. The pattern intensity-based similarity measure, originally developed for a rigid image registration algorithm, provides a powerful and efficient technique for solving the 2D/3D image registration problem, also in a non-rigid framework.
In one embodiment, block matching is performed, i.e. a small block centered around a point of interest is used in order to locally estimate the displacements at each desired point within the ROI. As known, when using block matching to register a first image onto a second image, the first image is divided into different blocks, typically rectangular boxes of equal size. Each point of interest, which may be a mesh node, or may be a non-node pixel that is surrounded by mesh nodes, is taken as the center of one of the blocks. These blocks are then translated so as to maximize a local similarity criterion, which in one embodiment is the pattern intensity based similarity measure, described above.
In block matching methods, it is generally assumed that each pixel in a block has the same motion, and a block matching algorithm is typically used to estimate the motion vectors for each block. In a block matching algorithm used in one embodiment, a search is conducted for a matching block in the second image, in a manner so as to maximize a measure of similarity, based on pattern intensity, between the respective blocks. The search is for a location of the maximum in the similarity measure function, the maximum representing the existence of a matching block in the second image. The search may be conducted within a search window that is defined around the point of interest and that contains the block.
In any block matching algorithm, it is important to optimize the search strategy, and to select an appropriate block size. For small blocks, the translational rigid model is typically assumed. Even though rigid rotations or some other complicated deformations exist, the rigid body translation model is valid for estimating the translations for the block center point. When rotations or other deformations exist in addition to the translations, the accuracy increases with decreasing block size, and decreases with increasing block size. With the use of smaller block sizes, however, the possibility of mismatching increases. In one embodiment, a block size selection strategy is adopted in which it is assumed that larger blocks are needed for larger displacements, and that smaller blocks are need for smaller displacements.
In
in Level 2 and Level 3, respectively, as indicated in the figure. In the lowest resolution level (Level 3), a large search range is used to enable estimation of large displacements. A very small search range (−2, +2) is used in the rest of the resolution levels.
The results at the lower resolution level serve to determine rough estimates of the displacements. The output at the lower level is then passed onto the subsequent higher resolution level. The estimated motion vector (in most cases, a translation vector) for the block is successively refined, using the higher resolution images. In the final matching results, the accuracy of the estimated translations depends on the spatial resolution of the highest resolution images (shown as level 1 in
There is some risk in multi-level matching. It is possible that the estimate at lower levels may fall in a local maximum, and far away from the global maximum that is being sought. In this case, further matchings at subsequent higher resolution levels may not converge to its global maximum. To overcome this risk, multiple candidates are used for the estimates, in one embodiment. Many candidates that have shown optimal matching results are passed on from the lower levels to the higher resolution levels. The more candidates that are used, the more reliable are the estimates. In one embodiment, the best candidates are ranked by the similarity measure function values.
In one embodiment, a similarity measure based on pattern intensity is used, in conjunction with multi-level block matching. As mentioned earlier, this similarity measure is a key element contributing to the success of the “FAST 6D algorithm,” described in the commonly owned '786 application, '717 application, and '785 application. In one embodiment, the similarity measure is determined by forming a difference image between the “live” (or near real time) x-ray projection images and the DRR images, and applying upon each pixel of the difference image a pattern intensity function. Specifically, the difference image Idiff(i,j) is formed by subtracting a corresponding pixel value of the pre-operative DRR image from each pixel value of the intra-operative x-ray projection image, within the ROI:
Idiff(i,j)=ILive(i,j)−IDRR(i,j) (4)
In equation (4), I(i,j) represents the image intensity value of a pixel located at the i-th row and j-th column of each pixel array for the respective image. Specifically, Idiff(i,j) represents an array of pixel values for a difference image formed by subtracting the corresponding pixel values of the second image from each pixel value of the first image. Ilive(i,j) represents the (i,j)-th pixel value of the first image of the object. IDRR(i,j) represents the (i,j)-th pixel value of the second image of the object. The similarity measure operates on this difference image, and is expressed as the summation of asymptotic functions of the gradients of the difference image over the pixels within a neighborhood R:
In equation (5) above, the constant σ is a weighting coefficient for the pattern intensity function. The sensitivity of the solution to the variation of x-ray image can be minimized by careful selection of this constant. The larger the weighting coefficient, the more stable the results. However, the choice of σ entails a tradeoff between stability and accuracy. When the value of σ is too large, some small details in the images cannot be reflected in the similarity measure. Based on the experiments, the empirical value for σ is in the range from about 4 to about 16, in one embodiment.
Equations (5) and (6) for pattern intensity have several advantages. First, the difference image filters out the low frequency part that predominantly consists of the soft tissues, and keeps the high frequency part that predominantly consists of the skeletal structures. This feature makes the algorithm robust to some brightness intensity difference between live and DRR images. Second, because of the asymptotic function, the measure is less affected by the pixels whose intensity value slightly deviates from its neighboring pixels. These types of pixels are thought to contain random noise. Third, because the asymptotic function quickly approaches to zero when the variable increases, large intensity differences such as image artifacts have the same effects on the similarity measure regardless of their magnitude. Due to this feature, the pattern intensity is less sensitive to image artifacts.
The estimation of local motion fields using block matching together with hierarchical mesh motion estimation, as well as the reconstruction of the full motion field from the plurality of locally estimated motion fields, are performed in steps 420-450 of the flowchart shown in
In one embodiment, a global translation of the entire image (measured as a translation of the image center of the image) is first estimated, then used as the initial estimates for all further local motion estimation. In other words, a rough estimate is made of the center displacement for the entire image, and is used as the starting estimate for all local displacements. Referring back to
After global motion estimation, the next step 430 (see
In particular, mesh motion estimation consists of finding the vertices of corresponding mesh elements in the other image, i.e. finding the corresponding mesh nodes in the other image, such that errors are minimized in the overall motion field. Typically, a number of mesh nodes are selected in one image, and the corresponding mesh nodes in the other image are estimated. For any pixel located within a mesh element (as opposed to being located on the vertices of the mesh elements), the mapping between different image acquisitions is performed through interpolation. The local motion vectors for such pixels are estimated by interpolating from the nodal motion vectors that were estimated for the mesh nodes that surround the pixel.
In one embodiment, hierarchical mesh motion estimation may be performed. By hierarchical mesh motion estimation, it is meant that nodal motion is estimated for the mesh nodes that define the mesh structure, for each of a plurality of mesh resolution levels. Motion estimation performed with a coarse mesh provides the initialization for the subsequent (finer) resolution levels of the mesh. To estimate the motion of each mesh node, multi-level block matching may be performed.
For most mesh nodes, the estimates of motion are reliable and accurate. For a few nodes where mismatching may occur and the estimation may not be reliable, the displacements need to be reconstructed by the surrounding node displacements. Accordingly, the next step in the registration algorithm flow chart in
Local motion estimation relies on the local image content. In some smooth local regions, mismatching may occur. During mesh motion estimation, the estimation in most nodes is pretty accurate. For a few nodes where mismatching occurs, the motions should be reconstructed from their surrounding nodes. What is known a priori is matter coherence of bone and tissue, and accordingly, the smoothness of local motion. In other words, the estimated local motion vectors are thought to be smooth and continuous, because of matter coherence. By imposing this physically-based smoothness constraint, a cost function is formulated to reconstruct the motion field.
In one embodiment, the cost function is expressed mathematically as follows:
E(d)=∫∫β(d−u)2dxdy+λ∫∫(d,x2+d,y2)dxdy (7)
In equation (7) above, E(d) represents the cost function, d represents a desired local estimate for a nodal motion vector at coordinates (x,y), u represents a locally estimated nodal motion vector at coordinates (x,y), and β represents a reliability constant that ranges from 0 to 1, where β=0 indicates an unreliable estimation, and 13=1 indicates a reliable estimation.
By performing a finite difference of the derivatives over the mesh grids, a discretized form for the cost function in equation (7) is expressed as:
E(di,j)=ΣΣβi,j(di,j−ui,j)2+λΣΣ[(di,j−di−1,j)2+(di,j−di,j−1)2] (8)
where ui,j represents the locally estimated translations, di,j is the local motion desired, βi,j=1 if the estimation is reliable and βi,j=0 if the estimation is unreliable.
The first term on the right side of equation (8) reflects the fidelity to the observed data in the reconstruction. The second term imposes the smoothness constraints on the motion field in two spatial directions.
The minimization of the cost function given by equation (8) results in a system of simultaneous linear equations
In one embodiment, the iterative algorithm of successive-over relaxation (SOR), which is fast and convergent, is used to solve the equations:
di,j(n+1)=di,j(n)−ω[(βi,j+4λ)di−1,j(n)−λ(di−1,j(n)+di−1,j(n)+i,j−1(n)+di,j+1(n))−βi,jui,j]/(βi,j+4λ) (10)
Once all the nodal motion vectors have been estimated at all the mesh nodes, the translations for any point (or pixel) inside the ROI can be computed by interpolation.
Assuming the motion vector (dx(i),dy(i)) for nine nodes, the motion vector (dx,dy) at the point of interest is computed using the following expressions:
where N(i) is the shape function for the node (i), and where N(i) for I=1, 2, . . . 9 are given as follows:
N(1)=(1−ξ)(1−η)/4−(N8+N5)/2,
N(2)=(1−ξ)(1−η)/4−(N5+N6)/2,
N(3)=(1+ξ)(1+η)/4−(N6+N7)/2,
N(4)=(1−ξ)(1+η)/4−(N7+N8)/2,
N(5)=(1−ξ2)(1−η)/2,
N(6)=(1−ξ)(1−η2)/2,
N(7)=(1−ξ2)(1+η)/2,
N(8)=(1−ξ)(1−η2)/2,
N(9)=(1−ξ2)(1−η2)/2 (12)
Using steps 420, 430, and 440, described above, the local motion vectors can be estimated for a plurality of points of interest within the ROI. The full motion field is obtained as a composite or superposition of all of the local motion vectors that are estimated for the many points of interest that have been selected for motion estimation.
The final step in the image registration process is target localization, namely deriving the target translations and rotations from the full motion field that has been determined. In one embodiment, non-rigid image registration seeks to determine a projection mapping or transformation between different coordinate systems in respective image acquisitions such that points in each space which correspond to the same anatomical point are mapped to each other. In one embodiment, the transformation is represented by a set of non-rigid transformation parameters (dxT, dyT, dzT, r, p, w), where (dxT, dyT, dzT) represent the translations of the target, and (r, p, w) represent rotations of the target.
In one embodiment, two orthogonal x-ray projections are used to solve for these six parameters. In this embodiment, the registration in each projection is performed individually, and the results of the registration for each projection are subsequently combined, to obtain the six 3D transformation parameters. In other embodiments, however, different projections or combinations thereof may be used to solve for the transformation parameters.
In
For projection A, the 2D motion field (dxA, dyA) is estimated by registering the x-ray image that is projected onto the image plane A, with the corresponding reference DRR image. For projection B, the 2D motion field (dxB, dyB) is estimated by registering the x-ray image that is projected onto the image plane B, with the corresponding reference DRR image. Given the 2D motion fields (dxA, dyA) for projection A, and (dxB, dyB) for projection B, the 3-D target translation (dxT,dyT,dzT), as well as the global rigid rotations (r, p, w), can be obtained for both projections A and B, by a straightforward mathematical operation.
Referring back to
dxT=(dxTA+dxTB)/2,
dyT=(dyTA−dyTB)/√{square root over (2)},
dzT=(dyTA+dyTB)/√{square root over (2)} (13)
The global rigid rotations (r, p, w) can be calculated from the motion fields (dxA, dyA) in projection A and (dxB, dyB) in projection B. Using the target as the rotation center, global rigid rotations are useful for position and rotation correction and compensation during initial patient alignment and treatment. Because the target translation is already calculated, the calculation of the global translation is not needed. To get the three rotations in 3D patient coordinates, three 2D in-plane rotations are first computed, including the in-plane rotations θA and θB in projections A and B, respectively, and the in-plane rotation θx in a plane perpendicular to the inferior-superior axis. Approximately, the global rotations can be expressed as:
r=θx,
p=(θB−θA)/√{square root over (2)},
w=(θB+θA)/√{square root over (2)}, (14)
Estimation of θA and θB is directly based the 2D motion fields in projections A and B, respectively. To estimate θx, a plane is first defined, which passes the target point and is perpendicular to the axis x in the 3D patient coordinate system. Then the motion field is calculated from the two motion fields (xA, yA) and (xB, yB) in projections A and B, respectively.
Assuming (dx, dy) is the motion field in the corresponding coordinate (x, y) and θ is the global rotation, when the rotation is small (<10°), the following transformation equation is valid:
Given (dx,dy) and (x,y) in many points, θ can be easily calculated using least square minimization method
Using equations (14) and (16) above, the average rigid transformation parameters can be obtained, in step 160 illustrated in
Using the results of non-rigid registration, obtained as described above, dynamic tracking of the targets can be performed during treatment delivery, by using the 4D mathematical model, and by monitoring the breathing (or other motion) cycle during delivery of radiation.
Dynamic tracking of the target 210 (i.e. the tumor/lesion 210) during treatment delivery can achieved by combining the 4D mathematical model obtained during 4D treatment planning, with the registration information provided by fiducial-less tracking, and monitoring of the patient breathing motion using surface markers. As explained above, the 4D mathematical model obtained during treatment planning relate the locations of the skeletal structures 230 and 232 to the locations of the target 210.
A number of approaches to dynamic tracking are possible, when using fiducial-less tracking in association with 4D planning and dynamic tracking of surface markers, as schematically shown in
In a second approach, the tumor/lesion 210 is located, using the location of the reference structures 232 and 230 obtained from fiducial-less tracking, and using a model that relates the location of the reference structures 232 and 230 to the lesion 210, obtained from 4D treatment planning. Next, a mathematical model is built that relates the motion of the surface markers 340 to the motion of the lesion 210. In image processing, it is known in the art to relate the motion of a first type of object with the motion of a second type of object, and to describe such a relation in terms of a mathematical model. Algorithms or software that are known and that may be commercially available can be used to built the mathematical model that relates the motion of the surface markers 340 to the motion of the lesion 210.
In a third approach, the locations of the reference structures that are determined from the 4D model obtained during 4D treatment planning are used to locate the lesion by interpolation. A mathematical model is then built that relates the motion of the surface markers 340 to the motion of the lesion 210. This third approach involves the least dependence on the 4-D planning model obtained from treatment planning.
In a final approach, the tumor or lesion 210 is directly tracked, using the 2D/3D registration techniques described in the above paragraphs. In this approach, the model relating the motion of surface markers to the location of the target 210 can be built directly, using just the results of the 2D/3D registration.
Once the targets have been located, using one of the approaches described above, radiation beam delivery can be implemented. The real time locations of the targets within the moving anatomical region, which are determined as described in the previous paragraphs, provide guidance for beam delivery. During treatment planning, the beam trajectories are initially defined with respect to a nominal patient co-ordinate system, perhaps chosen to orient with respect to one of the several CT studies acquired to cover the motion cycle. This epoch in the motion cycle is determined by analyzing the motion of the surface markers, and each radiation beam is to be turned on from this epoch.
In overview, the apparatus 500 includes: a target locater 510 that determines in real time the locations of the target(s) 210 relative to the reference structures 230 or 232 within the periodically moving anatomical region; a 4D treatment planning generator 520 that generates a 4D treatment plan as a function of the relative position in time within P for the moving anatomical region; and a treatment beam generator 524 that delivers therapeutic radiation to the targets in real time in accordance with the treatment plan. The treatment plan prescribes a desired radiation dose distribution to be delivered in real time to the targets, while accounting for a deformation of the moving anatomical region during the periodic motion. The treatment planning generator 520 may be connected to a treatment delivery controller (not shown) which controls the delivery of radiation, in accordance with the treatment plan generated by the treatment planning generator 520.
Also included may be a robot system (not shown), which typically has a fixed base and an articulated arm assembly at the distal end of which the treatment beam generator (e.g. x-ray source such as a linac) may be mounted. The robot system may move (and orient), in response to the directions of the delivery controller, the treatment beam generator (i.e. x-ray linac). The robot system and treatment beam generator are described in detail, for example in commonly owned U.S. Pat. No. 5,207,223, and U.S. patent application Ser. No. 10/814,451 characterized by attorney docket No. ACCL-120, both incorporated by reference herein in their entireties.
In one embodiment, the target locater 510 includes a 3D scanner 520; a DRR generator 530; an x-ray imaging system 540; an image registration subsystem 550; one or more motion sensors 560; a 4D model generator 570; and a target location computer 580. The 3D scanner 520 generates a plurality of 3D images Ij (j=1, . . . , p) of the anatomical region, at each of a succession of time points tj (j=1, . . . , p) within the cycle P. These 3D images may include, but are not limited to: a 3D CT scan; a 3D MRI scan; a 3D PET (positron emission tomography) scan; and a 3D ultrasound scan. The 3D scanner 520 can therefore be one of: a 3D CT scanner; a 3D PET scanner; a 3D MRI scanner; and a 3D ultrasound scanner.
The time points tj (j=1, . . . , p) are preferably chosen to substantially encompass a full range of the periodic motion of the anatomical region. For example, the time points may include: a first time point corresponding to a peak of the cycle P, a second time point corresponding to a trough of the cycle P, and a third time point disposed at an intermediate location between the peak and the trough of the cycle P. The motion sensors 560, which may be surface markers, for example, monitor the periodic motion of the anatomical region, and measure the cycle P. In this way, the motion sensors 560 generate time position data representative of the relative position within P of one or more desired time points.
The DRR generator 530 reconstructs DRRs from the 3D scan data, at each time point tj, by casting hypothetical rays through the volumetric 3D scan data from a known beam projection geometry, and integrating the 3D scan data along each ray. The x-ray imaging system 540 generates near real time 2D x-ray projection images of the targets 210 and the reference structures 230 and 232 within the moving anatomical region, by detecting x-ray imaging beams after the beams have traversed at least a portion of the anatomical region. These x-ray imaging beams are generated from the same beam projection geometry as used to generate the DRRs.
The image registration subsystem 550 registers the near real time x-ray projection images of the reference structures and/or the targets, with the DRRs of the reference structures and/or the targets, thereby determining the locations of the reference structures and/or the targets. In one embodiment, the image registration subsystem 550 includes: 1) an RO1 selector 620 configured to select an RO1 (region of interest) within the DRR, the ROI containing the treatment target and preferably at least one reference structure; 2) an image enhancer 630 configured to enhance the DRRs and the x-ray images by applying a filter operator to the DRR and to the x-ray image;
3) a similarity measure calculator 640 configured to determine a measure of similarity between the DRR and the x-ray image; 4) a motion field generator 650 configured to generate a 3D full motion field by estimating, for each of a plurality of resolution levels, one or more 2D local motion fields within the RO1, using the similarity measure; and
5) a parameter determiner 660 configured to determine a set of non-rigid transformation parameters that represent the difference in the position and orientation of the treatment target as shown in the x-ray image, as compared to the position and orientation of the treatment target as shown in the DRR, from the 3D full motion field.
The 4D model generator 570 generates a 4D model that describes a motion of the targets 210 relative to the reference structures 232 within the moving anatomical region, as a function of the relative position in time within the cycle P. The target location computer 580 computes the locations of the targets at the one or more desired time points. The target location computer 580 uses the 4D model constructed by the 4D model generator 570, to correlate the locations of the targets with the known locations of the reference structures, as determined by the image registration subsystem 550, and uses the time position data obtained by the motion sensors 560 to determine the relative position within P of each desired time point.
In one embodiment, the 4D model generator 570 includes a deformation model constructor 575 configured to construct a mathematical model that describes the deformation and motion of the anatomical region, as a function of the relative position in time within the cycle P. In this embodiment, the 4D model generator 570 derives the 4D model from the mathematical model constructed by the deformation model constructor 575. The deformation model constructor 575 extracts, from the plurality of images Ij generated by the 3D scanner 520, deformation data that contain information relating to the deformation and motion of the anatomical region.
In one embodiment, the deformation model constructor 575 extracts the deformation data from the plurality of images by registering each image that is taken at a time point tj within P, onto a consecutive image that is taken at a consecutive time point tj+1 within P. The information contained in the deformation data comprises information relating to the change in the position and orientation of the targets relative to the reference structures. The deformation model constructor 575 uses the deformation data, together with the time position data from the motion sensors 560 to mathematically correlate, for each time point tj, the relative position within P of the time point tj with the deformation and motion of the targets at the time point tj.
Using the 4D mathematical model generated by the 4D model generator 570, the results of non-rigid image registration as performed by the image registration subsystem 550, and the relative position in time within the breathing cycle as determined by the motion sensors 560, the target location computer 580 computes the locations of the tumors/lesions.
Using the 4D mathematical model generated by the 4D model generator 570, the 4D treatment plan generator 520 generates a desired radiation dose distribution that results from continuous radiation beam delivery through the non-rigidly moving anatomical region. Finally, the treatment beam generator 524 generates and treatment beams in accordance with the desired radiation dose distribution, and delivers them in real time to the targets.
In sum, a number of techniques have been described for dynamically tracking tumors/lesions in the anatomy that move, for example due to periodic motion such as respiration. These techniques combine 4D treatment planning, fiducial-less tracking of skeletal structures or targets, and dynamic tracking of surface markers with pertinent mathematical models, to achieve dynamic tracking of the tumors/lesions of interest.
While the invention has been particularly shown and described with reference to specific embodiments, it should be understood by those skilled in the art that many modifications and variations in form and detail may be made in the techniques and structures described and illustrated herein, without departing from the spirit and scope of the invention. Accordingly, the techniques and structures described and illustrated herein should be understood to be illustrative only and not limiting upon the scope of the present invention. The scope of the present invention is defined by the claims, which includes known equivalents and unforeseeable equivalents at the time of filing of this application.
Claims
1. A method, comprising:
- developing a four-dimensional model to describe a movement of a target region over time; and
- developing a radiation treatment plan based on the four-dimensional model, wherein developing the radiation treatment plan comprises determining a relative movement between a radiation beam path and the target region.
2. The method of claim 1, wherein developing the radiation treatment plan further comprises determining a gating of a radiation beam, based on the relative movement, to intersect the target region with the radiation beam path.
3. The method of claim 2, wherein determining the gating of the radiation beam comprises determining a duration to turn on the radiation beam when intersecting with the target region to give a desired dose distribution of the radiation treatment plan.
Type: Application
Filed: Dec 9, 2010
Publication Date: Apr 21, 2011
Applicant: Accuray, Inc. (Sunnyvale, CA)
Inventors: Euan Thomson (Los Gatos, CA), John Robinson Dooley (Castro Valley, CA), Gopinath Kuduvalli (San Jose, CA), James Wang (Palo Alto, CA), Eric Earnst (Saratoga, CA), Chris Raanes (Portola Valley, CA)
Application Number: 12/963,896
International Classification: A61B 6/00 (20060101);