Difference between revisions of "NA-MIC/Projects/Collaboration/MGH RadOnc"

From NAMIC Wiki
Jump to: navigation, search
 
(32 intermediate revisions by 2 users not shown)
Line 1: Line 1:
 +
__NOTOC__
 +
 +
===Adaptive RT for Head and Neck===
 +
 +
This first example shows the kind of anatomic changes that can occur in head and neck cancer.  The pre-treatment scan is in green, and the mid-treatment scan is in red.  The image on the left is the rigid registration, on the right is deformable registration. 
 +
 +
{|
 +
|[[Image:RadOnc_HN2_rigid.png|thumb|380px|Rigid registration]]
 +
|[[Image:RadOnc_HN2_deformable.png|thumb|380px|Deformable registration]]
 +
|}
 +
 +
Here is another pertinent example for head and neck.  In axial view, there appears to be some weight loss.  Note the change in positioning of the mandible, and also the twisting of the cervical spine between scans.  Also note the strong CT artifacts caused by dental fillings.  In both examples, registration of the soft palette is worse using deformable registration than rigid registration, probably due to these artifacts.
 +
 +
{|
 +
|[[Image:RadOnc_HN5_rigid.png|thumb|380px|Rigid registration]]
 +
|[[Image:RadOnc_HN5_deformable.jpeg|thumb|380px|Deformable registration]]
 +
|}
 +
 +
 +
 +
===Adaptive RT for Thorax===
 +
 +
This example shows anatomic change in the thorax.  The patient has a collapsed left lower lobe in the pre-treatment scan (top), which has recovered in the mid-treatment scan (bottom).  Notice there is some kind of fluid accumulation below the collapsed lung.
 +
 
{|
 
{|
|[[Image:ProjectWeek-2008.png|thumb|320px|Return to [[2008_Summer_Project_Week|Project Week Main Page]] ]]
+
|[[Image:RadOnc_LU1.png|thumb|380px|Lung cancer 1]]
|[[Image:abdomen.jpg|thumb|320px|Registration of abdomen.]]
 
 
|}
 
|}
  
__NOTOC__
+
Here is another example, showing the magnitude of tumor regression that can occur during treatment.  (Sorry, low resolution image).
  
===Key Investigators===
+
{|
* MGH: Greg Sharp, Marta Peroni
+
|[[Image:RadOnc_LU2.png|thumb|380px|Lung cancer 2]]
* NA-MIC: Steve Pieper, Wendy Plesniak
+
|}
* BWH: Nicole Aucoin (fiducials), Katie Hayes (CTest)
 
  
 +
Thorax is a special case.  Patient images are acquired using 4D-CT, and radiation treatment plan can be evaluated at each breathing phase. The volumes are aligned using deformable registration, and radiation dose from each phase is accumulated into a reference phase (e.g. exhale).  This process is called "4D treatment planning."
  
<div style="margin: 20px;">
+
Deformable registration of the phases within a 4D-CT is considered "easy".  The reason for this is:
  
<div style="width: 23%; float: left; padding-right: 3%;">
+
# Single-session imaging, so patient is already co-registered
 +
# Single-session imaging, so no anatomic change
 +
# High contrast of vessels against lung parenchema
  
<h1>Objective</h1>
+
However, the sliding of the lungs against the chest wall is difficult to model. We sometimes segment the images at the pleural boundary. This allows us to separate the moving set of organs from the non-moving set, which are registered separately. Ideally we would always do this, but segmentation is manual and therefore we usually skip this step.
The long term objective is to improve interactive tools for radiation therapy planning. Specifically, we would like to improve the reliability and usability of image registration and segmentation of critical organs.
 
  
</div>
+
{|
 +
|[[Image:RadOnc_LU3.png|thumb|300px|Segmentation at pleural boundary]]
 +
|[[Image:RadOnc_LU4_before.png|thumb|300px|4D-CT phases]]
 +
|[[Image:RadOnc_LU4_after.png|thumb|300px|Registration of 4D-CT phases]]
 +
|}
  
<div style="width: 33%; float: left; padding-right: 3%;">
+
===General Discussion of Registration===
  
<h1>Approach, Plan</h1>
+
Deformable registration is still not as reliable as it should be.  Here are some of my complaints:
Workshop goals
 
* Complete the CLP interface between our in-house registration software (plastimatch) and slicer3
 
* Get ideas how the user interface can be implemented
 
* Learn how to interface with slicer3 fiducial and label maps for interactive deformable registration
 
* Learn how CTest can be used for automated testing
 
  
</div>
+
# Image acquisition has residual artifacts which cause unrealistic deformations (dental artifacts in H&N, motion artifacts in thorax).
 +
# Registration algorithms are not always robust, and require experimentation and tuning.
 +
# Validation of registration results is not easy, since there are inadequate tools.
 +
# In thorax, temporal regularization is generally not done, because of slow algorithms and large memory footprints.
 +
# For cone-beam CT (and sometimes MR), intensity values are not globally stable.  The most common suggestion is adaptive histogram equalization, but isn't there a better way?
 +
# Interactive tools to repair (or reinitialize) registration are virtually non-existent.
  
<div style="width: 38%; float: left;">
+
===Segmentation===
  
<h1>Progress</h1>
+
Rad Onc departments use interactive segmentation every day for both research and patient care.
Since last year
+
Prior to treatment planning, the target and critical structures are delineated in CT. The current state of the art is manual segmentation in axial view. A outline tool, used delineate the boundary, is generally prefered over a paintbrush tool that fills pixels. Commercial products generally support some subset of the following tools to assist the operator.
* Plastimatch software released under open source software license (BSD-style).
 
* The scaffolding code for slicer3 CLP is implemented, but not working because plastimatch doesn't read nrrd (yet).
 
  
</div>
+
# contour interpolation between slices
 +
# boundary editing
 +
# mixed axial/coronal/sagittal drawing
 +
# livewire or intelligent scissors
 +
# drawing constraints (e.g. constraints on volume overlap/distance)
 +
# post-processing tools to nudge or smooth the boundary
  
<br style="clear: both;" />
+
There are many opportunities for using computer vision algorithms to improve interactive segmentation.  For example, using prior models of shape and intensity to improve interpolation.  Automatic segmentation also exists is several commercial products, each with impressive demos.  We have GE Adv Sim software which does model-based segmentation of structures such as spinal cord, lens & optic nerve, liver.  For H&N, atlas-based segmentation is the most popular approach.
  
===Progress===
+
Below are examples of segmentations of targets and critical structures for head and neck, and thorax.  These structures (or similar) would be manually delineated for every patient who gets 3D planning.
Under construction.
 
  
 
{|
 
{|
|[[Image:gcsGUI.png|thumb|320px|Return to [[2008_Summer_Project_Week|Project Week Main Page]] ]]
+
|[[Image:RadOnc_HN_seg.png|thumb|380px|Head and Neck segmentation]]
|[[Image:gcsImage.jpg|thumb|320px|Registration of abdomen.]]
+
|
 +
{| class="wikitable"
 +
|-
 +
! optic chiasm
 +
! med green
 +
|-
 +
! brain stem
 +
! dim green
 +
|-
 +
! spinal cord
 +
! bright green
 +
|-
 +
! left parotid
 +
! violet
 +
|-
 +
! right parotid
 +
! dim blue
 +
|-
 +
!oral cavity
 +
! cyan
 +
|-
 +
! mandible
 +
! pink
 +
|-
 +
! larynx
 +
! bright blue
 +
|-
 +
! esophagus
 +
! orange
 +
|-
 +
! (target)
 +
! red
 +
|-
 +
! (target)
 +
! yellow
 +
|}
 
|}
 
|}
  
</div>
+
{|
 
+
|[[Image:RadOnc_LU_seg.png|thumb|380px|Thorax segmentation]]
 
+
|
===References===
+
{| class="wikitable"
* GC Sharp, Z Wu, N Kandasamy, "A Data Structure for B-Spline Registration," AAPM 50, Houston TX, July 2008.
+
|-
* V Boldea, GC Sharp, SB Jiang, D Sarrut, “4D-CT lung motion estimation with deformable registration: quantification of motion nonlinearity and hysteresis,” Medical Physics, Vol 35, No 3, pp 1008-1018, March 2008.
+
! left lung
* Z Wu, E Rietzel, V Boldea, D Sarrut, GC Sharp, "Evaluation of deformable registration of patient lung 4DCT with sub-anatomical region segmentations," Medical Physics, Vol 35, No 2, pp 775-81, February 2008.
+
! dark red
* GC Sharp, N Kandasamy, H Singh, M Folkert, "GPU-based streaming architectures for fast cone-beam CT image reconstruction and demons deformable registration,” Physics in Medicine and Biology, Vol 52, No 19, pp 5771--83, October 7, 2007.
+
|-
 +
! right lung
 +
! green
 +
|-
 +
! esophagus
 +
! violet
 +
|-
 +
! heart
 +
! cyan
 +
|-
 +
! cord
 +
! yellow
 +
|-
 +
! (target)
 +
! light red
 +
|}
 +
|}

Latest revision as of 22:26, 8 June 2009

Home < NA-MIC < Projects < Collaboration < MGH RadOnc


Adaptive RT for Head and Neck

This first example shows the kind of anatomic changes that can occur in head and neck cancer. The pre-treatment scan is in green, and the mid-treatment scan is in red. The image on the left is the rigid registration, on the right is deformable registration.

Rigid registration
Deformable registration

Here is another pertinent example for head and neck. In axial view, there appears to be some weight loss. Note the change in positioning of the mandible, and also the twisting of the cervical spine between scans. Also note the strong CT artifacts caused by dental fillings. In both examples, registration of the soft palette is worse using deformable registration than rigid registration, probably due to these artifacts.

Rigid registration
Deformable registration


Adaptive RT for Thorax

This example shows anatomic change in the thorax. The patient has a collapsed left lower lobe in the pre-treatment scan (top), which has recovered in the mid-treatment scan (bottom). Notice there is some kind of fluid accumulation below the collapsed lung.

Lung cancer 1

Here is another example, showing the magnitude of tumor regression that can occur during treatment. (Sorry, low resolution image).

Lung cancer 2

Thorax is a special case. Patient images are acquired using 4D-CT, and radiation treatment plan can be evaluated at each breathing phase. The volumes are aligned using deformable registration, and radiation dose from each phase is accumulated into a reference phase (e.g. exhale). This process is called "4D treatment planning."

Deformable registration of the phases within a 4D-CT is considered "easy". The reason for this is:

  1. Single-session imaging, so patient is already co-registered
  2. Single-session imaging, so no anatomic change
  3. High contrast of vessels against lung parenchema

However, the sliding of the lungs against the chest wall is difficult to model. We sometimes segment the images at the pleural boundary. This allows us to separate the moving set of organs from the non-moving set, which are registered separately. Ideally we would always do this, but segmentation is manual and therefore we usually skip this step.

Segmentation at pleural boundary
4D-CT phases
Registration of 4D-CT phases

General Discussion of Registration

Deformable registration is still not as reliable as it should be. Here are some of my complaints:

  1. Image acquisition has residual artifacts which cause unrealistic deformations (dental artifacts in H&N, motion artifacts in thorax).
  2. Registration algorithms are not always robust, and require experimentation and tuning.
  3. Validation of registration results is not easy, since there are inadequate tools.
  4. In thorax, temporal regularization is generally not done, because of slow algorithms and large memory footprints.
  5. For cone-beam CT (and sometimes MR), intensity values are not globally stable. The most common suggestion is adaptive histogram equalization, but isn't there a better way?
  6. Interactive tools to repair (or reinitialize) registration are virtually non-existent.

Segmentation

Rad Onc departments use interactive segmentation every day for both research and patient care. Prior to treatment planning, the target and critical structures are delineated in CT. The current state of the art is manual segmentation in axial view. A outline tool, used delineate the boundary, is generally prefered over a paintbrush tool that fills pixels. Commercial products generally support some subset of the following tools to assist the operator.

  1. contour interpolation between slices
  2. boundary editing
  3. mixed axial/coronal/sagittal drawing
  4. livewire or intelligent scissors
  5. drawing constraints (e.g. constraints on volume overlap/distance)
  6. post-processing tools to nudge or smooth the boundary

There are many opportunities for using computer vision algorithms to improve interactive segmentation. For example, using prior models of shape and intensity to improve interpolation. Automatic segmentation also exists is several commercial products, each with impressive demos. We have GE Adv Sim software which does model-based segmentation of structures such as spinal cord, lens & optic nerve, liver. For H&N, atlas-based segmentation is the most popular approach.

Below are examples of segmentations of targets and critical structures for head and neck, and thorax. These structures (or similar) would be manually delineated for every patient who gets 3D planning.

Head and Neck segmentation
optic chiasm med green
brain stem dim green
spinal cord bright green
left parotid violet
right parotid dim blue
oral cavity cyan
mandible pink
larynx bright blue
esophagus orange
(target) red
(target) yellow
Thorax segmentation
left lung dark red
right lung green
esophagus violet
heart cyan
cord yellow
(target) light red