Difference between revisions of "2017 Winter Project Week/DeepInfer"
(4 intermediate revisions by the same user not shown) | |||
Line 1: | Line 1: | ||
__NOTOC__ | __NOTOC__ | ||
− | + | <gallery> | |
+ | Image:PW-Winter2017.png|link=2017_Winter_Project_Week#Projects|[[2017_Winter_Project_Week#Projects|Projects List]] | ||
+ | <!-- Use the "Upload file" link on the left and then add a line to this list like "File:MyAlgorithmScreenshot.png" --> | ||
+ | File:Deepinfer arch.png|The initial prototype of the DeepInfer with OpenIGTLink. | ||
+ | </gallery> | ||
Line 27: | Line 31: | ||
<!-- Objective bullet points --> | <!-- Objective bullet points --> | ||
* Redesign the architecture of the toolkit considering Docker as the deep learning model deployment engine. | * Redesign the architecture of the toolkit considering Docker as the deep learning model deployment engine. | ||
− | * Discuss about the | + | * Discuss about the implementation details of Slicer side. |
* Planning the structure of the cloud model repository. | * Planning the structure of the cloud model repository. | ||
| | | | ||
<!-- Approach and Plan bullet points --> | <!-- Approach and Plan bullet points --> | ||
− | * | + | * Study and evaluate different approaches for passing input images and prediction results between Slicer and Docker including CLI. |
+ | * Decide about the necessary fields in the metadata of the stored models. | ||
+ | * Implement Slicer side to talk with Docker. | ||
| | | | ||
<!-- Progress and Next steps bullet points (fill out at the end of project week) --> | <!-- Progress and Next steps bullet points (fill out at the end of project week) --> | ||
− | * | + | * Helpful discussions about the design of Slicer part. |
+ | * We will use Github as the repository for storing the metadata information about different trained models. The repository will be the same as Slicer extensions index repository. The maintainers of the models will make pull requests to put their model meta data on the repository. | ||
+ | * The metadata file fields: Author, Institution, Organ, Task, Modality, Accuracy, Training methodology, license, data source, network details, docker image location, version. | ||
+ | * Slicer side scenario for using a deep model: | ||
+ | * Docker generates the XML of CLI | ||
+ | * Slicer reads the XML (using QSlicerCLIModule) and generates the required GUI for the task | ||
+ | * User selects input/output | ||
+ | * The input images would be saved to a temporary directory | ||
+ | * The docker will be run with the parameters on the mounted temporary directory | ||
+ | * Slicer waits for completion of the task. | ||
+ | * Docker saves the results of processing to the temp directory | ||
+ | * Slicer loads the results. | ||
+ | * Next steps | ||
+ | * Implement instantiation of CLI UI from the XML description in Python. | ||
+ | * Test the slicer/docker communication on our prostate-segmenter model. | ||
+ | |||
|} | |} |
Latest revision as of 14:56, 13 January 2017
Home < 2017 Winter Project Week < DeepInfer
Key Investigators
- Alireza Mehrtash (BWH, UBC)
- Mehran Pesteie (UBC)
- Yang (Silvia) Yixin (Tianjin University)
- Tina Kapur (BWH)
- Sandy Wells (BWH)
- Purang Abolmaesumi (UBC)
- Andriy Fedorov (BWH)
Background and References
Deep learning models have outperformed some of the previous state-of-the-art approaches in medical image analysis. However, utilizing deep models during image-guided therapy procedures requires integration of several software components which is often a tedious taskfor clinical researchers. Hence, there is a gap between the state-of-the-art machine learning research and itsapplication in clinical setup.
DeepInfer enables 3D Slicer to connect to a powerful processing back-end either on the local machine or a remote processing server. Utilizing a repository of pre-trained, task-specific models, DeepInfer allows clinical researchers and biomedical engineers to choose and deploy a model on new data without the need for software development or configuration.
Project Description
Objective | Approach and Plan | Progress and Next Steps |
---|---|---|
|
|
* Docker generates the XML of CLI * Slicer reads the XML (using QSlicerCLIModule) and generates the required GUI for the task * User selects input/output * The input images would be saved to a temporary directory * The docker will be run with the parameters on the mounted temporary directory * Slicer waits for completion of the task. * Docker saves the results of processing to the temp directory * Slicer loads the results.
* Implement instantiation of CLI UI from the XML description in Python. * Test the slicer/docker communication on our prostate-segmenter model. |