Skip to content

MedicalImageAnalysisTutorials/Unsupervised-Domain-Adaptation-Framework-for-Vestibular-Schwannoma-and-Cochlea-Segmentation

 
 

Repository files navigation

Unsupervised-Domain-Adaptation-Framework-for-Vestibular-Schwannoma-and-Cochlea-Segmentation

An improved framework that performs domain adaptation on Vestibular Schwannomas and cochlea segmentation

  1. Due to the large file size of training and validation data, it is not feasible to upload it on github, users are required to download it from the crossMoDA 2021 challenge website: https://crossmoda.grand-challenge.org/

    • The training ceT1 MRI scans with the naming convention of crossmoda_XXX_ceT1.nii.gz should be placed under the directory of 'dataset\raw_dataset\training_data\source_sample\ceT1_MRI_scans'
    • The training ceT1 MRI segmentation masks with the naming convention of crossmoda_XXX_Label.nii.gz should be placed under the directory of 'dataset\raw_dataset\training_data\source_sample\ceT1_MRI_labels'
    • The target hrT2 MRI scans with the naming convention of crossmoda_XXX_hrT2.nii.gz should be placed under the directory of 'dataset\raw_dataset\training_data\target_sample'
    • The validation hrT2 MRI scans with the naming convention of crossmoda_XXX_hrT2.nii.gz should be placed under the directory of 'dataset\raw_dataset\validation_data'

    ** Please remove the sample files in those directories before moving them.

  2. Set up a new virtual environment

  3. Open command prompt

  4. locate command prompt to this directory

  5. run 'pip install -r requirements.txt'

  6. open the jupyter notebook file : 'Demo_code.ipynb'

About

An improved framework that performs domain adaptation on Vestibular Schwannomas and cochlea segmentation

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages

  • Jupyter Notebook 100.0%