Skip to content

thanhhau097/multimodal_brain_image_analysis

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 

Repository files navigation

Multimodal Brain Tumor Segmentation

Medical Application Overview (optional, no need to read)

IDEA

Original Result: 0.79574 0.89777 0.84114

Compare to paper: 80.12 90.62 84.54

  • Design new model: cascaded, because they are overlap
  • Multi-stage DMF
  • Apply variant of UNet architecture, compare to UNet architecture
  • Consider post-processing step to ignore 0.0 results.
Paper What can be handled? Methods Results Note
Modified DMF Net Use MFUnit in skip connections MFUnit 0.81131 0.90011 0.84194 memory consuming (6M parameters), just use MF Unit in skip connection
epoch 599 model
UNet++: A Nested U-Net Architecture for Medical Image Segmentation The re-designed skip pathways aim at reducing the semantic gap between the feature maps of the encoder and decoder sub-networks. We argue that the optimizer would deal with an easier learning task when the feature maps from the decoder and encoder networks are semantically similar nested and dense skip connections. memory consuming (too deep)
Attention U-Net: Learning Where to Look for the Pancreas suppress irrelevant regions in an input image while highlighting salient features attention gate module
Concurrent Spatial and Channel ‘Squeeze & Excitation’ in Fully Convolutional Networks recalibrating the feature maps adaptively, to boost meaningful features, while suppressing weak ones. We draw inspiration from the recently proposed squeeze & excitation (SE) module for channel recalibration of feature maps for image clas- sification SE modules
A NOVEL FOCAL TVERSKY LOSS FUNCTIONWITH IMPROVED ATTENTION U-NET FOR LESION SEGMENTATION highly imbalanced data and small ROI segmentation attention gate, focal Tversky loss function, multiscale input
Dense Multi-path U-Net for Ischemic Stroke Lesion Segmentation in Multiple Image Modalities
IVD-Net: Intervertebral disc localization and segmentation in MRI with a multi-modal UNet
instead of combining the available image modalities at the input, each of them is processed in a different path to better exploit their unique information. separate input, dense connection, inception module

Experiments

My code is located at: https://github.com/thanhhau097/pytorch-3dunet which was folked from https://github.com/wolny/pytorch-3dunet with modification.

Code references:

Model Params FLOPS Dice ET Dice WT Dice TC Note (Sum)
DMFNet 3.88M 80.12 (79.574) 90.62 (89.777) 84.54 (84.114) 255.28
DMFNet + MFUnit in skip connections 6.87M 81.131 90.011 84.194
BiFPNNet - 1 layer - 64 hidden (concatenate) 1.38M 79.643 90.633 84.919 255.195
BiFPNNet - 2 layer - 64 hidden (concatenate) 1.76M 80.075 90.678 85.043 255.796
BiFPNNet - 3 layer - 64 hidden (concatenate) 2.14M 81.191 89.791 84.423 255.405
DMFNet + multiscale inputs (PSP) 7M 77.853 89.636 84.723 (1 error file) good for WT and TC, bad for ET (may be because it is too small)
DMFNet + multiscale weighted inputs (PSP) 7M 79.471 90.284 84302
BiFPNNet - 1 layer - 128 hidden 80.518 89.458 83.669
BiFPNNet - 1 layer - 64 hidden (add) 1.07M
BiFPNNet - 2 layer - 64 hidden (add) 1.14M
BiFPNNet - 3 layer - 64 hidden (add) 1.21M
DMFNet + MFUnit in skip connections + interconnect
DMFNet + DMFUnit in skip connections 11300299 79.661 89.896 84.189
Attention Unet (one gate) 10881302 79.673 89.175 83.737
Attention Unet (single module) 11226614 79.431 89.708 82.755
Attention Unet (multi module) 12345818 79.571 89.42 83.14
DMFNet + csSE 4110041 79.653 89.908 84.566
DMFNet + PE (same paper with csSE) 4108946 71.56 82.421 71.082
DMFNet + attention gate, focal Tversky loss function
DMFNet + separate inputs (IVD architecture) 80.228 89.603 83.824

Note:

  • MFUnit is enough for multiscale and attention
  • We can improve by handling the difference between encoder features and decoder features, using multiscale input

Application of Medical Image Overview

I do a summarization of application from MICCAI 2019 papers here |

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published