Medical Application Overview (optional, no need to read)
Original Result: 0.79574 0.89777 0.84114
Compare to paper: 80.12 90.62 84.54
- Design new model: cascaded, because they are overlap
- Multi-stage DMF
- Apply variant of UNet architecture, compare to UNet architecture
- Consider post-processing step to ignore 0.0 results.
Paper | What can be handled? | Methods | Results | Note |
---|---|---|---|---|
Modified DMF Net | Use MFUnit in skip connections | MFUnit | 0.81131 0.90011 0.84194 | memory consuming (6M parameters), just use MF Unit in skip connection epoch 599 model |
UNet++: A Nested U-Net Architecture for Medical Image Segmentation | The re-designed skip pathways aim at reducing the semantic gap between the feature maps of the encoder and decoder sub-networks. We argue that the optimizer would deal with an easier learning task when the feature maps from the decoder and encoder networks are semantically similar | nested and dense skip connections. | memory consuming (too deep) | |
Attention U-Net: Learning Where to Look for the Pancreas | suppress irrelevant regions in an input image while highlighting salient features | attention gate module | ||
Concurrent Spatial and Channel ‘Squeeze & Excitation’ in Fully Convolutional Networks | recalibrating the feature maps adaptively, to boost meaningful features, while suppressing weak ones. We draw inspiration from the recently proposed squeeze & excitation (SE) module for channel recalibration of feature maps for image clas- sification | SE modules | ||
A NOVEL FOCAL TVERSKY LOSS FUNCTIONWITH IMPROVED ATTENTION U-NET FOR LESION SEGMENTATION | highly imbalanced data and small ROI segmentation | attention gate, focal Tversky loss function, multiscale input | ||
Dense Multi-path U-Net for Ischemic Stroke Lesion Segmentation in Multiple Image Modalities IVD-Net: Intervertebral disc localization and segmentation in MRI with a multi-modal UNet |
instead of combining the available image modalities at the input, each of them is processed in a different path to better exploit their unique information. | separate input, dense connection, inception module |
My code is located at: https://github.com/thanhhau097/pytorch-3dunet which was folked from https://github.com/wolny/pytorch-3dunet with modification.
Code references:
https://github.com/pykao/BraTS2018-tumor-segmentation: models, criterions, transforms [1]https://github.com/MIC-DKFZ/BraTS2017: dataset- https://github.com/China-LiuXiaopeng/BraTS-DMFNet: main (3)
- https://github.com/xf4j/brats18/tree/master/models: (pre_processingN4ITK)
- https://github.com/wolny/pytorch-3dunet [2] (5)
Model | Params | FLOPS | Dice ET | Dice WT | Dice TC | Note (Sum) |
---|---|---|---|---|---|---|
DMFNet | 3.88M | 80.12 (79.574) | 90.62 (89.777) | 84.54 (84.114) | 255.28 | |
DMFNet + MFUnit in skip connections | 6.87M | 81.131 | 90.011 | 84.194 | ||
BiFPNNet - 1 layer - 64 hidden (concatenate) | 1.38M | 79.643 | 90.633 | 84.919 | 255.195 | |
BiFPNNet - 2 layer - 64 hidden (concatenate) | 1.76M | 80.075 | 90.678 | 85.043 | 255.796 | |
BiFPNNet - 3 layer - 64 hidden (concatenate) | 2.14M | 81.191 | 89.791 | 84.423 | 255.405 | |
DMFNet + multiscale inputs (PSP) | 7M | 77.853 | 89.636 | 84.723 | (1 error file) good for WT and TC, bad for ET (may be because it is too small) | |
DMFNet + multiscale weighted inputs (PSP) | 7M | 79.471 | 90.284 | 84302 | ||
BiFPNNet - 1 layer - 128 hidden | 80.518 | 89.458 | 83.669 | |||
BiFPNNet - 1 layer - 64 hidden (add) | 1.07M | |||||
BiFPNNet - 2 layer - 64 hidden (add) | 1.14M | |||||
BiFPNNet - 3 layer - 64 hidden (add) | 1.21M | |||||
DMFNet + MFUnit in skip connections + interconnect | ||||||
DMFNet + DMFUnit in skip connections | 11300299 | 79.661 | 89.896 | 84.189 | ||
Attention Unet (one gate) | 10881302 | 79.673 | 89.175 | 83.737 | ||
Attention Unet (single module) | 11226614 | 79.431 | 89.708 | 82.755 | ||
Attention Unet (multi module) | 12345818 | 79.571 | 89.42 | 83.14 | ||
DMFNet + csSE | 4110041 | 79.653 | 89.908 | 84.566 | ||
DMFNet + PE (same paper with csSE) | 4108946 | 71.56 | 82.421 | 71.082 | ||
DMFNet + attention gate, focal Tversky loss function | ||||||
DMFNet + separate inputs (IVD architecture) | 80.228 | 89.603 | 83.824 |
- MFUnit is enough for multiscale and attention
- We can improve by handling the difference between encoder features and decoder features, using multiscale input
I do a summarization of application from MICCAI 2019 papers here |