Multi-stage prediction networks for data harmonization
Highlights
- Deep learning for dMRI harmonization, leveraging limited datasets using multi-task learning (multiple independent networks are merged to accomplish a final task).
Method
-
Task: Predict multiple high quality targets (modern platforms) from a single low-quality input (out-of-date platform)
- Standard methods
- Registration to a common template
- Align mean and variance from each platform (voxelwise/regionally/whole image)
- Deep learning methods
- CNN-RISH (Rotation-Invariant Spherical Harmonics)
- DIQT (Deeper Image Quality Transfer)
- SHResNet
All existing deep learning methods use a single CNN to accomplish a single task (mapping the source to a single target platform). Traveling heads datasets (one subject scanned on multiple scanners) are very limited, which is problematic for deep learning.
Solution
- Use multi-task learning to train multiple networks on independent tasks, then use the last layers as input to a new network adated for the specific task.
- The final prediction is a linear combination of the task-specific predictions.
Model figure
Experiments
Data
- MUSHAC dataset
- 10 healthy subjects scanned on 3 different scanners
- GE 3T (“out-of-date” platform), Siemens Prisma 3T, Siemens Connectom 3T
- Standard (st) protocol: 2.4mm, 30 directions; SotA (sa) protocol: 1.2mm, 60 directions
- Model input: SH coefficients after deconvolution, 6th order (28 coefficients), processed as patches of 11^3 voxels
- Target: Patches of 11^3 (harmonization) or 11^9 (harmonization + super-resolution)
Results
Comments
- MSP reuses many pretrained models, and adds many new sub-networks, so it has much more parameters than any single model, which might give it an advantage over others.
- The final MSP layer is still task-specific.
- No subjects were specifically left out for the test set.
- They seem to choose the task-specific models based on the test set evaluation, which may indicate data contamination.