๐ง ReCoSeg โ Residual-Guided Cross-Modal Diffusion for Brain Tumor Segmentation
๐ Goal:
Improve brain tumor segmentation accuracy by synthesizing missing modalities using diffusion models and leveraging residual errors as attention cues for efficient, interpretable refinement.
๐ง Domain: Medical Imaging & Semi-Supervised Deep Learning
๐ฏ Task: Brain Tumor Segmentation (Whole Tumor)
๐ Dataset: BraTS 2020 โ Multimodal MRI (355 subjects; FLAIR, T1, T2, T1ce)
Project Domain
Imaging
Task
Segmentation
The Goal:
Manual segmentation of brain tumors is time-consuming and dependent on expert annotations, while available labels may be sparse in real clinical settings. ReCoSeg addresses this by splitting the problem into two phases: cross-modal T1ce synthesis using diffusion and residual-guided segmentation to correct tumor-related discrepancies. The aim is to enhance accuracy, reduce labeling burden, and improve clinical interpretability.
1
The Challenge:
Methodology & Process
๐ Stage 1 โ Cross-Modal Diffusion Synthesis
A DDPM generates the T1ce modality from FLAIR, T1, and T2 MRI, trained with BCE + Dice loss to improve structural alignment.
๐ Stage 2 โ Residual-Guided Segmentation
The absolute difference between synthesized and real T1ce is computed as a residual map that highlights possible tumor regions.
Residuals + original modalities โ lightweight U-Net for final segmentation.
๐ Training Setup
Input: Axial slices 120ร120
Loss: BCE + Dice
Optimizer: Adam, LR=2e-4
Hardware: RTX 3090 GPU
Semi-supervised design for robustness with missing modalities
2
The Result
๐ Results & Findings
Model | Dice โ | IoU โ |
|---|---|---|
UNet2D | 0.784 | 0.736 |
UNet3D | 0.842 | 0.743 |
DDMM-Synth | 0.872 | 0.811 |
ReCoSeg (Proposed) | 0.917 | 0.853 |
โ
Key Insight:
Residual maps explain and correct segmentation mistakes increasing boundary precision while keeping the model lightweight and efficient.
๐ Visuals Included:
Real vs. Synthesized T1ce
Residual localization over tumor areas
Segmentation comparison vs. baselines
๐ Conclusion
ReCoSeg provides an interpretable, semi-supervised, and computationally efficient segmentation pipeline that learns from cross-modal discrepancies.
Its strong performance with limited annotations and missing modalities supports real-world clinical adoption and motivates future extension to 3D diffusion and multi-class segmentation.
View Manuscript here:
Recoseg
3








