TLDR: MInDI-3D is a new 3D deep learning model that uses an iterative denoising process to remove artifacts from sparse-view Cone Beam Computed Tomography (CBCT) images. This allows for significant reductions in patient radiation exposure (up to 8x) while maintaining or improving image quality. The model was trained on a large pseudo-CBCT dataset and validated on real-world patient scans, showing strong performance in artifact removal, generalization to different anatomies and scanner types, and clinical utility for patient positioning.
Medical imaging plays a crucial role in diagnosing and treating various conditions, especially in fields like radiation therapy. One such imaging modality is Cone Beam Computed Tomography (CBCT), which provides 3D X-ray images. While CBCT is widely used for tasks like patient positioning and tumor contouring, it faces significant challenges. These include image quality degradation due to artifacts from patient motion, metal implants, and, importantly, the need to reduce radiation exposure from repeated scans.
To address the concern of cumulative radiation exposure, researchers have explored ‘sparse-view CBCT,’ which involves using fewer X-ray projections. However, this often leads to streak artifacts that compromise image quality and clinical utility. Traditional methods struggle to reconstruct high-quality images from such limited data.
Introducing MInDI-3D: A New Approach to CBCT Image Enhancement
A recent research paper, MInDI-3D: Iterative Deep Learning in 3D for Sparse-view Cone Beam Computed Tomography, introduces a novel solution called MInDI-3D (Medical Inversion by Direct Iteration in 3D). This model is the first 3D conditional diffusion-based model designed specifically for removing artifacts in real-world sparse-view CBCT images, with the ultimate goal of reducing patient radiation exposure.
A key innovation of MInDI-3D is its extension of the ‘InDI’ (Inversion by Direct Iteration) concept from 2D to a full 3D volumetric approach for medical images. Unlike many 3D methods that compress data to manage memory, MInDI-3D operates directly in the 3D voxel space, preserving anatomical detail. It employs an iterative denoising process that refines the CBCT volume directly from sparse-view input, gradually enhancing image quality in incremental steps.
Training and Evaluation
To train MInDI-3D robustly, the researchers generated a large pseudo-CBCT dataset comprising 16,182 chest CT volumes from the public CT-RATE dataset. This addresses the common challenge of data scarcity in medical imaging. For testing, a real-world CBCT dataset called HyperSight, from 16 cancer patients, was used. This allowed for comprehensive evaluation, including quantitative metrics, scalability analysis, generalization tests, and a crucial clinical assessment.
Promising Results and Clinical Utility
The evaluation showed MInDI-3D’s effectiveness. It achieved significant improvements in image quality metrics, such as a 12.96 dB PSNR gain over uncorrected scans with only 50 projections on the pseudo-CBCT test set. Importantly, it enabled an 8x reduction in imaging radiation exposure. The model also demonstrated strong generalization capabilities, performing well on real-world scans from different anatomical sites (abdomen, breast, lung) and new CBCT scanner geometries, even though the training data was primarily chest CTs.
Scalability was also a highlight, with performance improving as more training data was used. MInDI-3D also offers a unique ‘perception-distortion trade-off,’ allowing users to adjust the number of iterative steps to prioritize either quantitative fidelity (pixel-level accuracy) or perceptual realism (how realistic the image appears to a human observer). This flexibility can be tailored to specific clinical needs.
A critical aspect of the study was the clinical assessment by 11 clinicians. They rated MInDI-3D’s output as sufficient for patient positioning across all anatomical sites (90-100% acceptance). For tasks like dose calculation and contouring, the acceptance rates were mixed but generally lower, particularly for abdomen scans, highlighting areas for future refinement. However, lung scans showed the highest acceptance for these tasks, possibly due to inherent anatomical advantages and consistency with the training data.
Also Read:
- MIND: A New AI Model for Adaptive Denoising of Medical Images
- Advancing Medical Image Fusion with MMIF-AMIN: A New Approach for Comprehensive Diagnostics
Future Directions
While MInDI-3D represents a significant advancement, future work will focus on further investigating the trade-off between perceived image quality and anatomical fidelity, especially with increased iteration steps. The goal is to ensure that visual improvements do not compromise diagnostic accuracy or introduce synthetic features that could mislead clinical interpretation. This research establishes conditional generative-based models as viable tools for sparse-view CBCT restoration, paving the way for advancements in adaptive radiotherapy.


