TLDR: Tooth-Diffusion is a novel AI model that uses a conditional diffusion framework to generate highly realistic 3D dental CBCT scans. It allows precise, fine-grained control over individual tooth presence and configuration, enabling virtual tooth addition, removal, and full dentition synthesis. The model integrates wavelet-based denoising, FiLM conditioning, and masked loss functions, demonstrating strong fidelity and generalization. This technology holds significant potential for surgical planning, patient communication, and data augmentation in dental AI workflows.
Dental Cone-Beam Computed Tomography (CBCT) scans are crucial tools in modern dentistry, providing high-resolution 3D images of teeth and jaw structures for diagnosis and treatment planning. However, generating anatomically realistic synthetic CBCT scans with precise control over individual teeth has been a significant challenge in medical image synthesis.
A novel approach, named Tooth-Diffusion, has been proposed to address this gap. This new conditional diffusion framework allows for the generation of 3D dental volumes with fine-grained control over the presence and configuration of individual teeth. The core of this method involves guiding the synthesis process using tooth-level binary attributes, meaning the model can be told which teeth should be present or absent.
How Tooth-Diffusion Works
The Tooth-Diffusion framework integrates several advanced techniques to achieve its remarkable capabilities. It utilizes a wavelet-based denoising diffusion model, which helps in processing the complex 3D data more efficiently by operating on multi-scale frequency components. This reduces computational demands while maintaining high fidelity in the generated images.
A key innovation is the use of Feature-wise Linear Modulation (FiLM) conditioning. This allows the network to dynamically adjust its internal features based on the desired tooth configuration, providing precise control over the synthesis process. Additionally, a masked L2 loss function is employed during training, which focuses the learning process specifically on the tooth regions, preventing the less informative background from dominating the learning.
Simulating Dental Scenarios
To make the model robust and clinically relevant, the researchers introduced unique augmentation strategies during training. These include simulating both tooth addition and tooth removal scenarios. For instance, in a tooth addition scenario, the model is trained to plausibly reconstruct missing teeth based on the surrounding anatomy. Conversely, in a tooth removal scenario, the model learns to suppress specified tooth regions, creating a realistic scan as if those teeth were never there or had been removed.
When simulating missing teeth, the method avoids simply creating empty holes. Instead, it uses an image-based inpainting strategy to fill the masked tooth cavity with anatomically plausible content, such as jawbone and air gradients, making the synthetic scans highly realistic. Data augmentation through horizontal flipping also enhances the training dataset, leveraging the natural symmetry of human dentition.
Also Read:
- A Unified AI Model for Comprehensive Facial and Dental Reconstruction
- Automated Dental Structure Mapping in CBCT Scans for Enhanced Patient Care
Impressive Results and Future Potential
The Tooth-Diffusion model was evaluated across various tasks, including full dentition synthesis, tooth addition, and tooth removal. Quantitative and qualitative assessments demonstrated strong fidelity and generalization. The model achieved low Fréchet Inception Distance (FID) scores, indicating high-quality image generation that closely resembles real scans. It also showed robust inpainting performance and high Structural Similarity Index Measure (SSIM) values, often above 0.91, even on previously unseen scans.
While the model performed exceptionally well for most teeth, some variations were noted for molars and wisdom teeth, likely due to their anatomical variability and scarcity in training datasets. Nevertheless, the ability to realistically modify dentition without requiring actual rescanning opens up significant opportunities for dental professionals. This includes enhanced surgical planning, improved patient communication through visual simulations of treatment outcomes, and targeted data augmentation for training other dental AI models.
This work represents a significant step forward in controllable 3D CBCT synthesis, offering a powerful tool for simulation and the development of more customizable generative models in dental imaging. For more detailed information, you can refer to the full research paper here.


