spot_img
HomeResearch & DevelopmentGenerating Robust AI Models: How Uncertainty Guides Synthetic Data...

Generating Robust AI Models: How Uncertainty Guides Synthetic Data Creation

TLDR: UnIACorN is a new method that uses both semantic labels and a novel uncertainty measure from unlabeled data to generate synthetic images. This allows it to create diverse, labeled datasets that bridge domain gaps and improve the performance of discriminative models (like segmentation networks) on unseen, challenging data, without needing explicit style transfer or additional manual labeling.

In the rapidly evolving landscape of artificial intelligence, particularly in fields like medical imaging and autonomous driving, the demand for high-quality, labeled datasets is immense. Discriminative models, such as those used for semantic segmentation, thrive on extensive and diverse training data. However, acquiring such data is often challenging due to high manual labeling costs, privacy concerns, and the inherent difficulty in capturing all possible variations or “domain shifts” that a model might encounter in real-world applications.

Traditional data augmentation techniques, like simple image transformations, offer limited help when faced with significant domain shifts. Generative models, especially controlled diffusion models like ControlNet, have emerged as powerful tools for creating synthetic data. While these models can produce high-quality, labeled images, they often tend to reproduce the characteristics of their original training distribution. This limitation means they might not effectively generate data that represents “unknown” parts of a distribution, which are crucial for making models robust to new, unseen data.

UnIACorN: A New Approach to Synthetic Data Generation

A recent research paper introduces a novel method called Uncertainty-Aware ControlNet, or UnIACorN, designed to overcome these limitations. The core idea behind UnIACorN is to integrate the concept of “uncertainty” directly into the image generation process. This uncertainty acts as a signal, indicating that a particular image or data point was not part of the original training distribution of a downstream task, such as a segmentation network.

UnIACorN employs a dual control mechanism during image generation. It combines a “semantic control” derived from labeled datasets with an “uncertainty control” learned from unlabeled datasets. This innovative approach allows the model to generate annotated data that exhibits high uncertainty from a target domain – essentially creating synthetic data from an unlabeled distribution, complete with labels. This is particularly valuable for scenarios where labeled data is scarce but unlabeled data is abundant, such as in medical imaging with new device types.

How UnIACorN Works

The UnIACorN framework consists of three main components: a denoising diffusion probabilistic model (DDPM), a Semantic-ControlNet, and an Uncertainty-ControlNet.

  • DDPM Pre-training: This foundational model is trained to generate a wide variety of images, ensuring it can produce images from both labeled and unlabeled distributions.
  • Semantic-ControlNet: This component is trained exclusively on labeled data. Its role is to generate images that correspond to given ground-truth labels, ensuring semantic consistency.
  • Uncertainty-ControlNet: This is where UnIACorN truly innovates. It’s trained using uncertainty measurements from both labeled and unlabeled images. The uncertainty is typically measured as the pixel-wise entropy of a pre-trained segmentation network’s predictions. By conditioning on this uncertainty, the network learns to generate images that are “difficult” for the segmentation model, effectively creating “known-unknown” information.

During the inference phase, both the Semantic-ControlNet and the Uncertainty-ControlNet work in parallel. Their noise predictions are fused using a weighted sum, allowing for fine-grained control over both the segmentation mask and the level of uncertainty in the generated image. The uncertainties are sampled from a distribution derived from the unlabeled target data, steering the generation process towards creating images from the unseen distribution.

The entire process involves three steps: first, pre-training the generative blocks and a segmentation model; second, generating a new training distribution by sampling uncertainties and using label maps; and finally, retraining the discriminative network on this newly generated data. This retraining helps reduce the model’s uncertainties and significantly improves segmentation results for out-of-distribution data.

Experimental Validation and Impact

The researchers evaluated UnIACorN primarily in a medical imaging context, specifically with retinal Optical Coherence Tomography (OCT) scans. They used high-quality Spectralis OCTs as the labeled source domain and lower-quality HOME-OCTs (from self-examination devices) as the unlabeled target domain, which exhibit a large domain shift. The goal was to synthesize data that induces high uncertainty in a pre-trained retinal layer segmentation model, then use this synthetic data to refine the model.

The results were compelling. UnIACorN significantly improved segmentation performance on HOME-OCT data compared to training only on real Spectralis data or using other generative approaches like CycleGAN and CUT. While CycleGAN achieved good domain adaptation in terms of style, it often failed to maintain semantic consistency, leading to inaccurately labeled synthetic data. CUT offered better semantic consistency but a less pronounced domain shift. UnIACorN, however, successfully generated labeled data with uncertainty levels similar to the real HOME-OCT data, without explicitly learning the style or domain properties of the target data.

Beyond medical imaging, UnIACorN’s generalization ability was demonstrated in a traffic scene experiment, using Cityscapes as the labeled domain and ACDC (Adverse Conditions Dataset) as the unlabeled domain. This showed that UnIACorN is not merely a style transfer method; it can represent arbitrary domain shifts based on uncertainty, matching real-world scenarios with diverse and partly unknown data domains. The generated images for adverse conditions like fog or night were more convincing and led to improved segmentation results.

Also Read:

Conclusion

UnIACorN represents a significant advancement in synthetic data generation, particularly for scenarios plagued by data scarcity and domain shifts. By introducing an uncertainty-based conditioning mechanism, it allows for the effective utilization of unlabeled data to create diverse, labeled datasets that challenge and improve downstream discriminative models. This dual control over semantic information and uncertainty enables the generation of “known unknowns,” leading to more robust and accurate AI systems without requiring additional manual supervision. For more details, you can refer to the full research paper here.

Karthik Mehta
Karthik Mehtahttps://blogs.edgentiq.com
Karthik Mehta is a data journalist known for his data-rich, insightful coverage of AI news and developments. Armed with a degree in Data Science from IIT Bombay and years of newsroom experience, Karthik merges storytelling with metrics to surface deeper narratives in AI-related events. His writing cuts through hype, revealing the real-world impact of Generative AI on industries, policy, and society. You can reach him out at: [email protected]

- Advertisement -

spot_img

Gen AI News and Updates

spot_img

- Advertisement -