Should I render or should AI Generate? Crafting Synthetic Semantic Segmentation Datasets with Controlled Generation

Loading...
Thumbnail Image

Identifiers

Publication date

Authors

Advisors

Other responsabilities

Journal Title

Bibliographic citation

O. A. Mures, M. Silva, M. Lijó-Sanchez, E. J. Padrón and J. A. Iglesias-Guitian, "Should I render or should AI Generate? Crafting Synthetic Semantic Segmentation Datasets with Controlled Generation," in IEEE Computer Graphics and Applications, doi: 10.1109/MCG.2025.3553494

Type of academic work

Academic degree

Abstract

[Abstract]: This work explores the integration of generative AI models for automatically generating synthetic image-labeled data. Our approach leverages controllable Diffusion Models to generate synthetic variations of semantically labeled images. Synthetic datasets for semantic segmentation struggle to represent real-world subtleties, such as different weather conditions or fine details, typically relying on costly simulations and rendering. However, Diffusion Models can generate diverse images using input text prompts and guidance images, like semantic masks. Our work introduces and tests a novel methodology for generating labeled synthetic images, with an initial focus on semantic segmentation, a demanding computer vision task. We showcase our approach in two distinct image segmentation domains, outperforming traditional computer graphics simulations in efficiently creating diverse datasets and training downstream models. We leverage generative models for crafting synthetically labeled images, posing the question: “Should I render or should AI generate?”. Our results endorse a paradigm shift towards controlled generation models

Description

Versión aceptada, publicada en modo "Early Publication" por el editor.

Rights

Atribución 3.0 España
Atribución 3.0 España

Except where otherwise noted, this item's license is described as Atribución 3.0 España