NoisyTwins: Class-Consistent and Diverse Image Generation Through StyleGANs
| dc.contributor.author | Rangwani H.; Bansal L.; Sharma K.; Karmali T.; Jampani V.; Babu R.V. | |
| dc.date.accessioned | 2025-05-23T11:16:54Z | |
| dc.description.abstract | StyleGANs are at the forefront of controllable image generation as they produce a latent space that is semantically disentangled, making it suitable for image editing and manipulation. However, the performance of StyleGANs severely degrades when trained via class-conditioning on large-scale long-tailed datasets. We find that one reason for degradation is the collapse of latents for each class in the W latent space. With NoisyTwins, we first introduce an effective and inexpensive augmentation strategy for class embeddings, which then decorrelates the latents based on self-supervision in the W space. This decorrelation mitigates collapse, ensuring that our method preserves intra-class diversity with class-consistency in image generation. We show the effectiveness of our approach on large-scale real-world long-tailed datasets of ImageNet-LT and iNaturalist 2019, where our method outperforms other methods by ∼ 19% on FID, establishing a new state-of-the-art. © 2023 IEEE. | |
| dc.identifier.doi | https://doi.org/10.1109/CVPR52729.2023.00580 | |
| dc.identifier.uri | http://172.23.0.11:4000/handle/123456789/6823 | |
| dc.relation.ispartofseries | Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition | |
| dc.title | NoisyTwins: Class-Consistent and Diverse Image Generation Through StyleGANs |