DISC-GAN: Disentangling Style and Content for Cluster-Specific Synthetic Underwater Image Generation
Varur, Sneha, Hanchinamani, Anirudh R, Bagewadi, Tarun S, Mudenagudi, Uma, Desai, Chaitra D, C, Sujata, Desai, Padmashree, Meharwade, Sumit
–arXiv.org Artificial Intelligence
In this paper, we propose a novel framework, Disentangled Style-Content GAN (DISC-GAN), which integrates style-content disentanglement with a cluster-specific training strategy towards photorealistic underwater image synthesis. The quality of synthetic underwater images is challenged by optical distortions due to phenomena such as color attenuation and turbidity. These phenomena are represented by distinct stylistic variations across different wa-terbodies, such as changes in tint and haze. While generative models are well-suited to capture complex patterns, they often lack the ability to model the non-uniform stylistic conditions of diverse underwater environments. T o address these challenges, we employ K-means clustering to partition a dataset into style-specific domains. W e use separate encoders to get latent spaces for style and content; we further integrate these latent representations via Adaptive Instance Normalization (AdaIN) and decode the result to produce the final synthetic image. The model is trained independently on each style cluster to preserve domain-specific characteristics. Our framework demonstrates state-of-the-art performance, obtaining a Structural Similarity Index (SSIM) of 0.9012, an average Peak Signal-to-Noise Ratio (PSNR) of 32.5118 dB, and a Fr echet Inception Distance (FID) of 13.3728.
arXiv.org Artificial Intelligence
Oct-14-2025