Back to Search
Start Over
Multi-Class Guided GAN for Remote-Sensing Image Synthesis Based on Semantic Labels.
- Source :
- Remote Sensing; Jan2025, Vol. 17 Issue 2, p344, 28p
- Publication Year :
- 2025
-
Abstract
- In the scenario of limited labeled remote-sensing datasets, the model's performance is constrained by the insufficient availability of data. Generative model-based data augmentation has emerged as a promising solution to this limitation. While existing generative models perform well in natural scene domains (e.g., faces and street scenes), their performance in remote sensing is hindered by severe data imbalance and the semantic similarity among land-cover classes. To tackle these challenges, we propose the Multi-Class Guided GAN (MCGGAN), a novel network for generating remote-sensing images from semantic labels. Our model features a dual-branch architecture with a global generator that captures the overall image structure and a multi-class generator that improves the quality and differentiation of land-cover types. To integrate these generators, we design a shared-parameter encoder for consistent feature encoding across two branches, and a spatial decoder that synthesizes outputs from the class generators, preventing overlap and confusion. Additionally, we employ perceptual loss ( L V G G ) to assess perceptual similarity between generated and real images, and texture matching loss ( L T ) to capture fine texture details. To evaluate the quality of image generation, we tested multiple models on two custom datasets (one from Chongzhou, Sichuan Province, and another from Wuzhen, Zhejiang Province, China) and a public dataset LoveDA. The results show that MCGGAN achieves improvements of 52.86 in FID, 0.0821 in SSIM, and 0.0297 in LPIPS compared to the Pix2Pix baseline. We also conducted comparative experiments to assess the semantic segmentation accuracy of the U-Net before and after incorporating the generated images. The results show that data augmentation with the generated images leads to an improvement of 4.47% in FWIoU and 3.23% in OA across the Chongzhou and Wuzhen datasets. Experiments show that MCGGAN can be effectively used as a data augmentation approach to improve the performance of downstream remote-sensing image segmentation tasks. [ABSTRACT FROM AUTHOR]
Details
- Language :
- English
- ISSN :
- 20724292
- Volume :
- 17
- Issue :
- 2
- Database :
- Complementary Index
- Journal :
- Remote Sensing
- Publication Type :
- Academic Journal
- Accession number :
- 182445404
- Full Text :
- https://doi.org/10.3390/rs17020344