Back to Search Start Over

Multi-Class Guided GAN for Remote-Sensing Image Synthesis Based on Semantic Labels.

Authors :
Niu, Zhenye
Li, Yuxia
Gong, Yushu
Zhang, Bowei
He, Yuan
Zhang, Jinglin
Tian, Mengyu
He, Lei
Source :
Remote Sensing; Jan2025, Vol. 17 Issue 2, p344, 28p
Publication Year :
2025

Abstract

In the scenario of limited labeled remote-sensing datasets, the model's performance is constrained by the insufficient availability of data. Generative model-based data augmentation has emerged as a promising solution to this limitation. While existing generative models perform well in natural scene domains (e.g., faces and street scenes), their performance in remote sensing is hindered by severe data imbalance and the semantic similarity among land-cover classes. To tackle these challenges, we propose the Multi-Class Guided GAN (MCGGAN), a novel network for generating remote-sensing images from semantic labels. Our model features a dual-branch architecture with a global generator that captures the overall image structure and a multi-class generator that improves the quality and differentiation of land-cover types. To integrate these generators, we design a shared-parameter encoder for consistent feature encoding across two branches, and a spatial decoder that synthesizes outputs from the class generators, preventing overlap and confusion. Additionally, we employ perceptual loss ( L V G G ) to assess perceptual similarity between generated and real images, and texture matching loss ( L T ) to capture fine texture details. To evaluate the quality of image generation, we tested multiple models on two custom datasets (one from Chongzhou, Sichuan Province, and another from Wuzhen, Zhejiang Province, China) and a public dataset LoveDA. The results show that MCGGAN achieves improvements of 52.86 in FID, 0.0821 in SSIM, and 0.0297 in LPIPS compared to the Pix2Pix baseline. We also conducted comparative experiments to assess the semantic segmentation accuracy of the U-Net before and after incorporating the generated images. The results show that data augmentation with the generated images leads to an improvement of 4.47% in FWIoU and 3.23% in OA across the Chongzhou and Wuzhen datasets. Experiments show that MCGGAN can be effectively used as a data augmentation approach to improve the performance of downstream remote-sensing image segmentation tasks. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
20724292
Volume :
17
Issue :
2
Database :
Complementary Index
Journal :
Remote Sensing
Publication Type :
Academic Journal
Accession number :
182445404
Full Text :
https://doi.org/10.3390/rs17020344