-
Notifications
You must be signed in to change notification settings - Fork 5
Description
Description
I am seeking insight regarding significant artifacts appearing during ESRGAN training on the SEN2NAIP dataset. These appear to be convolutional or checkerboard artifacts and are visible across the entire image, particularly in the low-frequency regions.
Training Details
The issue persists across different training stages:
PSNR-oriented pre-training: Artifacts are present even when training with only content loss (no Discriminator).
Full GAN training: The problem remains visible until early stopping.
Attempted Mitigations
To address this (an issue already present in the original repository), I have implemented several features from the original ESRGAN paper, though they have not yet resolved the problem:
Smaller initialization: Adjusted weights initialization to improve stability.
Relativistic Average Loss (RaGAN): Implemented for the GAN phase.
Perpetual Pre-training: Added the g_pretrain_steps: -1 feature to allow continuous generator pre-training with content loss, as suggested in the original literature.
This justifies some new keys available in the attached configurations.
Configurations
I have attached the configuration files used:
Pre-training config: Generator-only with content loss.
Full training config: Resumed training with combined losses.
And some visual examples:
Please zoom into the low-frequency parts of the attached W&B examples to better visualize the pattern.