-
Notifications
You must be signed in to change notification settings - Fork 2
Open
Description
Hi, thanks for sharing a great work!
I'm trying to use the repo for paired 2D image enhancement. I'm using 2 datasets, one for grayscale image enhancement and one for RGB color normalization tasks.
One of the main issue is that both datasets have 512x512 resolution and I can't use the patching trick for the experiments I'm doing.
I managed to train a diffusion model for the image enhancement task, but after three days of training I get small improvements in image quality. To train the mode I used bf16 tensor format and the following settings:
# Settings
lr : 0.0002
max_epochs : 2000
num_cpu : 16
exp_name : exp1
dataset : exp1_dsT # Just override this by using -ds {name}
dataset_val : exp1_dsV # Just override this by using -ds {name}
new : True
size : 512
L2_loss : False
channels : 32
batch_size : 4
timesteps : 1000 # should be 1000
conditional : True # used only in Image2Image
image_dropout : 0.5 # Does not work with Label2Image
flip : False # used only in Image2Image reverses the predicton directon
image_mode : True
lambda_ssim : 0.0
# patch_size :
# Always recomeded
learned_variance : False
linear : False
model_name : unet
The loss is improving for the first 250k iterations but then stops decreasing.

Do you have any advice on how to tune the hyperparameters to achieve better results using high resolution images?
Metadata
Metadata
Assignees
Labels
No labels