Phil Wang
|
e527002472
|
take care of saving and loading functions on the diffusion prior and decoder training classes
|
2022-05-22 15:10:15 -07:00 |
|
Phil Wang
|
bb86ab2404
|
update sample, and set default gradient clipping value for decoder training
|
2022-05-16 17:38:30 -07:00 |
|
Phil Wang
|
c7ea8748db
|
default decoder learning rate to what was in the paper
|
2022-05-16 13:33:54 -07:00 |
|
Phil Wang
|
13382885d9
|
final update to dalle2 repository for a while - sampling from prior in chunks automatically with max_batch_size keyword given
|
2022-05-16 12:57:31 -07:00 |
|
Phil Wang
|
164d9be444
|
use a decorator and take care of sampling in chunks (max_batch_size keyword), in case one is sampling a huge grid of images
|
2022-05-16 12:34:28 -07:00 |
|
Phil Wang
|
89ff04cfe2
|
final tweak to EMA class
|
2022-05-16 11:54:34 -07:00 |
|
Phil Wang
|
f4016f6302
|
allow for overriding use of EMA during sampling in decoder trainer with use_non_ema keyword, also fix some issues with automatic normalization of images and low res conditioning image if latent diffusion is in play
|
2022-05-16 11:18:30 -07:00 |
|
Phil Wang
|
dab106d4e5
|
back to no_grad for now, also keep track and restore unet devices in one_unet_in_gpu contextmanager
|
2022-05-16 09:36:14 -07:00 |
|
Phil Wang
|
bb151ca6b1
|
unet_number on decoder trainer only needs to be passed in if there is greater than 1 unet, so that unconditional training of a single ddpm is seamless (experiment in progress locally)
|
2022-05-16 09:17:17 -07:00 |
|