zion
fe19b508ca
Distributed Training of the Prior ( #112 )
...
* distributed prior trainer
better EMA support
update load and save methods of prior
* update prior training script
add test evalution & ema validation
add more tracking metrics
small cleanup
2022-06-19 08:46:14 -07:00
Phil Wang
6651eafa93
one more residual, after seeing good results on unconditional generation locally
2022-06-16 11:18:02 -07:00
Phil Wang
e6bb75e5ab
fix missing residual for highest resolution of the unet
2022-06-15 20:09:43 -07:00
Giorgos Zachariadis
b4c3e5b854
changed str in order to avoid confusions and collisions with Python ( #147 )
2022-06-15 13:41:16 -07:00
Phil Wang
b7f9607258
make memory efficient unet design from imagen toggle-able
2022-06-15 13:40:26 -07:00
Phil Wang
2219348a6e
adopt similar unet architecture as imagen
2022-06-15 12:18:21 -07:00
Phil Wang
9eea9b9862
add p2 loss reweighting for decoder training as an option
2022-06-14 10:58:57 -07:00
Phil Wang
5d958713c0
fix classifier free guidance for image hiddens summed to time hiddens, thanks to @xvjiarui for finding this bug
2022-06-13 21:01:50 -07:00
Phil Wang
0f31980362
cleanup
2022-06-07 17:31:38 -07:00
Phil Wang
350a3d6045
0.6.16
2022-06-06 08:45:46 -07:00
Kashif Rasul
1a81670718
fix quadratic_beta_schedule ( #141 )
2022-06-06 08:45:14 -07:00
Phil Wang
934c9728dc
some cleanup
2022-06-04 16:54:15 -07:00
Phil Wang
ce4b0107c1
0.6.13
2022-06-04 13:26:57 -07:00
zion
64c2f9c4eb
implement ema warmup from @crowsonkb ( #140 )
2022-06-04 13:26:34 -07:00
Phil Wang
22cc613278
ema fix from @nousr
2022-06-03 19:44:36 -07:00
zion
83517849e5
ema module fixes ( #139 )
2022-06-03 19:43:51 -07:00
Phil Wang
708809ed6c
lower beta2 for adam down to 0.99, based on https://openreview.net/forum?id=2LdBqxc1Yv
2022-06-03 10:26:28 -07:00
Phil Wang
9cc475f6e7
fix update_every within EMA
2022-06-03 10:21:05 -07:00
Phil Wang
ffd342e9d0
allow for an option to constrain the variance interpolation fraction coming out from the unet for learned variance, if it is turned on
2022-06-03 09:34:57 -07:00
Phil Wang
8cc278447e
just cast to right types for blur sigma and kernel size augs
2022-06-02 11:21:58 -07:00
Phil Wang
38cd62010c
allow for random blur sigma and kernel size augmentations on low res conditioning (need to reread paper to see if the augmentation value needs to be fed into the unet for conditioning as well)
2022-06-02 11:11:25 -07:00
Ryan Russell
1cc288af39
Improve Readability ( #133 )
...
Signed-off-by: Ryan Russell <git@ryanrussell.org >
2022-06-01 13:28:02 -07:00
Phil Wang
a851168633
make youtokentome optional package, due to reported installation difficulties
2022-06-01 09:25:35 -07:00
Phil Wang
1ffeecd0ca
lower default ema beta value
2022-05-31 11:55:21 -07:00
Phil Wang
3df899f7a4
patch
2022-05-31 09:03:43 -07:00
Aidan Dempster
09534119a1
Fixed non deterministic optimizer creation ( #130 )
2022-05-31 09:03:20 -07:00
Phil Wang
6f8b90d4d7
add packaging package
2022-05-30 11:45:00 -07:00
Phil Wang
b588286288
fix version
2022-05-30 11:06:34 -07:00
Phil Wang
b693e0be03
default number of resnet blocks per layer in unet to 2 (in imagen it was 3 for base 64x64)
2022-05-30 10:06:48 -07:00
Phil Wang
a0bed30a84
additional conditioning on image embedding by summing to time embeddings (for FiLM like conditioning in subsequent layers), from passage found in paper by @mhh0318
2022-05-30 09:26:51 -07:00
zion
44d4b1bba9
overhaul prior dataloader ( #122 )
...
add readme for loader
2022-05-29 07:39:59 -07:00
Phil Wang
b8af2210df
make sure diffusion prior can be instantiated from pydantic class without clip
2022-05-26 08:47:30 -07:00
Phil Wang
f4fe6c570d
allow for full customization of number of resnet blocks per down or upsampling layers in unet, as in imagen
2022-05-26 08:33:31 -07:00
zion
1ed0f9d80b
use deterministic optimizer params ( #116 )
2022-05-25 09:31:43 -07:00
zion
d7a0a2ce4b
add more support for configuring prior ( #113 )
2022-05-25 09:06:50 -07:00
Phil Wang
f23fab7ef7
switch over to scale shift conditioning, as it seems like Imagen and Glide used it and it may be important
2022-05-24 21:46:12 -07:00
Phil Wang
857b9fbf1e
allow for one to stop grouping out weight decayable parameters, to debug optimizer state dict problem
2022-05-24 21:42:32 -07:00
Phil Wang
8864fd0aa7
bring in the dynamic thresholding technique from the Imagen paper, which purportedly improves classifier free guidance for the cascading ddpm
2022-05-24 18:15:14 -07:00
Phil Wang
fa533962bd
just use an assert to make sure clip image channels is never different than the channels of the diffusion prior and decoder, if clip is given
2022-05-22 22:43:14 -07:00
Phil Wang
276abf337b
fix and cleanup image size determination logic in decoder
2022-05-22 22:28:45 -07:00
Phil Wang
ae42d03006
allow for saving of additional fields on save method in trainers, and return loaded objects from the load method
2022-05-22 22:14:25 -07:00
Phil Wang
4d346e98d9
allow for config driven creation of clip-less diffusion prior
2022-05-22 20:36:20 -07:00
Phil Wang
5c397c9d66
move neural network creations off the configuration file into the pydantic classes
2022-05-22 19:18:18 -07:00
Phil Wang
0f4edff214
derived value for image preprocessing belongs to the data config class
2022-05-22 18:42:40 -07:00
Phil Wang
501a8c7c46
small cleanup
2022-05-22 15:39:38 -07:00
Phil Wang
49de72040c
fix decoder trainer optimizer loading (since there are multiple for each unet), also save and load step number correctly
2022-05-22 15:21:00 -07:00
Phil Wang
e527002472
take care of saving and loading functions on the diffusion prior and decoder training classes
2022-05-22 15:10:15 -07:00
Phil Wang
c12e067178
let the pydantic config base model take care of loading configuration from json path
2022-05-22 14:47:23 -07:00
Phil Wang
c6629c431a
make training splits into its own pydantic base model, validate it sums to 1, make decoder script cleaner
2022-05-22 14:43:22 -07:00
Phil Wang
a1ef023193
use pydantic to manage decoder training configs + defaults and refactor training script
2022-05-22 14:27:40 -07:00