Phil Wang
|
6651eafa93
|
one more residual, after seeing good results on unconditional generation locally
|
2022-06-16 11:18:02 -07:00 |
|
Phil Wang
|
e6bb75e5ab
|
fix missing residual for highest resolution of the unet
|
2022-06-15 20:09:43 -07:00 |
|
Giorgos Zachariadis
|
b4c3e5b854
|
changed str in order to avoid confusions and collisions with Python (#147)
|
2022-06-15 13:41:16 -07:00 |
|
Phil Wang
|
b7f9607258
|
make memory efficient unet design from imagen toggle-able
|
2022-06-15 13:40:26 -07:00 |
|
Phil Wang
|
2219348a6e
|
adopt similar unet architecture as imagen
|
2022-06-15 12:18:21 -07:00 |
|
Phil Wang
|
9eea9b9862
|
add p2 loss reweighting for decoder training as an option
|
2022-06-14 10:58:57 -07:00 |
|
Phil Wang
|
5d958713c0
|
fix classifier free guidance for image hiddens summed to time hiddens, thanks to @xvjiarui for finding this bug
|
2022-06-13 21:01:50 -07:00 |
|
Phil Wang
|
0f31980362
|
cleanup
|
2022-06-07 17:31:38 -07:00 |
|
Phil Wang
|
350a3d6045
|
0.6.16
|
2022-06-06 08:45:46 -07:00 |
|
Kashif Rasul
|
1a81670718
|
fix quadratic_beta_schedule (#141)
|
2022-06-06 08:45:14 -07:00 |
|
Phil Wang
|
934c9728dc
|
some cleanup
|
2022-06-04 16:54:15 -07:00 |
|
Phil Wang
|
ce4b0107c1
|
0.6.13
|
2022-06-04 13:26:57 -07:00 |
|
zion
|
64c2f9c4eb
|
implement ema warmup from @crowsonkb (#140)
|
2022-06-04 13:26:34 -07:00 |
|
Phil Wang
|
22cc613278
|
ema fix from @nousr
|
2022-06-03 19:44:36 -07:00 |
|
zion
|
83517849e5
|
ema module fixes (#139)
|
2022-06-03 19:43:51 -07:00 |
|
Phil Wang
|
708809ed6c
|
lower beta2 for adam down to 0.99, based on https://openreview.net/forum?id=2LdBqxc1Yv
|
2022-06-03 10:26:28 -07:00 |
|
Phil Wang
|
9cc475f6e7
|
fix update_every within EMA
|
2022-06-03 10:21:05 -07:00 |
|
Phil Wang
|
ffd342e9d0
|
allow for an option to constrain the variance interpolation fraction coming out from the unet for learned variance, if it is turned on
|
2022-06-03 09:34:57 -07:00 |
|
Phil Wang
|
8cc278447e
|
just cast to right types for blur sigma and kernel size augs
|
2022-06-02 11:21:58 -07:00 |
|
Phil Wang
|
38cd62010c
|
allow for random blur sigma and kernel size augmentations on low res conditioning (need to reread paper to see if the augmentation value needs to be fed into the unet for conditioning as well)
|
2022-06-02 11:11:25 -07:00 |
|
Ryan Russell
|
1cc288af39
|
Improve Readability (#133)
Signed-off-by: Ryan Russell <git@ryanrussell.org>
|
2022-06-01 13:28:02 -07:00 |
|
Phil Wang
|
a851168633
|
make youtokentome optional package, due to reported installation difficulties
|
2022-06-01 09:25:35 -07:00 |
|
Phil Wang
|
1ffeecd0ca
|
lower default ema beta value
|
2022-05-31 11:55:21 -07:00 |
|
Phil Wang
|
3df899f7a4
|
patch
|
2022-05-31 09:03:43 -07:00 |
|
Aidan Dempster
|
09534119a1
|
Fixed non deterministic optimizer creation (#130)
|
2022-05-31 09:03:20 -07:00 |
|
Phil Wang
|
6f8b90d4d7
|
add packaging package
|
2022-05-30 11:45:00 -07:00 |
|
Phil Wang
|
b588286288
|
fix version
|
2022-05-30 11:06:34 -07:00 |
|
Phil Wang
|
b693e0be03
|
default number of resnet blocks per layer in unet to 2 (in imagen it was 3 for base 64x64)
|
2022-05-30 10:06:48 -07:00 |
|
Phil Wang
|
a0bed30a84
|
additional conditioning on image embedding by summing to time embeddings (for FiLM like conditioning in subsequent layers), from passage found in paper by @mhh0318
|
2022-05-30 09:26:51 -07:00 |
|
zion
|
44d4b1bba9
|
overhaul prior dataloader (#122)
add readme for loader
|
2022-05-29 07:39:59 -07:00 |
|
Phil Wang
|
b8af2210df
|
make sure diffusion prior can be instantiated from pydantic class without clip
|
2022-05-26 08:47:30 -07:00 |
|
Phil Wang
|
f4fe6c570d
|
allow for full customization of number of resnet blocks per down or upsampling layers in unet, as in imagen
|
2022-05-26 08:33:31 -07:00 |
|
zion
|
1ed0f9d80b
|
use deterministic optimizer params (#116)
|
2022-05-25 09:31:43 -07:00 |
|
zion
|
d7a0a2ce4b
|
add more support for configuring prior (#113)
|
2022-05-25 09:06:50 -07:00 |
|
Phil Wang
|
f23fab7ef7
|
switch over to scale shift conditioning, as it seems like Imagen and Glide used it and it may be important
|
2022-05-24 21:46:12 -07:00 |
|
Phil Wang
|
857b9fbf1e
|
allow for one to stop grouping out weight decayable parameters, to debug optimizer state dict problem
|
2022-05-24 21:42:32 -07:00 |
|
Phil Wang
|
8864fd0aa7
|
bring in the dynamic thresholding technique from the Imagen paper, which purportedly improves classifier free guidance for the cascading ddpm
|
2022-05-24 18:15:14 -07:00 |
|
Phil Wang
|
fa533962bd
|
just use an assert to make sure clip image channels is never different than the channels of the diffusion prior and decoder, if clip is given
|
2022-05-22 22:43:14 -07:00 |
|
Phil Wang
|
276abf337b
|
fix and cleanup image size determination logic in decoder
|
2022-05-22 22:28:45 -07:00 |
|
Phil Wang
|
ae42d03006
|
allow for saving of additional fields on save method in trainers, and return loaded objects from the load method
|
2022-05-22 22:14:25 -07:00 |
|
Phil Wang
|
4d346e98d9
|
allow for config driven creation of clip-less diffusion prior
|
2022-05-22 20:36:20 -07:00 |
|
Phil Wang
|
5c397c9d66
|
move neural network creations off the configuration file into the pydantic classes
|
2022-05-22 19:18:18 -07:00 |
|
Phil Wang
|
0f4edff214
|
derived value for image preprocessing belongs to the data config class
|
2022-05-22 18:42:40 -07:00 |
|
Phil Wang
|
501a8c7c46
|
small cleanup
|
2022-05-22 15:39:38 -07:00 |
|
Phil Wang
|
49de72040c
|
fix decoder trainer optimizer loading (since there are multiple for each unet), also save and load step number correctly
|
2022-05-22 15:21:00 -07:00 |
|
Phil Wang
|
e527002472
|
take care of saving and loading functions on the diffusion prior and decoder training classes
|
2022-05-22 15:10:15 -07:00 |
|
Phil Wang
|
c12e067178
|
let the pydantic config base model take care of loading configuration from json path
|
2022-05-22 14:47:23 -07:00 |
|
Phil Wang
|
c6629c431a
|
make training splits into its own pydantic base model, validate it sums to 1, make decoder script cleaner
|
2022-05-22 14:43:22 -07:00 |
|
Phil Wang
|
a1ef023193
|
use pydantic to manage decoder training configs + defaults and refactor training script
|
2022-05-22 14:27:40 -07:00 |
|
Phil Wang
|
8b0d459b25
|
move config parsing logic to own file, consider whether to find an off-the-shelf solution at future date
|
2022-05-21 10:30:10 -07:00 |
|