Phil Wang
|
0d82dff9c5
|
in ddim, noise should be predicted after x0 is maybe clipped, thanks to @lukovnikov for pointing this out in another repository
|
2022-09-01 09:40:47 -07:00 |
|
Phil Wang
|
8bbc956ff1
|
fix bug with misnamed variable in diffusion prior network
|
2022-08-31 17:19:05 -07:00 |
|
Phil Wang
|
6fb7e91343
|
fix ddim to use alpha_cumprod
|
2022-08-31 07:40:46 -07:00 |
|
Phil Wang
|
ba58ae0bf2
|
add two asserts to diffusion prior to ensure matching image embedding dimensions for clip, diffusion prior network, and what was set on diffusion prior
|
2022-08-28 10:11:37 -07:00 |
|
Phil Wang
|
1cc5d0afa7
|
upgrade to best downsample
|
2022-08-25 10:37:02 -07:00 |
|
Phil Wang
|
59fa101c4d
|
fix classifier free guidance for diffusion prior, thanks to @jaykim9870 for spotting the issue
|
2022-08-23 08:29:01 -07:00 |
|
Phil Wang
|
083508ff8e
|
cast attention matrix back to original dtype pre-softmax in attention
|
2022-08-20 10:56:01 -07:00 |
|
Phil Wang
|
7762edd0ff
|
make it work for @ethancohen123
|
2022-08-19 11:28:58 -07:00 |
|
Phil Wang
|
27f19ba7fa
|
make sure diffusion prior trainer can operate with no warmup
|
2022-08-15 14:27:40 -07:00 |
|
Phil Wang
|
8f38339c2b
|
give diffusion prior trainer cosine annealing lr too
|
2022-08-15 07:38:01 -07:00 |
|
Phil Wang
|
6b9b4b9e5e
|
add cosine annealing lr schedule
|
2022-08-15 07:29:56 -07:00 |
|
Phil Wang
|
44e09d5a4d
|
add weight standardization behind feature flag, which may potentially work well with group norm
|
2022-08-14 11:34:45 -07:00 |
|
Phil Wang
|
34806663e3
|
make it so diffusion prior p_sample_loop returns unnormalized image embeddings
|
2022-08-13 10:03:40 -07:00 |
|
Phil Wang
|
dc816b1b6e
|
dry up some code around handling unet outputs with learned variance
|
2022-08-12 15:25:03 -07:00 |
|
Phil Wang
|
05192ffac4
|
fix self conditioning shape in diffusion prior
|
2022-08-12 12:30:03 -07:00 |
|
Phil Wang
|
9440411954
|
make self conditioning technique work with diffusion prior
|
2022-08-12 12:20:51 -07:00 |
|
Phil Wang
|
7c5477b26d
|
bet on the new self-conditioning technique out of geoffrey hintons group
|
2022-08-12 11:36:08 -07:00 |
|
Phil Wang
|
be3bb868bf
|
add gradient checkpointing for all resnet blocks
|
2022-08-02 19:21:44 -07:00 |
|
Phil Wang
|
451de34871
|
enforce clip anytorch version
|
2022-07-30 10:07:55 -07:00 |
|
Phil Wang
|
f22e8c8741
|
make open clip available for use with dalle2 pytorch
|
2022-07-30 09:02:31 -07:00 |
|
Phil Wang
|
87432e93ad
|
quick fix for linear attention
|
2022-07-29 13:17:12 -07:00 |
|
Phil Wang
|
d167378401
|
add cosine sim for self attention as well, as a setting
|
2022-07-29 12:48:20 -07:00 |
|
Phil Wang
|
2d67d5821e
|
change up epsilon in layernorm the case of using fp16, thanks to @Veldrovive for figuring out this stabilizes training
|
2022-07-29 12:41:02 -07:00 |
|
Phil Wang
|
748c7fe7af
|
allow for cosine sim cross attention, modify linear attention in attempt to resolve issue on fp16
|
2022-07-29 11:12:18 -07:00 |
|
Phil Wang
|
80046334ad
|
make sure entire readme runs without errors
|
2022-07-28 10:17:43 -07:00 |
|
Phil Wang
|
36fb46a95e
|
fix readme and a small bug in DALLE2 class
|
2022-07-28 08:33:51 -07:00 |
|
Phil Wang
|
07abfcf45b
|
rescale values in linear attention to mitigate overflows in fp16 setting
|
2022-07-27 12:27:38 -07:00 |
|
Phil Wang
|
406e75043f
|
add upsample combiner feature for the unets
|
2022-07-26 10:46:04 -07:00 |
|
Phil Wang
|
9646dfc0e6
|
fix path_or_state bug
|
2022-07-26 09:47:54 -07:00 |
|
Phil Wang
|
62043acb2f
|
fix repaint
|
2022-07-24 15:29:06 -07:00 |
|
Phil Wang
|
417ff808e6
|
1.0.3
|
2022-07-22 13:16:57 -07:00 |
|
Phil Wang
|
48a1302428
|
1.0.2
|
2022-07-20 23:01:51 -07:00 |
|
Phil Wang
|
76d08498cc
|
diffusion prior training updates from @nousr
|
2022-07-20 18:05:27 -07:00 |
|
Phil Wang
|
06c65b60d2
|
1.0.0
|
2022-07-19 19:08:17 -07:00 |
|
Phil Wang
|
4b912a38c6
|
0.26.2
|
2022-07-19 17:50:36 -07:00 |
|
Phil Wang
|
291377bb9c
|
@jacobwjs reports dynamic thresholding works very well and 0.95 is a better value
|
2022-07-19 11:31:56 -07:00 |
|
Phil Wang
|
723bf0abba
|
complete inpainting ability using inpaint_image and inpaint_mask passed into sample function for decoder
|
2022-07-19 09:26:55 -07:00 |
|
Phil Wang
|
d88c7ba56c
|
fix a bug with ddim and predict x0 objective
|
2022-07-18 19:04:26 -07:00 |
|
Phil Wang
|
da8e99ada0
|
fix sample bug
|
2022-07-18 13:50:22 -07:00 |
|
Phil Wang
|
6afb886cf4
|
complete imagen-like noise level conditioning
|
2022-07-18 13:43:57 -07:00 |
|
Phil Wang
|
a2ee3fa3cc
|
offer way to turn off initial cross embed convolutional module, for debugging upsampler artifacts
|
2022-07-15 17:29:10 -07:00 |
|
Phil Wang
|
a58a370d75
|
takes care of a grad strides error at https://github.com/lucidrains/DALLE2-pytorch/issues/196 thanks to @YUHANG-Ma
|
2022-07-14 15:28:34 -07:00 |
|
Phil Wang
|
1662bbf226
|
protect against random cropping for base unet
|
2022-07-14 12:49:43 -07:00 |
|
Phil Wang
|
a34f60962a
|
let the neural network peek at the low resolution conditioning one last time before making prediction, for upsamplers
|
2022-07-14 10:27:04 -07:00 |
|
Phil Wang
|
0b40cbaa54
|
just always use nearest neighbor interpolation when resizing for low resolution conditioning, for https://github.com/lucidrains/DALLE2-pytorch/pull/181
|
2022-07-13 20:59:43 -07:00 |
|
Phil Wang
|
f141144a6d
|
allow for using classifier free guidance for some unets but not others, by passing in a tuple of cond_scale during sampling for decoder, just in case it is causing issues for upsamplers
|
2022-07-13 13:12:30 -07:00 |
|
Phil Wang
|
f988207718
|
hack around some inplace error, also make sure for openai clip text encoding, only tokens after eos_id is masked out
|
2022-07-13 12:56:02 -07:00 |
|
Phil Wang
|
b2073219f0
|
foolproof sampling for decoder to always use eval mode (and restore training state afterwards)
|
2022-07-13 10:21:00 -07:00 |
|
Phil Wang
|
cc0f7a935c
|
fix non pixel shuffle upsample
|
2022-07-13 10:16:02 -07:00 |
|
Phil Wang
|
95a512cb65
|
fix a potential bug with conditioning with blurred low resolution image, blur should be applied only 50% of the time
|
2022-07-13 10:11:49 -07:00 |
|