Phil Wang
|
e76e89f9eb
|
remove text masking altogether in favor of deriving from text encodings (padded text encodings must be pad value of 0.)
|
2022-07-12 15:40:31 -07:00 |
|
Phil Wang
|
e055793e5d
|
shoutout for @MalumaDev
|
2022-07-11 16:12:35 -07:00 |
|
Phil Wang
|
4173e88121
|
more accurate readme
|
2022-07-09 20:57:26 -07:00 |
|
Phil Wang
|
b7e22f7da0
|
complete ddim integration of diffusion prior as well as decoder for each unet, feature complete for https://github.com/lucidrains/DALLE2-pytorch/issues/157
|
2022-07-09 17:25:34 -07:00 |
|
Phil Wang
|
46be8c32d3
|
fix a potential issue in the low resolution conditioner, when downsampling and then upsampling using resize right, thanks to @marunine
|
2022-07-07 09:41:49 -07:00 |
|
Phil Wang
|
900f086a6d
|
fix condition_on_text_encodings in dalle2 orchestrator class, fix readme
|
2022-07-07 07:43:41 -07:00 |
|
zion
|
b3e646fd3b
|
add readme for prior (#159)
* add readme for prior
* offload prior info in main readme
* typos
|
2022-07-06 20:50:52 -07:00 |
|
Phil Wang
|
6d477d7654
|
link to dalle2 laion
|
2022-07-05 11:43:07 -07:00 |
|
Phil Wang
|
531fe4b62f
|
status
|
2022-07-05 10:46:55 -07:00 |
|
Phil Wang
|
a922a539de
|
bring back convtranspose2d upsampling, allow for nearest upsample with hyperparam, change kernel size of last conv to 1, make configurable, cleanup
|
2022-07-01 09:21:47 -07:00 |
|
Phil Wang
|
6a11b9678b
|
bring in the skip connection scaling factor, used by imagen in their unets, cite original paper using it
|
2022-06-26 21:59:55 -07:00 |
|
Phil Wang
|
b90364695d
|
fix remaining issues with deriving cond_on_text_encodings from child unet settings
|
2022-06-26 21:07:42 -07:00 |
|
Phil Wang
|
a5b9fd6ca8
|
product management
|
2022-06-24 08:15:05 -07:00 |
|
Phil Wang
|
56883910fb
|
cleanup
|
2022-06-20 11:14:55 -07:00 |
|
Phil Wang
|
893f270012
|
project management
|
2022-06-20 10:00:22 -07:00 |
|
Phil Wang
|
0215237fc6
|
update status
|
2022-06-19 09:42:24 -07:00 |
|
Phil Wang
|
41ca896413
|
depend on huggingface accelerate, move appreciation thread up for visibility
|
2022-06-19 08:50:35 -07:00 |
|
Phil Wang
|
9eea9b9862
|
add p2 loss reweighting for decoder training as an option
|
2022-06-14 10:58:57 -07:00 |
|
Ryan Russell
|
1cc288af39
|
Improve Readability (#133)
Signed-off-by: Ryan Russell <git@ryanrussell.org>
|
2022-06-01 13:28:02 -07:00 |
|
Phil Wang
|
f12a7589c5
|
commit to trying out grid attention
|
2022-05-26 12:56:10 -07:00 |
|
Phil Wang
|
645e207441
|
credit assignment
|
2022-05-26 08:16:03 -07:00 |
|
Phil Wang
|
00743b3a0b
|
update
|
2022-05-26 08:12:25 -07:00 |
|
Phil Wang
|
01589aff6a
|
cite maxvit properly
|
2022-05-26 07:12:25 -07:00 |
|
Phil Wang
|
8864fd0aa7
|
bring in the dynamic thresholding technique from the Imagen paper, which purportedly improves classifier free guidance for the cascading ddpm
|
2022-05-24 18:15:14 -07:00 |
|
Phil Wang
|
72bf159331
|
update
|
2022-05-24 08:25:40 -07:00 |
|
Phil Wang
|
e5e47cfecb
|
link to aidan's test run
|
2022-05-23 12:41:46 -07:00 |
|
Phil Wang
|
4d346e98d9
|
allow for config driven creation of clip-less diffusion prior
|
2022-05-22 20:36:20 -07:00 |
|
Phil Wang
|
2b1fd1ad2e
|
product management
|
2022-05-22 19:23:40 -07:00 |
|
zion
|
82a2ef37d9
|
Update README.md (#109)
block in a section that links to available pre-trained models for those who are interested
|
2022-05-22 19:22:30 -07:00 |
|
Phil Wang
|
4e49373fc5
|
project management
|
2022-05-22 15:27:40 -07:00 |
|
Phil Wang
|
e527002472
|
take care of saving and loading functions on the diffusion prior and decoder training classes
|
2022-05-22 15:10:15 -07:00 |
|
Phil Wang
|
a1ef023193
|
use pydantic to manage decoder training configs + defaults and refactor training script
|
2022-05-22 14:27:40 -07:00 |
|
Phil Wang
|
8aab69b91e
|
final thought
|
2022-05-21 10:47:45 -07:00 |
|
Phil Wang
|
ebaa0d28c2
|
product management
|
2022-05-21 10:30:52 -07:00 |
|
Phil Wang
|
b895f52843
|
appreciation section
|
2022-05-21 08:32:12 -07:00 |
|
Phil Wang
|
bb86ab2404
|
update sample, and set default gradient clipping value for decoder training
|
2022-05-16 17:38:30 -07:00 |
|
Phil Wang
|
ae056dd67c
|
samples
|
2022-05-16 13:46:35 -07:00 |
|
Phil Wang
|
033d6b0ce8
|
last update
|
2022-05-16 13:38:33 -07:00 |
|
Phil Wang
|
13382885d9
|
final update to dalle2 repository for a while - sampling from prior in chunks automatically with max_batch_size keyword given
|
2022-05-16 12:57:31 -07:00 |
|
Phil Wang
|
c3d4a7ffe4
|
update working unconditional decoder example
|
2022-05-16 12:50:07 -07:00 |
|
Phil Wang
|
5562ec6be2
|
status updates
|
2022-05-16 12:01:54 -07:00 |
|
Phil Wang
|
fad7481479
|
todo
|
2022-05-15 17:00:25 -07:00 |
|
Phil Wang
|
123658d082
|
cite Ho et al, since cascading ddpm is now trainable
|
2022-05-15 16:56:53 -07:00 |
|
Phil Wang
|
11d4e11f10
|
allow for training unconditional ddpm or cascading ddpms
|
2022-05-15 16:54:56 -07:00 |
|
Phil Wang
|
0f0011caf0
|
todo
|
2022-05-15 14:28:35 -07:00 |
|
Phil Wang
|
5ec34bebe1
|
cleanup readme
|
2022-05-15 12:29:26 -07:00 |
|
Phil Wang
|
f7cd4a0992
|
product management
|
2022-05-15 11:21:12 -07:00 |
|
Phil Wang
|
68e7d2f241
|
make sure gradient accumulation feature works even if all arguments passed in are keyword arguments
|
2022-05-15 11:16:16 -07:00 |
|
Phil Wang
|
89de5af63e
|
experiment tracker agnostic
|
2022-05-15 09:56:40 -07:00 |
|
Phil Wang
|
b0cd5f24b6
|
take care of gradient accumulation automatically for researchers, by passing in a max_batch_size on the decoder or diffusion prior trainer forward
|
2022-05-14 17:04:09 -07:00 |
|