Phil Wang
|
46be8c32d3
|
fix a potential issue in the low resolution conditioner, when downsampling and then upsampling using resize right, thanks to @marunine
|
2022-07-07 09:41:49 -07:00 |
|
Phil Wang
|
900f086a6d
|
fix condition_on_text_encodings in dalle2 orchestrator class, fix readme
|
2022-07-07 07:43:41 -07:00 |
|
zion
|
b3e646fd3b
|
add readme for prior (#159)
* add readme for prior
* offload prior info in main readme
* typos
|
2022-07-06 20:50:52 -07:00 |
|
Phil Wang
|
6d477d7654
|
link to dalle2 laion
|
2022-07-05 11:43:07 -07:00 |
|
Phil Wang
|
531fe4b62f
|
status
|
2022-07-05 10:46:55 -07:00 |
|
Phil Wang
|
a922a539de
|
bring back convtranspose2d upsampling, allow for nearest upsample with hyperparam, change kernel size of last conv to 1, make configurable, cleanup
|
2022-07-01 09:21:47 -07:00 |
|
Phil Wang
|
6a11b9678b
|
bring in the skip connection scaling factor, used by imagen in their unets, cite original paper using it
|
2022-06-26 21:59:55 -07:00 |
|
Phil Wang
|
b90364695d
|
fix remaining issues with deriving cond_on_text_encodings from child unet settings
|
2022-06-26 21:07:42 -07:00 |
|
Phil Wang
|
a5b9fd6ca8
|
product management
|
2022-06-24 08:15:05 -07:00 |
|
Phil Wang
|
56883910fb
|
cleanup
|
2022-06-20 11:14:55 -07:00 |
|
Phil Wang
|
893f270012
|
project management
|
2022-06-20 10:00:22 -07:00 |
|
Phil Wang
|
0215237fc6
|
update status
|
2022-06-19 09:42:24 -07:00 |
|
Phil Wang
|
41ca896413
|
depend on huggingface accelerate, move appreciation thread up for visibility
|
2022-06-19 08:50:35 -07:00 |
|
Phil Wang
|
9eea9b9862
|
add p2 loss reweighting for decoder training as an option
|
2022-06-14 10:58:57 -07:00 |
|
Ryan Russell
|
1cc288af39
|
Improve Readability (#133)
Signed-off-by: Ryan Russell <git@ryanrussell.org>
|
2022-06-01 13:28:02 -07:00 |
|
Phil Wang
|
f12a7589c5
|
commit to trying out grid attention
|
2022-05-26 12:56:10 -07:00 |
|
Phil Wang
|
645e207441
|
credit assignment
|
2022-05-26 08:16:03 -07:00 |
|
Phil Wang
|
00743b3a0b
|
update
|
2022-05-26 08:12:25 -07:00 |
|
Phil Wang
|
01589aff6a
|
cite maxvit properly
|
2022-05-26 07:12:25 -07:00 |
|
Phil Wang
|
8864fd0aa7
|
bring in the dynamic thresholding technique from the Imagen paper, which purportedly improves classifier free guidance for the cascading ddpm
|
2022-05-24 18:15:14 -07:00 |
|
Phil Wang
|
72bf159331
|
update
|
2022-05-24 08:25:40 -07:00 |
|
Phil Wang
|
e5e47cfecb
|
link to aidan's test run
|
2022-05-23 12:41:46 -07:00 |
|
Phil Wang
|
4d346e98d9
|
allow for config driven creation of clip-less diffusion prior
|
2022-05-22 20:36:20 -07:00 |
|
Phil Wang
|
2b1fd1ad2e
|
product management
|
2022-05-22 19:23:40 -07:00 |
|
zion
|
82a2ef37d9
|
Update README.md (#109)
block in a section that links to available pre-trained models for those who are interested
|
2022-05-22 19:22:30 -07:00 |
|
Phil Wang
|
4e49373fc5
|
project management
|
2022-05-22 15:27:40 -07:00 |
|
Phil Wang
|
e527002472
|
take care of saving and loading functions on the diffusion prior and decoder training classes
|
2022-05-22 15:10:15 -07:00 |
|
Phil Wang
|
a1ef023193
|
use pydantic to manage decoder training configs + defaults and refactor training script
|
2022-05-22 14:27:40 -07:00 |
|
Phil Wang
|
8aab69b91e
|
final thought
|
2022-05-21 10:47:45 -07:00 |
|
Phil Wang
|
ebaa0d28c2
|
product management
|
2022-05-21 10:30:52 -07:00 |
|
Phil Wang
|
b895f52843
|
appreciation section
|
2022-05-21 08:32:12 -07:00 |
|
Phil Wang
|
bb86ab2404
|
update sample, and set default gradient clipping value for decoder training
|
2022-05-16 17:38:30 -07:00 |
|
Phil Wang
|
ae056dd67c
|
samples
|
2022-05-16 13:46:35 -07:00 |
|
Phil Wang
|
033d6b0ce8
|
last update
|
2022-05-16 13:38:33 -07:00 |
|
Phil Wang
|
13382885d9
|
final update to dalle2 repository for a while - sampling from prior in chunks automatically with max_batch_size keyword given
|
2022-05-16 12:57:31 -07:00 |
|
Phil Wang
|
c3d4a7ffe4
|
update working unconditional decoder example
|
2022-05-16 12:50:07 -07:00 |
|
Phil Wang
|
5562ec6be2
|
status updates
|
2022-05-16 12:01:54 -07:00 |
|
Phil Wang
|
fad7481479
|
todo
|
2022-05-15 17:00:25 -07:00 |
|
Phil Wang
|
123658d082
|
cite Ho et al, since cascading ddpm is now trainable
|
2022-05-15 16:56:53 -07:00 |
|
Phil Wang
|
11d4e11f10
|
allow for training unconditional ddpm or cascading ddpms
|
2022-05-15 16:54:56 -07:00 |
|
Phil Wang
|
0f0011caf0
|
todo
|
2022-05-15 14:28:35 -07:00 |
|
Phil Wang
|
5ec34bebe1
|
cleanup readme
|
2022-05-15 12:29:26 -07:00 |
|
Phil Wang
|
f7cd4a0992
|
product management
|
2022-05-15 11:21:12 -07:00 |
|
Phil Wang
|
68e7d2f241
|
make sure gradient accumulation feature works even if all arguments passed in are keyword arguments
|
2022-05-15 11:16:16 -07:00 |
|
Phil Wang
|
89de5af63e
|
experiment tracker agnostic
|
2022-05-15 09:56:40 -07:00 |
|
Phil Wang
|
b0cd5f24b6
|
take care of gradient accumulation automatically for researchers, by passing in a max_batch_size on the decoder or diffusion prior trainer forward
|
2022-05-14 17:04:09 -07:00 |
|
Phil Wang
|
b494ed81d4
|
take care of backwards within trainer classes for diffusion prior and decoder, readying to take care of gradient accumulation as well (plus, unsure if loss should be backwards within autocast block)
|
2022-05-14 15:49:24 -07:00 |
|
Phil Wang
|
ff3474f05c
|
normalize conditioning tokens outside of cross attention blocks
|
2022-05-14 14:23:52 -07:00 |
|
Light-V
|
6f76652d11
|
fix typo in README.md (#85)
The default config for clip from openai should be ViT-B/32
|
2022-05-11 13:38:16 -07:00 |
|
Phil Wang
|
908088cfea
|
wrap up cross embed layer feature
|
2022-05-10 12:19:34 -07:00 |
|