Phil Wang
|
f12a7589c5
|
commit to trying out grid attention
|
2022-05-26 12:56:10 -07:00 |
|
Phil Wang
|
645e207441
|
credit assignment
|
2022-05-26 08:16:03 -07:00 |
|
Phil Wang
|
00743b3a0b
|
update
|
2022-05-26 08:12:25 -07:00 |
|
Phil Wang
|
01589aff6a
|
cite maxvit properly
|
2022-05-26 07:12:25 -07:00 |
|
Phil Wang
|
8864fd0aa7
|
bring in the dynamic thresholding technique from the Imagen paper, which purportedly improves classifier free guidance for the cascading ddpm
|
2022-05-24 18:15:14 -07:00 |
|
Phil Wang
|
72bf159331
|
update
|
2022-05-24 08:25:40 -07:00 |
|
Phil Wang
|
e5e47cfecb
|
link to aidan's test run
|
2022-05-23 12:41:46 -07:00 |
|
Phil Wang
|
4d346e98d9
|
allow for config driven creation of clip-less diffusion prior
|
2022-05-22 20:36:20 -07:00 |
|
Phil Wang
|
2b1fd1ad2e
|
product management
|
2022-05-22 19:23:40 -07:00 |
|
zion
|
82a2ef37d9
|
Update README.md (#109)
block in a section that links to available pre-trained models for those who are interested
|
2022-05-22 19:22:30 -07:00 |
|
Phil Wang
|
4e49373fc5
|
project management
|
2022-05-22 15:27:40 -07:00 |
|
Phil Wang
|
e527002472
|
take care of saving and loading functions on the diffusion prior and decoder training classes
|
2022-05-22 15:10:15 -07:00 |
|
Phil Wang
|
a1ef023193
|
use pydantic to manage decoder training configs + defaults and refactor training script
|
2022-05-22 14:27:40 -07:00 |
|
Phil Wang
|
8aab69b91e
|
final thought
|
2022-05-21 10:47:45 -07:00 |
|
Phil Wang
|
ebaa0d28c2
|
product management
|
2022-05-21 10:30:52 -07:00 |
|
Phil Wang
|
b895f52843
|
appreciation section
|
2022-05-21 08:32:12 -07:00 |
|
Phil Wang
|
bb86ab2404
|
update sample, and set default gradient clipping value for decoder training
|
2022-05-16 17:38:30 -07:00 |
|
Phil Wang
|
ae056dd67c
|
samples
|
2022-05-16 13:46:35 -07:00 |
|
Phil Wang
|
033d6b0ce8
|
last update
|
2022-05-16 13:38:33 -07:00 |
|
Phil Wang
|
13382885d9
|
final update to dalle2 repository for a while - sampling from prior in chunks automatically with max_batch_size keyword given
|
2022-05-16 12:57:31 -07:00 |
|
Phil Wang
|
c3d4a7ffe4
|
update working unconditional decoder example
|
2022-05-16 12:50:07 -07:00 |
|
Phil Wang
|
5562ec6be2
|
status updates
|
2022-05-16 12:01:54 -07:00 |
|
Phil Wang
|
fad7481479
|
todo
|
2022-05-15 17:00:25 -07:00 |
|
Phil Wang
|
123658d082
|
cite Ho et al, since cascading ddpm is now trainable
|
2022-05-15 16:56:53 -07:00 |
|
Phil Wang
|
11d4e11f10
|
allow for training unconditional ddpm or cascading ddpms
|
2022-05-15 16:54:56 -07:00 |
|
Phil Wang
|
0f0011caf0
|
todo
|
2022-05-15 14:28:35 -07:00 |
|
Phil Wang
|
5ec34bebe1
|
cleanup readme
|
2022-05-15 12:29:26 -07:00 |
|
Phil Wang
|
f7cd4a0992
|
product management
|
2022-05-15 11:21:12 -07:00 |
|
Phil Wang
|
68e7d2f241
|
make sure gradient accumulation feature works even if all arguments passed in are keyword arguments
|
2022-05-15 11:16:16 -07:00 |
|
Phil Wang
|
89de5af63e
|
experiment tracker agnostic
|
2022-05-15 09:56:40 -07:00 |
|
Phil Wang
|
b0cd5f24b6
|
take care of gradient accumulation automatically for researchers, by passing in a max_batch_size on the decoder or diffusion prior trainer forward
|
2022-05-14 17:04:09 -07:00 |
|
Phil Wang
|
b494ed81d4
|
take care of backwards within trainer classes for diffusion prior and decoder, readying to take care of gradient accumulation as well (plus, unsure if loss should be backwards within autocast block)
|
2022-05-14 15:49:24 -07:00 |
|
Phil Wang
|
ff3474f05c
|
normalize conditioning tokens outside of cross attention blocks
|
2022-05-14 14:23:52 -07:00 |
|
Light-V
|
6f76652d11
|
fix typo in README.md (#85)
The default config for clip from openai should be ViT-B/32
|
2022-05-11 13:38:16 -07:00 |
|
Phil Wang
|
908088cfea
|
wrap up cross embed layer feature
|
2022-05-10 12:19:34 -07:00 |
|
Phil Wang
|
8dc8a3de0d
|
product management
|
2022-05-10 11:51:38 -07:00 |
|
Phil Wang
|
35f89556ba
|
bring in the cross embed layer from Crossformer paper for initial convolution in unet
|
2022-05-10 11:50:38 -07:00 |
|
Phil Wang
|
fc8fce38fb
|
make sure cascading DDPM can be trained unconditionally, to ready for CLI one command training for the public
|
2022-05-10 10:48:10 -07:00 |
|
Phil Wang
|
a1bfb03ba4
|
project management
|
2022-05-10 10:13:51 -07:00 |
|
Phil Wang
|
b1e7b5f6bb
|
make sure resnet groups in unet is finely customizable
|
2022-05-10 10:12:50 -07:00 |
|
Phil Wang
|
64f7be1926
|
some cleanup
|
2022-05-09 16:50:21 -07:00 |
|
Kumar R
|
8647cb5e76
|
Val loss changes, with quite a few other changes. This is in place of the earlier PR(https://github.com/lucidrains/DALLE2-pytorch/pull/67) (#77)
* Val_loss changes - no rebased with lucidrains' master.
* Val Loss changes - now rebased with lucidrains' master
* train_diffusion_prior.py updates
* dalle2_pytorch.py updates
* __init__.py changes
* Update train_diffusion_prior.py
* Update dalle2_pytorch.py
* Update train_diffusion_prior.py
* Update train_diffusion_prior.py
* Update dalle2_pytorch.py
* Update train_diffusion_prior.py
* Update train_diffusion_prior.py
* Update train_diffusion_prior.py
* Update train_diffusion_prior.py
* Update README.md
* Update README.md
* Update README.md
* Update README.md
* Update README.md
* Update README.md
* Update README.md
* Update README.md
* Update README.md
|
2022-05-09 08:53:29 -07:00 |
|
Phil Wang
|
53c189e46a
|
give more surface area for attention in diffusion prior
|
2022-05-09 08:08:11 -07:00 |
|
Phil Wang
|
c87b84a259
|
todo
|
2022-05-07 09:21:08 -07:00 |
|
Phil Wang
|
8b05468653
|
todo
|
2022-05-07 08:33:45 -07:00 |
|
Piero Rolando
|
fd53fa17db
|
Fix a typo in README (#70)
Change "pyhon" for "python" (correct)
|
2022-05-06 16:53:36 -07:00 |
|
Phil Wang
|
09e9eaa5a6
|
project management
|
2022-05-06 09:00:22 -07:00 |
|
Phil Wang
|
e6d752cf4a
|
reprioritize
|
2022-05-06 08:55:26 -07:00 |
|
Phil Wang
|
0be1e0d64c
|
support CoCa, which seems to be better than CLIP (has an autoregressive text encoder) https://arxiv.org/abs/2205.01917
|
2022-05-06 08:27:12 -07:00 |
|
Phil Wang
|
98df1ba51e
|
add diffusion prior trainer, which automatically takes care of the exponential moving average (training and sampling), as well as mixed precision, gradient clipping
|
2022-05-06 08:11:09 -07:00 |
|