Aidan Dempster
15acc03bd4
Add a dataloader for training the decoder ( #57 )
...
* Added dataloader and updated requirements
* Added option to set embedding shard width separately from webdataset shard length.
There must be a better way to do this.
* Changed embedding loader to read using fsspec
* Moved the loader into a more compatible location
* Removed unnecessary package
* Fixed typo (Embeding -> Embedding)
* Simplified example embedding finder code to remove unnecessary get_file_list function
* Added example usage of ImageEmbeddingDataset
* Changed the name of create_dataloader to be more verbose
Added a dataloaders __init__.py
2022-05-05 07:08:45 -07:00
Phil Wang
896f19786d
remove convnext blocks, they are illsuited for generative work, validated by early experimental results at https://github.com/lucidrains/video-diffusion-pytorch
0.0.101
2022-05-05 07:07:21 -07:00
Phil Wang
aec5575d09
take a bet on resize right, given Katherine is using it
0.0.100
2022-05-04 19:26:45 -07:00
Phil Wang
9773f10d6c
use inference mode whenever possible, cleanup
0.0.99
2022-05-04 15:25:05 -07:00
Phil Wang
a6bf8ddef6
advertise laion
2022-05-04 15:04:05 -07:00
Phil Wang
86e692d24f
fix random crop probability
0.0.98
2022-05-04 11:52:24 -07:00
Phil Wang
97b751209f
allow for last unet in the cascade to be trained on crops, if it is convolution-only
0.0.97
2022-05-04 11:48:48 -07:00
Phil Wang
74103fd8d6
product management
2022-05-04 11:20:50 -07:00
Phil Wang
1992d25cad
project management
0.0.96
2022-05-04 11:18:54 -07:00
Phil Wang
5b619c2fd5
make sure some hyperparameters for unet block is configurable
2022-05-04 11:18:32 -07:00
Phil Wang
9359ad2e91
0.0.95
0.0.95
2022-05-04 10:53:05 -07:00
Phil Wang
9ff228188b
offer old resnet blocks, from the original DDPM paper, just in case convnexts are unsuitable for generative work
2022-05-04 10:52:58 -07:00
Kumar R
2d9963d30e
Reporting metrics - Cosine similarity. ( #55 )
...
* Update train_diffusion_prior.py
* Delete train_diffusion_prior.py
* Cosine similarity logging.
* Update train_diffusion_prior.py
* Report Cosine metrics every N steps.
2022-05-04 08:04:36 -07:00
Phil Wang
58d9b422f3
0.0.94
0.0.94
2022-05-04 07:42:33 -07:00
Ray Bell
44b319cb57
add missing import ( #56 )
2022-05-04 07:42:20 -07:00
Phil Wang
c30f380689
final reminder
2022-05-03 08:18:53 -07:00
Phil Wang
e4e884bb8b
keep all doors open
2022-05-03 08:17:02 -07:00
Phil Wang
803ad9c17d
product management again
2022-05-03 08:15:25 -07:00
Phil Wang
a88dd6a9c0
todo
2022-05-03 08:09:02 -07:00
Kumar R
72c16b496e
Update train_diffusion_prior.py ( #53 )
2022-05-02 22:44:57 -07:00
z
81d83dd7f2
defaults align with paper ( #52 )
...
Co-authored-by: nousr <>
2022-05-02 13:52:11 -07:00
Phil Wang
fa66f7e1e9
todo
2022-05-02 12:57:15 -07:00
Phil Wang
aa8d135245
allow laion to experiment with normformer in diffusion prior
2022-05-02 11:35:00 -07:00
Phil Wang
70282de23b
add ability to turn on normformer settings, given @borisdayma reported good results and some personal anecdata
0.0.93
2022-05-02 11:33:15 -07:00
Phil Wang
83f761847e
todo
2022-05-02 10:52:39 -07:00
Phil Wang
11469dc0c6
makes more sense to keep this as True as default, for stability
0.0.92
2022-05-02 10:50:55 -07:00
Romain Beaumont
2d25c89f35
Fix passing of l2norm_output to DiffusionPriorNetwork ( #51 )
2022-05-02 10:48:16 -07:00
Phil Wang
3fe96c208a
add ability to train diffusion prior with l2norm on output image embed
2022-05-02 09:53:20 -07:00
Phil Wang
0fc6c9cdf3
provide option to l2norm the output of the diffusion prior
0.0.91
2022-05-02 09:41:03 -07:00
Phil Wang
7ee0ecc388
mixed precision for training diffusion prior + save optimizer and scaler states
2022-05-02 09:31:04 -07:00
Phil Wang
1924c7cc3d
fix issue with mixed precision and gradient clipping
0.0.90
2022-05-02 09:20:19 -07:00
Phil Wang
f7df3caaf3
address not calculating average eval / test loss when training diffusion prior https://github.com/lucidrains/DALLE2-pytorch/issues/49
2022-05-02 08:51:41 -07:00
Phil Wang
fc954ee788
fix calculation of adaptive weight for vit-vqgan, thanks to @CiaoHe
0.0.89
2022-05-02 07:58:14 -07:00
Phil Wang
c1db2753f5
todo
2022-05-01 18:02:30 -07:00
Phil Wang
ad87bfe28f
switch to using linear attention for the sparse attention layers within unet, given success in GAN projects
0.0.88
2022-05-01 17:59:03 -07:00
Phil Wang
76c767b1ce
update deps, commit to using webdatasets, per @rom1504 consultation
2022-05-01 12:22:15 -07:00
Phil Wang
d991b8c39c
just clip the diffusion prior network parameters
2022-05-01 12:01:08 -07:00
Phil Wang
902693e271
todo
2022-05-01 11:57:08 -07:00
Phil Wang
35cd63982d
add gradient clipping, make sure weight decay is configurable, make sure learning rate is actually passed into get_optimizer, make sure model is set to training mode at beginning of each epoch
2022-05-01 11:55:38 -07:00
Kumar R
53ce6dfdf6
All changes implemented, current run happening. Link to wandb run in comments. ( #43 )
...
* Train DiffusionPrior with pre-computed embeddings
This is in response to https://github.com/lucidrains/DALLE2-pytorch/issues/29 - more metrics will get added.
2022-05-01 11:46:59 -07:00
Phil Wang
ad8d7a368b
product management
2022-05-01 11:26:21 -07:00
Phil Wang
b8cf1e5c20
more attention
0.0.87
2022-05-01 11:00:33 -07:00
Phil Wang
94aaa08d97
product management
2022-05-01 09:43:10 -07:00
Phil Wang
8b9bbec7d1
project management
0.0.86
2022-05-01 09:32:57 -07:00
Phil Wang
1bb9fc9829
add convnext backbone for vqgan-vae, still need to fix groupnorms in resnet encdec
2022-05-01 09:32:24 -07:00
Phil Wang
5e421bd5bb
let researchers do the hyperparameter search
0.0.85
2022-05-01 08:46:21 -07:00
Phil Wang
67fcab1122
add MLP based time conditioning to all convnexts, in addition to cross attention. also add an initial convolution, given convnext first depthwise conv
0.0.84
2022-05-01 08:41:02 -07:00
Phil Wang
5bfbccda22
port over vqgan vae trainer
2022-05-01 08:09:15 -07:00
Phil Wang
989275ff59
product management
2022-04-30 16:57:56 -07:00
Phil Wang
56408f4a40
project management
2022-04-30 16:57:02 -07:00