From 1992d25cadfe0ef3f8f931bb81209aeb1d98648a Mon Sep 17 00:00:00 2001 From: Phil Wang Date: Wed, 4 May 2022 11:18:54 -0700 Subject: [PATCH] project management --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index 523d46a..f7b6e6a 100644 --- a/README.md +++ b/README.md @@ -821,6 +821,7 @@ Once built, images will be saved to the same directory the command is invoked - [x] just take care of the training for the decoder in a wrapper class, as each unet in the cascade will need its own optimizer - [x] bring in tools to train vqgan-vae - [x] add convnext backbone for vqgan-vae (in addition to vit [vit-vqgan] + resnet) +- [x] make sure DDPMs can be run with traditional resnet blocks (but leave convnext as an option for experimentation) - [ ] become an expert with unets, cleanup unet code, make it fully configurable, port all learnings over to https://github.com/lucidrains/x-unet (test out unet² in ddpm repo) - [ ] copy the cascading ddpm code to a separate repo (perhaps https://github.com/lucidrains/denoising-diffusion-pytorch) as the main contribution of dalle2 really is just the prior network - [ ] transcribe code to Jax, which lowers the activation energy for distributed training, given access to TPUs @@ -835,7 +836,6 @@ Once built, images will be saved to the same directory the command is invoked - [ ] make sure for the latter unets in the cascade, one can train on crops for learning super resolution (constrain the unet to be only convolutions in that case, or allow conv-like attention with rel pos bias) - [ ] interface out the vqgan-vae so a pretrained one can be pulled off the shelf to validate latent diffusion + DALL-E2 - [ ] make sure FILIP works with DALL-E2 from x-clip https://arxiv.org/abs/2111.07783 -- [ ] make sure DDPMs can be run with traditional resnet blocks (but leave convnext as an option for experimentation) ## Citations