Compare commits

...

3 Commits

Author SHA1 Message Date
Phil Wang
14ddbc159c cleanup 2022-04-13 18:24:32 -07:00
Phil Wang
0692f1699f favorite quote 2022-04-13 18:17:59 -07:00
Phil Wang
26c4534bc3 readme 2022-04-13 18:11:55 -07:00
3 changed files with 6 additions and 5 deletions

View File

@@ -199,7 +199,7 @@ dalle2 = DALLE2(
decoder = decoder decoder = decoder
) )
# send the text as a string if you want to use the simple tokenizer from DALL-E1 # send the text as a string if you want to use the simple tokenizer from DALLE v1
# or you can do it as token ids, if you have your own tokenizer # or you can do it as token ids, if you have your own tokenizer
texts = ['glistening morning dew on a flower petal'] texts = ['glistening morning dew on a flower petal']
@@ -214,8 +214,6 @@ Let's see the whole script below
import torch import torch
from dalle2_pytorch import DALLE2, DiffusionPriorNetwork, DiffusionPrior, Unet, Decoder, CLIP from dalle2_pytorch import DALLE2, DiffusionPriorNetwork, DiffusionPrior, Unet, Decoder, CLIP
import torch
clip = CLIP( clip = CLIP(
dim_text = 512, dim_text = 512,
dim_image = 512, dim_image = 512,
@@ -303,6 +301,8 @@ images = dalle2(['cute puppy chasing after a squirrel'])
Everything in this readme should run without error Everything in this readme should run without error
For the layperson, no worries, training will all be automated into a CLI tool, at least for small scale training.
## Training CLI (wip) ## Training CLI (wip)
<a href="https://github.com/lucidrains/stylegan2-pytorch">template</a> <a href="https://github.com/lucidrains/stylegan2-pytorch">template</a>
@@ -364,3 +364,5 @@ Everything in this readme should run without error
primaryClass = {cs.LG} primaryClass = {cs.LG}
} }
``` ```
*Creating noise from data is easy; creating data from noise is generative modeling.* - Yang Song's <a href="https://arxiv.org/abs/2011.13456">paper</a>

View File

@@ -877,7 +877,6 @@ class DALLE2(nn.Module):
text = [text] if not isinstance(text, (list, tuple)) else text text = [text] if not isinstance(text, (list, tuple)) else text
text = tokenizer.tokenize(text).to(device) text = tokenizer.tokenize(text).to(device)
print(text.shape, type(text))
image_embed = self.prior.sample(text, num_samples_per_batch = self.prior_num_samples) image_embed = self.prior.sample(text, num_samples_per_batch = self.prior_num_samples)
images = self.decoder.sample(image_embed) images = self.decoder.sample(image_embed)
return images return images

View File

@@ -10,7 +10,7 @@ setup(
'dream = dalle2_pytorch.cli:dream' 'dream = dalle2_pytorch.cli:dream'
], ],
}, },
version = '0.0.5', version = '0.0.6',
license='MIT', license='MIT',
description = 'DALL-E 2', description = 'DALL-E 2',
author = 'Phil Wang', author = 'Phil Wang',