take care of gradient accumulation automatically for researchers, by passing in a max_batch_size on the decoder or diffusion prior trainer forward

This commit is contained in:
Phil Wang
2022-05-14 17:04:09 -07:00
parent b494ed81d4
commit b0cd5f24b6
3 changed files with 85 additions and 16 deletions

View File

@@ -10,7 +10,7 @@ setup(
'dream = dalle2_pytorch.cli:dream'
],
},
version = '0.2.24',
version = '0.2.26',
license='MIT',
description = 'DALL-E 2',
author = 'Phil Wang',