always use sandwich norm for attention layer

This commit is contained in:
Phil Wang
2022-05-14 12:13:41 -07:00
parent 9faab59b23
commit d1f02e8f49
2 changed files with 4 additions and 7 deletions

View File

@@ -10,7 +10,7 @@ setup(
'dream = dalle2_pytorch.cli:dream'
],
},
version = '0.2.18',
version = '0.2.19',
license='MIT',
description = 'DALL-E 2',
author = 'Phil Wang',