mirror of
https://github.com/lucidrains/DALLE2-pytorch.git
synced 2025-12-19 09:44:19 +01:00
quick fix for linear attention
This commit is contained in:
@@ -1490,7 +1490,8 @@ class LinearAttention(nn.Module):
|
|||||||
self,
|
self,
|
||||||
dim,
|
dim,
|
||||||
dim_head = 32,
|
dim_head = 32,
|
||||||
heads = 8
|
heads = 8,
|
||||||
|
**kwargs
|
||||||
):
|
):
|
||||||
super().__init__()
|
super().__init__()
|
||||||
self.scale = dim_head ** -0.5
|
self.scale = dim_head ** -0.5
|
||||||
|
|||||||
@@ -1 +1 @@
|
|||||||
__version__ = '1.4.3'
|
__version__ = '1.4.4'
|
||||||
|
|||||||
Reference in New Issue
Block a user