mirror of
https://github.com/lucidrains/DALLE2-pytorch.git
synced 2025-12-19 09:44:19 +01:00
add ability to specify full self attention on specific stages in the unet
This commit is contained in:
@@ -216,6 +216,7 @@ class UnetConfig(BaseModel):
|
||||
cond_on_text_encodings: bool = None
|
||||
cond_dim: int = None
|
||||
channels: int = 3
|
||||
self_attn: ListOrTuple(int)
|
||||
attn_dim_head: int = 32
|
||||
attn_heads: int = 16
|
||||
|
||||
|
||||
Reference in New Issue
Block a user