|
31 | 31 | from kernels import get_kernel |
32 | 32 |
|
33 | 33 | activation = get_kernel("kernels-community/activation", revision="add_more_act") |
34 | | - silu_kernel = activation.silu |
| 34 | + silu_kernel = activation.layers.Silu |
35 | 35 |
|
36 | 36 |
|
37 | 37 | class AdaLayerNorm(nn.Module): |
@@ -69,7 +69,7 @@ def __init__( |
69 | 69 | if not DIFFUSERS_ENABLE_HUB_KERNELS: |
70 | 70 | self.silu = nn.SiLU() |
71 | 71 | else: |
72 | | - self.silu = silu_kernel |
| 72 | + self.silu = silu_kernel() |
73 | 73 | self.linear = nn.Linear(embedding_dim, output_dim) |
74 | 74 | self.norm = nn.LayerNorm(output_dim // 2, norm_eps, norm_elementwise_affine) |
75 | 75 |
|
@@ -159,7 +159,7 @@ def __init__(self, embedding_dim: int, num_embeddings: Optional[int] = None, nor |
159 | 159 | if not DIFFUSERS_ENABLE_HUB_KERNELS: |
160 | 160 | self.silu = nn.SiLU() |
161 | 161 | else: |
162 | | - self.silu = silu_kernel |
| 162 | + self.silu = silu_kernel() |
163 | 163 | self.linear = nn.Linear(embedding_dim, 6 * embedding_dim, bias=bias) |
164 | 164 | if norm_type == "layer_norm": |
165 | 165 | self.norm = nn.LayerNorm(embedding_dim, elementwise_affine=False, eps=1e-6) |
@@ -201,7 +201,7 @@ def __init__(self, embedding_dim: int, norm_type="layer_norm", bias=True): |
201 | 201 | if not DIFFUSERS_ENABLE_HUB_KERNELS: |
202 | 202 | self.silu = nn.SiLU() |
203 | 203 | else: |
204 | | - self.silu = silu_kernel |
| 204 | + self.silu = silu_kernel() |
205 | 205 | self.linear = nn.Linear(embedding_dim, 3 * embedding_dim, bias=bias) |
206 | 206 | if norm_type == "layer_norm": |
207 | 207 | self.norm = nn.LayerNorm(embedding_dim, elementwise_affine=False, eps=1e-6) |
@@ -356,7 +356,7 @@ def __init__( |
356 | 356 | if not DIFFUSERS_ENABLE_HUB_KERNELS: |
357 | 357 | self.silu = nn.SiLU() |
358 | 358 | else: |
359 | | - self.silu = silu_kernel |
| 359 | + self.silu = silu_kernel() |
360 | 360 | self.linear = nn.Linear(conditioning_embedding_dim, embedding_dim * 2, bias=bias) |
361 | 361 | if norm_type == "layer_norm": |
362 | 362 | self.norm = LayerNorm(embedding_dim, eps, elementwise_affine, bias) |
|
0 commit comments