Skip to content

Commit f834263

Browse files
authored
remove deprecated message (#1619)
1 parent 1280744 commit f834263

File tree

1 file changed

+0
-6
lines changed

1 file changed

+0
-6
lines changed

torchtune/modules/transformer.py

Lines changed: 0 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,6 @@
1111
from torch import nn
1212
from torchtune.modules import MultiHeadAttention
1313
from torchtune.modules.attention_utils import _MaskType
14-
from torchtune.utils._logging import deprecated
1514

1615

1716
class TransformerSelfAttentionLayer(nn.Module):
@@ -619,11 +618,6 @@ def forward(
619618
return output
620619

621620

622-
@deprecated(
623-
msg="Please use torchtune.modules.TransformerDecoder instead. \
624-
If you need an example, see torchtune.models.qwen2._component_builders.py \
625-
and how to implement torch.modules.TiedLinear for the output projection."
626-
)
627621
class TiedEmbeddingTransformerDecoder(nn.Module):
628622
"""
629623
Transformer Decoder with tied embedding weight. A key difference between

0 commit comments

Comments
 (0)