We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 1280744 commit f834263Copy full SHA for f834263
torchtune/modules/transformer.py
@@ -11,7 +11,6 @@
11
from torch import nn
12
from torchtune.modules import MultiHeadAttention
13
from torchtune.modules.attention_utils import _MaskType
14
-from torchtune.utils._logging import deprecated
15
16
17
class TransformerSelfAttentionLayer(nn.Module):
@@ -619,11 +618,6 @@ def forward(
619
618
return output
620
621
622
-@deprecated(
623
- msg="Please use torchtune.modules.TransformerDecoder instead. \
624
-If you need an example, see torchtune.models.qwen2._component_builders.py \
625
-and how to implement torch.modules.TiedLinear for the output projection."
626
-)
627
class TiedEmbeddingTransformerDecoder(nn.Module):
628
"""
629
Transformer Decoder with tied embedding weight. A key difference between
0 commit comments