Skip to content

Commit 370b06c

Browse files
committed
Add warning about inactive NJT development to related docs
1 parent 50c91dd commit 370b06c

File tree

2 files changed

+6
-2
lines changed

2 files changed

+6
-2
lines changed

intermediate_source/transformer_building_blocks.py

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -79,6 +79,10 @@
7979
# sequence lengths. They eliminate the need for the bug-prone practices of explicit
8080
# padding and masking (think ``key_padding_mask`` in ``nn.MultiHeadAttention``).
8181
#
82+
# ```{warning}
83+
# Nested tensors are not currently under active development. Use at your own risk.
84+
# ```
85+
#
8286
# * `scaled_dot_product_attention <https://pytorch.org/tutorials/intermediate/scaled_dot_product_attention_tutorial.html>`_
8387
#
8488
# ``scaled_dot_product_attention`` is a primitive for

unstable_source/nestedtensor.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -3,6 +3,8 @@
33
Getting Started with Nested Tensors
44
===============================================================
55
6+
**Warning: Nested tensors are not currently under active development. Use at your own risk.**
7+
68
Nested tensors generalize the shape of regular dense tensors, allowing for representation
79
of ragged-sized data.
810
@@ -21,8 +23,6 @@
2123
they are invaluable for building transformers that can efficiently operate on ragged sequential
2224
inputs. Below, we present an implementation of multi-head attention using nested tensors that,
2325
combined usage of ``torch.compile``, out-performs operating naively on tensors with padding.
24-
25-
Nested tensors are currently a prototype feature and are subject to change.
2626
"""
2727

2828
import numpy as np

0 commit comments

Comments
 (0)