|
7 | 7 | from typing import Any, Callable, Literal
|
8 | 8 |
|
9 | 9 | import torch
|
10 |
| -import transformers |
11 | 10 | from tensordict import lazy_stack, TensorDict, TensorDictBase
|
12 | 11 | from torch.utils.data import DataLoader
|
13 | 12 | from torchrl.data import Composite, NonTensor
|
@@ -116,7 +115,7 @@ def __init__(
|
116 | 115 | batch_size: tuple | torch.Size | None = None,
|
117 | 116 | system_prompt: str | None = None,
|
118 | 117 | apply_template: bool | None = None,
|
119 |
| - tokenizer: transformers.AutoTokenizer | None = None, |
| 118 | + tokenizer: transformers.AutoTokenizer | None = None, # noqa: F821 |
120 | 119 | template_kwargs: dict[str, Any] | None = None,
|
121 | 120 | system_role: str = "system",
|
122 | 121 | user_role: str = "user",
|
@@ -309,7 +308,7 @@ def __init__(
|
309 | 308 | batch_size_dl: int = 1,
|
310 | 309 | seed: int | None = None,
|
311 | 310 | group_repeats: bool = False,
|
312 |
| - tokenizer: transformers.AutoTokenizer | None = None, |
| 311 | + tokenizer: transformers.AutoTokenizer | None = None, # noqa: F821 |
313 | 312 | device: torch.device | None = None,
|
314 | 313 | template_kwargs: dict[str, Any] | None = None,
|
315 | 314 | apply_template: bool | None = None,
|
|
0 commit comments