We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 84422d9 commit 1163769Copy full SHA for 1163769
examples/models/llama/runner/generation.py
@@ -110,9 +110,7 @@ def generate( # noqa: C901
110
print(f"{self.tokenizer.decode_token(current_token)}", end="", flush=True)
111
tokens = prompt_tokens + [current_token]
112
113
- i = 0
114
while len(tokens) < max_seq_len:
115
- print(f"{i} out of {self.max_seq_len} max tokens generated")
116
if self.use_kv_cache:
117
logits = self.forward(
118
tokens=torch.tensor(
@@ -142,7 +140,6 @@ def generate( # noqa: C901
142
140
):
143
141
break
144
145
- i += 1
146
147
print("\n")
148
0 commit comments