Skip to content

Commit 1163769

Browse files
committed
Remove token count printing
1 parent 84422d9 commit 1163769

File tree

1 file changed

+0
-3
lines changed

1 file changed

+0
-3
lines changed

examples/models/llama/runner/generation.py

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -110,9 +110,7 @@ def generate( # noqa: C901
110110
print(f"{self.tokenizer.decode_token(current_token)}", end="", flush=True)
111111
tokens = prompt_tokens + [current_token]
112112

113-
i = 0
114113
while len(tokens) < max_seq_len:
115-
print(f"{i} out of {self.max_seq_len} max tokens generated")
116114
if self.use_kv_cache:
117115
logits = self.forward(
118116
tokens=torch.tensor(
@@ -142,7 +140,6 @@ def generate( # noqa: C901
142140
):
143141
break
144142

145-
i += 1
146143
print(f"{self.tokenizer.decode_token(current_token)}", end="", flush=True)
147144
print("\n")
148145

0 commit comments

Comments
 (0)