Skip to content

Commit 6d897ec

Browse files
committed
address review comments
Signed-off-by: h-guo18 <[email protected]>
1 parent 04bed24 commit 6d897ec

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

examples/speculative_decoding/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -86,7 +86,7 @@ We can optionally use smaller vocab size for the draft model for faster training
8686
python calibrate_draft_vocab.py --model meta-llama/Llama-3.2-1B-Instruct --data Daring-Anteater/train.jsonl --draft_vocab_size 32000 --save_dir draft_vocab_cache
8787
```
8888

89-
This will produce a `d2t.pt` file in `save_dir`, which is the mapping from draft vocabs to full vocab that will be read by our draft model later.
89+
This will produce a `d2t.pt` file in `save_dir`, which is the mapping from draft token to target token. During inference, draft tokens can be mapped back to target tokens by `target_token = draft_token + d2t[draft_token]`.
9090

9191
### (Optional) Configuring Draft Model
9292

0 commit comments

Comments
 (0)