Why does imatrix not tokenize special tokens? #395
Replies: 2 comments 7 replies
-
When the Back then I had the concept that the calibration data for chat/instruction tuned models need to contain actual instruction tuning datasets. And, instead of blindly dividing the calibration data into chunks of In any case, it would be interesting to see if including special tokens, using non equal-size chunks, etc., in the |
Beta Was this translation helpful? Give feedback.
-
@bartowski1182 I see you submitted this PR in mainline. You are welcome. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Recently there's been some discussion (and I've also experimented slightly) around adding chat tokens to the imatrix dataset and tokenizing them, a change from the default behaviour, so I was curious why the original implementation avoided tokenizing them
Was it just an arbitrary decision or was there a reason at the time?
Beta Was this translation helpful? Give feedback.
All reactions