-
Notifications
You must be signed in to change notification settings - Fork 12.7k
Open
Labels
Description
Name and Version
version: 6111 (7ad67ba) win cuda
Operating systems
Windows
Which llama.cpp modules do you know to be affected?
Other (Please specify in the next section)
Command line
llama-perplexity.exe -f wiki.test.raw --chunks 128 --n-gpu-layers 999 --model gpt-oss-20b-mxfp4.gguf
llama-perplexity.exe -f wiki.test.raw --chunks 128 --n-gpu-layers 999 --model openai_gpt-oss-20b-Q8_0.gguf
Problem description & steps to reproduce
I tried two different versions of the 20b model, but result is the same, perplexity skyrockets to the moon.
Other unrelated models don't have this problem.
First Bad Commit
No response
Relevant log output
gpt-oss-20b-mxfp4.gguf
Final estimate: PPL = 195.7201 +/- 3.45525
baopenai_gpt-oss-20b-Q8_0.gguf
Final estimate: PPL = 196.2824 +/- 3.46283