Deepseek v3/r1 poisoned prompt? #621
magikRUKKOLA
started this conversation in
General
Replies: 1 comment 9 replies
-
What sampler settings do you use? |
Beta Was this translation helpful? Give feedback.
9 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
[UPDATE]: the current bug IS UNRELATED to the Deepseek inference implementation! Its
a bug of the llm itselfjust a stupidity of mine.[UPDATE2:] solved! #621 (reply in thread) the issue is related to the usage of the special tokens in the prompt.
I tried different quants and settings. The answer of Deepseek V3 or R1 is surprisingly stupid. Initially I thought its a bug in a RAM then I realized I have ECC ram. So ... its about 20k context prompt and Deepseek answers more or less the following absolute nonsense:
Any idea what the hell is going on?
basically you can dump the prompt and send it like (if you use mods):
lol i never had such problems with this model at all. All I have done is downloaded some content from @Thireus repo and threw it onto the llm. The result is quite surprising. Let me try to reduce the prompt to figure out what may cause a bug.
the prompt:
Beta Was this translation helpful? Give feedback.
All reactions