-
|
Hi! So I have managed to get ubergarm's 235B quant to work on a 6 year old workstation with 2*2080TI's, 64GB RAM and a pretty fast (new) SSD. I have encountered some wierd issues with trying to use multiple GPUs though:
Command used (some variation of): Am I just doing something wrong or is there some genuine bug here? |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 6 replies
-
|
Split mode "row" does not work for MoE models (and I'm not sure if it works for dense models as I don't have access to a multi-GPU system, so have not tested since forking). I'm pretty sure split mode "row" does not work for MoE models in mainline With two or more GPU's you may need a more complicated tensor override recipe to get the best possible performance out of the system. For two identical GPU's I think you could start by using note how much VRAM this has used on each GPU, and then change to e.g. (I'm just guessing, as I don't have access to a multi-GPU system). Note that the tensor overrides are processed in the order they were defined on the command line. So, in the above example, we don't need to be specific about experts tensor layers going to the CPU because the ones that we want to stay on the GPU (layers 0,1 on CUDA0, layers 2,3 on CUDA1) were already handled, so all remaining experts go to the CPU. If the GPUs are different, then it may be better to just manually define with |
Beta Was this translation helpful? Give feedback.
@matt23654
First I'm not sure where this came from but a lot of folks keep using
-ot "^blk\.[3-9]\.ffn_.*_exps\.=CPU"which misses some other ffn layers without theexpsas the naming convention on Qwen3 is a bit different than DeepSeek for example.One other tip for multi-gpu is to recompile with
-DGGML_SCHED_MAX_COPIES=1Look here for more discussions and examples: https://huggingface.co/ubergarm/Qwen3-235B-A22B-GGUF/discussions/1#681642d4a383b2fb9aa3bd8c
Keep us posted how you get along, as some others have reported success with multi-gpu once they get the arguments just right for their specific systems!