Replies: 7 comments 5 replies
-
Pretty sweet |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
You should go into detail how you did this across multiple I'd love to read/watch that |
Beta Was this translation helpful? Give feedback.
-
Thanks for showing your results. Just started to use dllama and it looks very nice!
This is the command I used (v0.12.7):
|
Beta Was this translation helpful? Give feedback.
-
Hello b4rtaz, could you tell me what is diffference between evaluation and prediction |
Beta Was this translation helpful? Give feedback.
-
How did you run it successfully on 2 x Raspberry Pi 5 8GB ? When I run it on this scenario, worker node show Listening on 0.0.0.0:9999... My command on root node is ./dllama inference --model models/deepseek_r1_distill_llama_8b_q40/dllama_model_deepseek_r1_distill_llama_8b_q40.m --tokenizer models/deepseek_r1_distill_llama_8b_q40/dllama_tokenizer_deepseek_r1_distill_llama_8b_q40.t --buffer-float-type q80 --steps 16 --nthreads 4 --workers 192.168.1.100:9999 |
Beta Was this translation helpful? Give feedback.
-
Do you know why it only show blank and dot ? 💡 RopeScaling: f=8.0, l=1.0, h=4.0, o=8192 👱 User
🤖 Assistant . . . . . . . . |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Model:
deepseek_r1_distill_llama_8b_q40
Version:
0.12.2
2 x Raspberry Pi 5 8GB
4 x Raspberry Pi 5 8GB
Beta Was this translation helpful? Give feedback.
All reactions