Skip to content

Conversation

@kerthcet
Copy link
Contributor

@kerthcet kerthcet commented Aug 20, 2024

Hi, llmaz is a platform to serve large language models on Kubernetes, llama.cpp is an vital part of it for CPU inference as well as GPU part. Here's an example:

apiVersion: llmaz.io/v1alpha1
kind: OpenModel
metadata:
  name: qwen2-0-5b-gguf
spec:
  familyName: qwen2
  source:
    modelHub:
      modelID: Qwen/Qwen2-0.5B-Instruct-GGUF
      filename: qwen2-0_5b-instruct-q5_k_m.gguf
apiVersion: inference.llmaz.io/v1alpha1
kind: Playground
metadata:
  name: qwen2-0-5b
spec:
  replicas: 1
  modelClaim:
    modelName: qwen2-0-5b-gguf
  backendConfig:
    name: llamacpp
    args:
    - -fa # use flash attention

This is all your need to do, then you can serve models with llama.cpp on Kubernetes.

Thanks!

@kerthcet
Copy link
Contributor Author

kindly ping @ggerganov

@kerthcet
Copy link
Contributor Author

Hi @ggerganov let me know if this is a suitable integration for llama.cpp, actually, we use llama.cpp a lot in our platform and we use it for all kinds of tests as well for cost saving. I believe it's a great showcase about llama.cpp in cloud. Thanks!

@ggerganov ggerganov merged commit 53e4db1 into ggml-org:master Feb 26, 2025
1 check passed
mglambda pushed a commit to mglambda/llama.cpp that referenced this pull request Mar 8, 2025
arthw pushed a commit to arthw/llama.cpp that referenced this pull request Mar 19, 2025
mostlyuseful pushed a commit to mostlyuseful/llama.cpp that referenced this pull request May 12, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants