From 895269874f9228b55ef44928e3d905f5d61ac521 Mon Sep 17 00:00:00 2001 From: Utsav Popat Date: Mon, 19 May 2025 21:34:02 -0400 Subject: [PATCH 1/4] update ubuntu version --- .github/workflows/pr.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/pr.yml b/.github/workflows/pr.yml index 83b79c972..85e832de1 100644 --- a/.github/workflows/pr.yml +++ b/.github/workflows/pr.yml @@ -9,7 +9,7 @@ concurrency: jobs: lint: - runs-on: ubuntu-20.04 + runs-on: ubuntu-latest steps: - uses: actions/checkout@v4 - uses: ./.github/actions/setup-python/ From 4082fee800acc6d410fe7368458bf9a981bd8fbd Mon Sep 17 00:00:00 2001 From: Utsav Popat <20182650+parallelipopat@users.noreply.github.com> Date: Tue, 20 May 2025 01:40:28 +0000 Subject: [PATCH 2/4] lint fixes --- llama-cpp-server/config.yaml | 2 +- orpheus-best-performance/model/model.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/llama-cpp-server/config.yaml b/llama-cpp-server/config.yaml index 9286e6349..64ebb7a85 100644 --- a/llama-cpp-server/config.yaml +++ b/llama-cpp-server/config.yaml @@ -1,6 +1,6 @@ base_image: image: alphatozeta/llama-cpp-server:0.4 -build_commands: +build_commands: - pip install git+https://github.com/huggingface/transformers.git hf-xet model_metadata: repo_id: google/gemma-3-27b-it-qat-q4_0-gguf diff --git a/orpheus-best-performance/model/model.py b/orpheus-best-performance/model/model.py index 00c498f71..fec6e2d00 100644 --- a/orpheus-best-performance/model/model.py +++ b/orpheus-best-performance/model/model.py @@ -276,7 +276,7 @@ async def predict( async def audio_stream(req_id: str): token_gen = await self._engine.predict(model_input, request) - + if isinstance(token_gen, StreamingResponse): token_gen = token_gen.body_iterator From a5b6eed420ded24a96bf35f374668aca493501e4 Mon Sep 17 00:00:00 2001 From: Utsav Popat <20182650+parallelipopat@users.noreply.github.com> Date: Tue, 20 May 2025 01:51:56 +0000 Subject: [PATCH 3/4] more changes --- llama-cpp-server/README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llama-cpp-server/README.md b/llama-cpp-server/README.md index 4cbaa7b2a..01cdb580b 100644 --- a/llama-cpp-server/README.md +++ b/llama-cpp-server/README.md @@ -23,4 +23,4 @@ cd llama.cpp docker build -t local/llama.cpp:server-cuda --target server -f .devops/cuda.Dockerfile . ``` -You can then push this image to a container registry of your choice and then replace the base_image in the config.yaml \ No newline at end of file +You can then push this image to a container registry of your choice and then replace the base_image in the config.yaml From 454b57526adb4f76f6191e93840a14578221a158 Mon Sep 17 00:00:00 2001 From: Utsav Popat <20182650+parallelipopat@users.noreply.github.com> Date: Tue, 20 May 2025 01:52:16 +0000 Subject: [PATCH 4/4] ruff fixes --- orpheus-best-performance/model/model.py | 1 - 1 file changed, 1 deletion(-) diff --git a/orpheus-best-performance/model/model.py b/orpheus-best-performance/model/model.py index fec6e2d00..6d590ac93 100644 --- a/orpheus-best-performance/model/model.py +++ b/orpheus-best-performance/model/model.py @@ -3,7 +3,6 @@ import torch import fastapi from snac import SNAC -import struct from pathlib import Path import numpy as np from fastapi.responses import StreamingResponse