feat: Replace mock LLM with local SmolLM2-135M inference#3014
Draft
ps48 wants to merge 2 commits intoopen-telemetry:mainfrom
Draft
feat: Replace mock LLM with local SmolLM2-135M inference#3014ps48 wants to merge 2 commits intoopen-telemetry:mainfrom
ps48 wants to merge 2 commits intoopen-telemetry:mainfrom
Conversation
Signed-off-by: ps48 <pshenoy36@gmail.com>
Signed-off-by: ps48 <pshenoy36@gmail.com>
|
Real inference, nice! One thought on the instrumentation side: The
The LLM service can stay simple: Flask + llama-cpp-python + Flask auto-instrumentation. The GenAI telemetry makes more sense on the client side where the Setup would just be: from opentelemetry.instrumentation.openai_v2 import OpenAIInstrumentor
OpenAIInstrumentor().instrument()What do you think? |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Changes
Summary
llama-cpp-python, removing the need for an external API while still exposing an OpenAI-compatible/v1/chat/completionsendpointproduct_reviews_server.pyby dropping the custom tool-calling flow in favor of a direct chat completion with review context in the promptsrc/llm/requirements.txtto 1.39.1 / 0.60b1.env,.env.override, andsrc/product-reviews/README.mdto reflect the new default modelMerge Requirements
For new features contributions, please make sure you have completed the following
essential items:
CHANGELOG.mdupdated to document new feature additionsMaintainers will not merge until the above have been completed. If you're unsure
which docs need to be changed ping the
@open-telemetry/demo-approvers.