-
Notifications
You must be signed in to change notification settings - Fork 157
[LLM Observability] Add new landing page #782
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Merged
Merged
Changes from 1 commit
Commits
Show all changes
33 commits
Select commit
Hold shift + click to select a range
ee50689
Add new LLM obs page
alaudazzi d9c6011
Edits
alaudazzi 6149699
Add links to integrations pages
alaudazzi 746261f
Add links to otel distros
alaudazzi bedaf9f
More edits
alaudazzi 8f64355
Add image for the first use case
alaudazzi df0544e
Add more images for use cases
alaudazzi c2f23af
Merge branch 'main' into llm-observability
alaudazzi e4160d3
Merge branch 'main' into llm-observability
alaudazzi c10ad22
Add integrations and instrumentaions tables
alaudazzi 399a4ca
Merge branch 'llm-observability' of github.com:elastic/docs-content i…
alaudazzi 0162235
Update solutions/observability/apps/llm-observability.md
alaudazzi fc7cb75
Update solutions/observability/apps/llm-observability.md
alaudazzi f9f41eb
Update opentelemtry link
alaudazzi e001d45
Integrate Daniela's suggestions
alaudazzi 6511d42
Integrate Christoph's suggestion
alaudazzi d676101
Merge branch 'main' into llm-observability
alaudazzi f3f9d96
Merge branch 'main' into llm-observability
alaudazzi c684a86
Merge branch 'main' into llm-observability
alaudazzi d5a83b5
Update link to quickstart
alaudazzi 47a0230
Merge branch 'main' into llm-observability
alaudazzi 94770d0
Merge branch 'main' into llm-observability
alaudazzi 873f16c
Integrate Daniela's feedback
alaudazzi 252433d
Merge branch 'main' into llm-observability
alaudazzi 08b482b
Update AWS Boto link
alaudazzi 0e9f4f9
Update solutions/observability/apps/llm-observability.md
alaudazzi d268c0d
Change heading level for GS
alaudazzi c07b0a4
Merge branch 'main' into llm-observability
alaudazzi 88c5646
Replace table with link to supported technologies
alaudazzi b63caed
Merge branch 'llm-observability' of github.com:elastic/docs-content i…
alaudazzi b38951a
Merge branch 'main' into llm-observability
alaudazzi a26136b
Merge branch 'main' into llm-observability
alaudazzi 13bbc39
Add links to the source column
alaudazzi File filter
Filter by extension
Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
There are no files selected for viewing
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,60 @@ | ||
--- | ||
navigation_title: "LLM Observability" | ||
alaudazzi marked this conversation as resolved.
Show resolved
Hide resolved
|
||
--- | ||
|
||
# LLM Observability | ||
|
||
While LLMs hold incredible transformative potential, they also bring complex challenges in reliability, performance, and cost management. Traditional monitoring tools require an evolved set of observability capabilities to ensure these models operate efficiently and effectively. | ||
To keep your LLM-powered applications reliable, efficient, cost-effective, and easy to troubleshoot, Elastic provides a powerful LLM observability framework including key metrics, logs, and traces, along with pre-configured, out-of-the-box dashboards that deliver deep insights into model prompts and responses, performance, usage, and costs. | ||
Elastic’s end-to-end LLM observability is delivered through the following methods: | ||
|
||
- Metrics and logs ingestion for LLM APIs (via Elastic integrations) | ||
- APM tracing for OpenAI Models (via instrumentation) | ||
|
||
## Metrics and logs ingestion for LLM APIs (via Elastic integrations) | ||
|
||
Elastic’s LLM integrations now support the most widely adopted models, including OpenAI, Azure OpenAI, and a diverse range of models hosted on Amazon Bedrock and Google Vertex AI: | ||
|
||
- Amazon Bedrock | ||
- Azure OpenAI | ||
- GCP Vertex AI | ||
- OpenAI | ||
|
||
## APM tracing for OpenAI Models (via instrumentation) | ||
|
||
Elastic offers specialized OpenTelemetry Protocol (OTLP) tracing for applications leveraging OpeAI models hosted on OpenAI, Azure, and Amazon Bedrock, providing a detailed view of request flows. This tracing capability captures critical insights, including the specific models used, request duration, errors encountered, token consumption per request, and the interaction between prompts and responses. Ideal for troubleshooting, APM tracing allows you to find exactly where the issue is happening with precision and efficiency in your OpenAI-powered application. | ||
alaudazzi marked this conversation as resolved.
Outdated
Show resolved
Hide resolved
|
||
|
||
You can Instrument the application with one of the following OpenTelemetry API: | ||
|
||
- Python | ||
- Node.js | ||
- Java | ||
|
||
## Getting started | ||
alaudazzi marked this conversation as resolved.
Outdated
Show resolved
Hide resolved
|
||
|
||
Check these instructions on how to setup and collect OpenTelemetry data for your LLM applications [create a link to https://github.com/elastic/opentelemetry/pull/100/files#diff-965570d21670c0ee4bba4b303960e5fe83b285f66b001ff8f31f0413f65a9d47 once the content is finalized and merged] | ||
|
||
## Use cases | ||
|
||
Understand LLM performance and reliability | ||
For an SRE team optimizing a customer support system powered by Azure OpenAI, Elastic’s Azure OpenAI integration provides critical insights. They can quickly identify which model variants experience higher latency or error rates, enabling smarter decisions on model deployment or even switching providers based on real-time performance metrics. | ||
|
||
[image] | ||
|
||
## Troubleshoot OpenAI-powered applications | ||
|
||
Consider an enterprise utilizing an OpenAI model for real-time user interactions. Encountering unexplained delays, an SRE can use OpenAI tracing to dissect the transaction pathway, identify if one specific API call or model invocation is the bottleneck, and monitor a request to see the exact prompt and response between the user and the LLM. Such insight is pivotal for prompt resolution. | ||
|
||
[image] | ||
|
||
## Addressing Cost and Usage Concerns | ||
|
||
For cost-sensitive deployments, being acutely aware of which LLM configurations are more cost-effective is crucial. Elastic’s dashboards, pre-configured to display model usage patterns, help mitigate unnecessary spending effectively. You can find out-of-the box dashboards for Azure OpenAI, OpenAI, Amazon Bedrock, and Google VertexAI models. | ||
|
||
[image] | ||
|
||
## Understand Compliance with Guardrails in Amazon Bedrock | ||
|
||
With the Elastic Amazon Bedrock integration for Guardrails, SREs can swiftly address security concerns, like verifying if certain user interactions prompt policy violations. Elastic's observability logs clarify whether guardrails rightly blocked potentially harmful responses, bolstering compliance assurance. | ||
|
||
[image] |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Uh oh!
There was an error while loading. Please reload this page.