Skip to content

Commit bedaf9f

Browse files
committed
More edits
1 parent 746261f commit bedaf9f

File tree

1 file changed

+7
-6
lines changed

1 file changed

+7
-6
lines changed

solutions/observability/apps/llm-observability.md

Lines changed: 7 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -8,8 +8,8 @@ While LLMs hold incredible transformative potential, they also bring complex cha
88
To keep your LLM-powered applications reliable, efficient, cost-effective, and easy to troubleshoot, Elastic provides a powerful LLM observability framework including key metrics, logs, and traces, along with pre-configured, out-of-the-box dashboards that deliver deep insights into model prompts and responses, performance, usage, and costs.
99
Elastic’s end-to-end LLM observability is delivered through the following methods:
1010

11-
- Metrics and logs ingestion for LLM APIs (via Elastic integrations)
12-
- APM tracing for OpenAI Models (via instrumentation)
11+
- Metrics and logs ingestion for LLM APIs (via [Elastic integrations](https://www.elastic.co/guide/en/integrations/current/introduction.html))
12+
- APM tracing for OpenAI Models (via [instrumentation](https://github.com/elastic/opentelemetry))
1313

1414
## Metrics and logs ingestion for LLM APIs (via Elastic integrations)
1515

@@ -36,20 +36,21 @@ Check these instructions on how to setup and collect OpenTelemetry data for your
3636

3737
## Use cases
3838

39-
Understand LLM performance and reliability
40-
For an SRE team optimizing a customer support system powered by Azure OpenAI, Elastic’s Azure OpenAI integration provides critical insights. They can quickly identify which model variants experience higher latency or error rates, enabling smarter decisions on model deployment or even switching providers based on real-time performance metrics.
39+
### Understand LLM performance and reliability
40+
41+
For an SRE team optimizing a customer support system powered by Azure OpenAI, Elastic’s [Azure OpenAI integration](https://www.elastic.co/guide/en/integrations/current/azure_openai.html) provides critical insights. They can quickly identify which model variants experience higher latency or error rates, enabling smarter decisions on model deployment or even switching providers based on real-time performance metrics.
4142

4243
[image]
4344

4445
### Troubleshoot OpenAI-powered applications
4546

46-
Consider an enterprise utilizing an OpenAI model for real-time user interactions. Encountering unexplained delays, an SRE can use OpenAI tracing to dissect the transaction pathway, identify if one specific API call or model invocation is the bottleneck, and monitor a request to see the exact prompt and response between the user and the LLM. Such insight is pivotal for prompt resolution.
47+
Consider an enterprise utilizing an OpenAI model for real-time user interactions. Encountering unexplained delays, an SRE can use OpenAI tracing to dissect the transaction pathway, identify if one specific API call or model invocation is the bottleneck, and monitor a request to see the exact prompt and response between the user and the LLM.
4748

4849
[image]
4950

5051
### Addressing cost and usage concerns
5152

52-
For cost-sensitive deployments, being acutely aware of which LLM configurations are more cost-effective is crucial. Elastic’s dashboards, pre-configured to display model usage patterns, help mitigate unnecessary spending effectively. You can find out-of-the box dashboards for Azure OpenAI, OpenAI, Amazon Bedrock, and Google VertexAI models.
53+
For cost-sensitive deployments, being acutely aware of which LLM configurations are more cost-effective is crucial. Elastic’s dashboards, pre-configured to display model usage patterns, help mitigate unnecessary spending effectively. You can use out-of-the-box dashboards for Azure OpenAI, OpenAI, Amazon Bedrock, and Google VertexAI models.
5354

5455
[image]
5556

0 commit comments

Comments
 (0)