Skip to content

[Do not merge] [Obs AI Assistant] Adds docs for connecting to a local LLM with the Obs AI Assistant #2536

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wants to merge 22 commits into
base: main
Choose a base branch
from

Conversation

viduni94
Copy link
Contributor

@viduni94 viduni94 commented Aug 12, 2025

Closes https://github.com/elastic/obs-ai-assistant-team/issues/322

[DO NOT MERGE] This PR can only be merged once llama support is complete.

This PR adds documentation about how to connect to a local LLM with the Observability AI Assistant.

@viduni94 viduni94 self-assigned this Aug 12, 2025
@viduni94 viduni94 requested review from a team as code owners August 12, 2025 21:32
@viduni94 viduni94 added the documentation Improvements or additions to documentation label Aug 12, 2025
Setup is now complete. You can use the model you’ve loaded in LM Studio to power Elastic’s generative AI features.

::::{note}
While local (open-weight) LLMs offer greater privacy and control, they generally do not match the raw performance and advanced reasoning capabilities of proprietary models by LLM providers mentioned in [here](/solutions/observability/observability-ai-assistant.md#set-up-the-ai-assistant-obs-ai-set-up)
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Added a small disclaimer here - let me know if I should remove it.

@viduni94
Copy link
Contributor Author

I'm waiting for the preview to be generated to double check the links and formatting.

Copy link

github-actions bot commented Aug 12, 2025

@viduni94
Copy link
Contributor Author

@SrdjanLL

Is it a must to use Nginx for this purpose? If not, I'd suggest mentioning we have an example of setting up NGINX reverse proxy, but to be less prescriptive, because customers may be using other reverse proxies (HAProxy, Envoy Proxy, Caddy, etc.).

The reverse proxy configuration in the doc I've linked mentions the follows. While the user can use alternatives to Nginx for testing purposes, the recommendation is to use Nginx to collect telemetry via the Elastic's Nginx integration.

For testing, you can use alternatives to Nginx such as Azure Dev Tunnels or Ngrok, but using Nginx makes it easy to collect additional telemetry and monitor its status by using Elastic’s native Nginx integration. While this example uses cloud infrastructure, it could also be replicated locally without an internet connection.

@viduni94
Copy link
Contributor Author

Hi @mdbirnstiehl
The PR is ready for review. I've addressed the comments from @benironside

Let me know if I need to add anything to target this for Serverless and ECH 9.2
Thank you

@viduni94 viduni94 requested a review from benironside August 15, 2025 18:31
@arturoliduena
Copy link

Thanks, @viduni94, for adding this documentation. I have created a task to add vLLM documentation: elastic/kibana#232052

@arturoliduena
Copy link

@viduni94, should we add the minimum requirement to run llama3.3? WDYT?
e.g, Memory GPU, Nvidia drivers, disk size, etc

@viduni94
Copy link
Contributor Author

viduni94 commented Aug 18, 2025

@viduni94, should we add the minimum requirement to run llama3.3? WDYT? e.g, Memory GPU, Nvidia drivers, disk size, etc

It might not be necessary to add this because it's general information and not specific to Elastic.

@viduni94
Copy link
Contributor Author

viduni94 commented Aug 18, 2025

Thanks, @viduni94, for adding this documentation. I have created a task to add vLLM documentation: elastic/kibana#232052

Thanks for the review @arturoliduena
It might not be necessary (but a good to have) to add how to connect via vLLM. The most important part of the documentation is connector creation since that's the only part specific to Elastic. The other sections (connecting to LM studio) is added for completeness. Let's discuss with the team.

Copy link
Contributor

@mdbirnstiehl mdbirnstiehl left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thank you putt this together! I've added some comments and suggestions. Let me know if you have any questions.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
documentation Improvements or additions to documentation v9.2.0
Projects
None yet
Development

Successfully merging this pull request may close these issues.

6 participants