Skip to content

Llama agent

igardev edited this page Aug 11, 2025 · 10 revisions

Llama Agent

Requred servers

  • Tools server
  • Chat server (if search_source tool is used)
  • Embeddings server (if search_source tool is used)

How to use it

Llama Agent is still in development, but could produce some results with intlligent models with tools support.

The best wey to prepare the environment for the agent is by selecting and Orchestra (group of models). So, below is the standard workflow:

  1. Select "Show Llama Agent" from llama-vscode menu or Ctrl+Shift+A to show Llama Agent.
  2. Click "Orhestra" button and select orchestra, which supports agent, for your needes. This will download the required models and start llama.cpp servers with them. For the external servers (like OpenRouter) llama-vscode will ask for api key if needed.
  3. Write your request and send it with Enter or the Send button.

Optional

  • You could add files to the context with the @ button.
  • You could select source code and activate the agent (Ctrl+Shift+A) to attache the selected lines to the contxt
  • You could choose the tools to be used from "Select Tools" button (on the right side of "New Chat" button). If you have installed and started MCP Servers in VS Code, their tools will be available for selection too. Don't forget to click the OK button after changing the tool selection.

Click button "Stop Orchestra" to stop the servers. Click button "Selected Models" to show details about the selected models

Clone this wiki locally