|
| 1 | +<details> |
| 2 | +<summary>Specific guidance for Azure</summary> |
| 3 | +<div markdown="1"> |
| 4 | +There are 2 hosted model options for Azure: |
| 5 | + |
| 6 | +- Machine Learning Services (MLS) |
| 7 | +- Cognitive Services (CS) |
| 8 | + |
| 9 | +TrustGraph's *Azure* is for integration with MLS. *Azure OpenAI* is for |
| 10 | +integration with CS. If you are using the *Azure* / MLS integration, you |
| 11 | +should make sure you know your model endpoint, and the token granted for the |
| 12 | +endpoint, and configure these values thus: |
| 13 | +``` |
| 14 | +export AZURE_ENDPOINT=https://ENDPOINT.API.HOST.GOES.HERE/ |
| 15 | +export AZURE_TOKEN=TOKEN-GOES-HERE |
| 16 | +``` |
| 17 | +If you are using the *Azure OpenAI* / CS integration, you should make sure |
| 18 | +you know your model endpoint, the token and configure them thus: |
| 19 | +``` |
| 20 | +export AZURE_ENDPOINT=https://ENDPOINT.API.HOST.GOES.HERE/ |
| 21 | +export AZURE_TOKEN=TOKEN-GOES-HERE |
| 22 | +``` |
| 23 | +</div> |
| 24 | +</details> |
| 25 | + |
| 26 | +<details> |
| 27 | +<summary>Specific guidance for AWS Bedrock</summary> |
| 28 | +<div markdown="1"> |
| 29 | +To use Bedrock, you need to have AWS credentials provisioned. |
| 30 | +The easiest way is to create an IAM user, and create credentials for this |
| 31 | +user. When you provision the user, you will be asked to give the user |
| 32 | +permissions. To allow Bedrock access, the `AmazonBedrockFullAccess` |
| 33 | +role should be added. |
| 34 | + |
| 35 | +You would then provision credentials which would give you an *access key ID* |
| 36 | +and a *secret access key*. You should pick the identifier of an |
| 37 | +AWS region to connect to e.g. `eu-west-2`. In order to prepare to deploy, |
| 38 | +you should set three environment variables using the information. |
| 39 | + |
| 40 | +``` |
| 41 | +export AWS_ACCESS_KEY_ID=ID-KEY-HERE |
| 42 | +export AWS_SECRET_ACCESS_KEY=TOKEN-GOES-HERE |
| 43 | +export AWS_DEFAULT_REGION=AWS-REGION-HERE |
| 44 | +``` |
| 45 | + |
| 46 | +Note: You should be very careful with AWS cloud credentials provisioned |
| 47 | +this way: if lost or leaked this provides a malicious person access to the |
| 48 | +AWS resources you gave this user. |
| 49 | +</div> |
| 50 | +</details> |
| 51 | + |
| 52 | +<details> |
| 53 | +<summary>Specific guidance for Anthropic Claude</summary> |
| 54 | +<div markdown="1"> |
| 55 | +To use Anthropic's Claude models directly, sign up for API access at |
| 56 | +[console.anthropic.com](https://console.anthropic.com/). Create an API key |
| 57 | +from the dashboard. Set the key as an environment variable: |
| 58 | + |
| 59 | +``` |
| 60 | +export CLAUDE_KEY=sk-ant-api03-xxxxx |
| 61 | +``` |
| 62 | +</div> |
| 63 | +</details> |
| 64 | + |
| 65 | +<details> |
| 66 | +<summary>Specific guidance for Cohere</summary> |
| 67 | +<div markdown="1"> |
| 68 | +To use Cohere's models, sign up at [cohere.com](https://cohere.com/) and |
| 69 | +create an API key from your dashboard. Set the key as an environment variable: |
| 70 | + |
| 71 | +``` |
| 72 | +export COHERE_KEY=your-cohere-api-key-here |
| 73 | +``` |
| 74 | +</div> |
| 75 | +</details> |
| 76 | + |
| 77 | +<details> |
| 78 | +<summary>Specific guidance for Google AI Studio</summary> |
| 79 | +<div markdown="1"> |
| 80 | +To use Google's Gemini models via AI Studio, visit |
| 81 | +[aistudio.google.com](https://aistudio.google.com/) and generate an API key. |
| 82 | +Set the key as an environment variable: |
| 83 | + |
| 84 | +``` |
| 85 | +export GOOGLE_AI_STUDIO_KEY=your-api-key-here |
| 86 | +``` |
| 87 | +</div> |
| 88 | +</details> |
| 89 | + |
| 90 | +<details> |
| 91 | +<summary>Specific guidance for Llamafile / llama.cpp server</summary> |
| 92 | +<div markdown="1"> |
| 93 | +If running a llamafile or llama.cpp server locally, configure the URL to point |
| 94 | +to your server. The URL must include the `/v1` path: |
| 95 | + |
| 96 | +``` |
| 97 | +export LLAMAFILE_URL=http://your-server-host:port/v1 |
| 98 | +``` |
| 99 | + |
| 100 | +If running on the same host as your containers, use `host.containers.internal` |
| 101 | +as the hostname (e.g., `http://host.containers.internal:7000/v1`). |
| 102 | + |
| 103 | +See also: [Container networking and self-hosted models](container-networking) |
| 104 | +</div> |
| 105 | +</details> |
| 106 | + |
| 107 | +<details> |
| 108 | +<summary>Specific guidance for LMStudio</summary> |
| 109 | +<div markdown="1"> |
| 110 | +If running LMStudio locally, configure the URL to point to your LMStudio server. |
| 111 | +LMStudio typically runs on port 1234: |
| 112 | + |
| 113 | +``` |
| 114 | +export LMSTUDIO_URL=http://your-server-host:1234 |
| 115 | +``` |
| 116 | + |
| 117 | +If running on the same host as your containers, use `host.containers.internal` |
| 118 | +as the hostname (e.g., `http://host.containers.internal:1234`). |
| 119 | + |
| 120 | +See also: [Container networking and self-hosted models](container-networking) |
| 121 | +</div> |
| 122 | +</details> |
| 123 | + |
| 124 | +<details> |
| 125 | +<summary>Specific guidance for Mistral AI</summary> |
| 126 | +<div markdown="1"> |
| 127 | +To use Mistral's API, sign up at [console.mistral.ai](https://console.mistral.ai/) |
| 128 | +and create an API key. Set the key as an environment variable: |
| 129 | + |
| 130 | +``` |
| 131 | +export MISTRAL_TOKEN=your-mistral-api-key-here |
| 132 | +``` |
| 133 | +</div> |
| 134 | +</details> |
| 135 | + |
| 136 | +<details> |
| 137 | +<summary>Specific guidance for Ollama</summary> |
| 138 | +<div markdown="1"> |
| 139 | +If running Ollama locally, configure the URL to point to your Ollama server. |
| 140 | +Ollama typically runs on port 11434: |
| 141 | + |
| 142 | +``` |
| 143 | +export OLLAMA_HOST=http://your-server-host:11434 |
| 144 | +``` |
| 145 | + |
| 146 | +If running on the same host as your containers, use `host.containers.internal` |
| 147 | +as the hostname (e.g., `http://host.containers.internal:11434`). |
| 148 | + |
| 149 | +See also: [Container networking and self-hosted models](container-networking) |
| 150 | +</div> |
| 151 | +</details> |
| 152 | + |
| 153 | +<details> |
| 154 | +<summary>Specific guidance for OpenAI</summary> |
| 155 | +<div markdown="1"> |
| 156 | +To use OpenAI's API, sign up at [platform.openai.com](https://platform.openai.com/) |
| 157 | +and create an API key. Set the key as an environment variable: |
| 158 | + |
| 159 | +``` |
| 160 | +export OPENAI_TOKEN=your-openai-api-key-here |
| 161 | +``` |
| 162 | + |
| 163 | +Many other services provide OpenAI-compatible APIs. You can use these by setting |
| 164 | +the `OPENAI_BASE_URL` environment variable to point to the alternative service: |
| 165 | + |
| 166 | +``` |
| 167 | +export OPENAI_BASE_URL=http://your-server-host:8000/v1 |
| 168 | +``` |
| 169 | +</div> |
| 170 | +</details> |
| 171 | + |
| 172 | +<details> |
| 173 | +<summary>Specific guidance for Google Cloud VertexAI</summary> |
| 174 | +<div markdown="1"> |
| 175 | +To use Google Cloud VertexAI, you need to create a service account with |
| 176 | +appropriate permissions and download its credentials file. |
| 177 | + |
| 178 | +1. In Google Cloud Console, create a service account |
| 179 | +2. Grant the service account permissions to invoke VertexAI models (e.g., |
| 180 | + `Vertex AI User` role - use minimal permissions, not admin roles) |
| 181 | +3. Create and download a JSON key file for the service account |
| 182 | +4. Save the key file as `vertexai/private.json` in your deployment directory |
| 183 | + |
| 184 | +{: .warning } |
| 185 | +**Important**: Service account credentials provide access to your Google Cloud |
| 186 | +resources. Never commit `private.json` to version control. Use minimal |
| 187 | +permissions - grant only what's needed for VertexAI model invocation, not |
| 188 | +administrator roles. |
| 189 | + |
| 190 | +After placing the file, you may need to adjust file permissions as described |
| 191 | +earlier in the configuration unpacking section: |
| 192 | + |
| 193 | +```sh |
| 194 | +chmod 644 vertexai/private.json |
| 195 | +``` |
| 196 | + |
| 197 | +On SELinux systems, also run: |
| 198 | + |
| 199 | +```sh |
| 200 | +sudo chcon -Rt svirt_sandbox_file_t vertexai/ |
| 201 | +``` |
| 202 | +</div> |
| 203 | +</details> |
| 204 | + |
| 205 | +<details> |
| 206 | +<summary>Specific guidance for vLLM</summary> |
| 207 | +<div markdown="1"> |
| 208 | +If running vLLM locally, configure the URL to point to your vLLM server. |
| 209 | +The URL should include the `/v1` path: |
| 210 | + |
| 211 | +``` |
| 212 | +export VLLM_URL=http://your-server-host:port/v1 |
| 213 | +``` |
| 214 | + |
| 215 | +If running on the same host as your containers, use `host.containers.internal` |
| 216 | +as the hostname (e.g., `http://host.containers.internal:8000/v1`). |
| 217 | + |
| 218 | +See also: [Container networking and self-hosted models](container-networking) |
| 219 | +</div> |
| 220 | +</details> |
0 commit comments