Skip to content

Commit 6043b49

Browse files
authored
Merge pull request #1506 from vandijm/local-llm-processor
Local llm processor
2 parents fac3b01 + d756b0e commit 6043b49

File tree

5 files changed

+7
-8
lines changed

5 files changed

+7
-8
lines changed

cloud-infrastructure/private-cloud-and-edge/compute-cloud-at-customer/local-llm/README.md

Lines changed: 7 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
*Last Update: 23 November 2024*
1+
*Last Update: 27 November 2024*
22

33
<br><h1 align="center">Local LLM Inferencing and Interaction<br>Using the Ollama Open Source Tool</h1>
44
<p align="center"><img align="centre" src="./images/ollama-logo.png" width="10%" style="float:right"/></p>
@@ -65,8 +65,7 @@ Create a VM in a public subnet following these guidelines:
6565
10. Ensure that the VM is accessible via `ssh`
6666
11. Configure the proxy setup if required (described below)
6767
12. Update your local host's `/etc/hosts` file to reflect your public IP address for `llm-host`
68-
13. Should you have a proxy'd network follow the instruction in the "Proxy Settings" section below prior to performing the next step
69-
14. Perform an OS update in `llm-host` before proceeding:
68+
13. Perform an OS update in `llm-host` before proceeding:
7069

7170
```
7271
sudo dnf update
@@ -267,16 +266,16 @@ ollama help serve
267266

268267
Download and test your first LLM (and you will notice the population of `/mnt/llm-repo` with data by running `ls -lR /mnt/llm-repo`):
269268

270-
<p><img src="./images/ollama-pull-and-test.png" title="Ollama pull/test Llama3.2" width="75%" style="float:right"/></p>
269+
<p><img src="./images/ollama-pull-and-test.png" title="Ollama pull/test mistral" width="75%" style="float:right"/></p>
271270

272271
Run some more tests from your client to test the APIs:
273272

274273
```
275274
$ curl http://llm-host:11434/api/tags
276275
$ curl http://llm-host:11434/api/ps
277276
$ curl -X POST http://llm-host:11434/api/generate -d '{
278-
"model": "llama3.2",
279-
"prompt":"Hello Llama3.2!",
277+
"model": "mistral",
278+
"prompt":"Hello Mistral!",
280279
"stream": false
281280
}'
282281
```
@@ -294,8 +293,8 @@ Install any of the GUI clients mentioned previously and test the connectivity an
294293
1. Create a Remote Models Provider
295294
2. Name it appropriately
296295
3. The Service Endpoint is `http://llm-host:11434`
297-
4. "Fetch Models" (that are already installed, in this case `llama3.2`)
298-
5. This step can be repeated as new models are added
296+
4. "Fetch Models" (that are already installed, in this case `mistral`)
297+
5. This step can be repeated as new models are added from the Ollama model repository
299298

300299
Example output as follows:
301300

30.1 KB
Loading
Loading
-179 KB
Loading
-14.9 KB
Loading

0 commit comments

Comments
 (0)