1
- * Last Update: 23 November 2024*
1
+ * Last Update: 27 November 2024*
2
2
3
3
<br ><h1 align =" center " >Local LLM Inferencing and Interaction<br >Using the Ollama Open Source Tool</h1 >
4
4
<p align =" center " ><img align =" centre " src =" ./images/ollama-logo.png " width =" 10% " style =" float :right " /></p >
@@ -65,8 +65,7 @@ Create a VM in a public subnet following these guidelines:
65
65
10 . Ensure that the VM is accessible via ` ssh `
66
66
11 . Configure the proxy setup if required (described below)
67
67
12 . Update your local host's ` /etc/hosts ` file to reflect your public IP address for ` llm-host `
68
- 13 . Should you have a proxy'd network follow the instruction in the "Proxy Settings" section below prior to performing the next step
69
- 14 . Perform an OS update in ` llm-host ` before proceeding:
68
+ 13 . Perform an OS update in ` llm-host ` before proceeding:
70
69
71
70
```
72
71
sudo dnf update
@@ -267,16 +266,16 @@ ollama help serve
267
266
268
267
Download and test your first LLM (and you will notice the population of ` /mnt/llm-repo ` with data by running ` ls -lR /mnt/llm-repo ` ):
269
268
270
- <p ><img src =" ./images/ollama-pull-and-test.png " title =" Ollama pull/test Llama3.2 " width =" 75% " style =" float :right " /></p >
269
+ <p ><img src =" ./images/ollama-pull-and-test.png " title =" Ollama pull/test mistral " width =" 75% " style =" float :right " /></p >
271
270
272
271
Run some more tests from your client to test the APIs:
273
272
274
273
```
275
274
$ curl http://llm-host:11434/api/tags
276
275
$ curl http://llm-host:11434/api/ps
277
276
$ curl -X POST http://llm-host:11434/api/generate -d '{
278
- "model": "llama3.2 ",
279
- "prompt":"Hello Llama3.2 !",
277
+ "model": "mistral ",
278
+ "prompt":"Hello Mistral !",
280
279
"stream": false
281
280
}'
282
281
```
@@ -294,8 +293,8 @@ Install any of the GUI clients mentioned previously and test the connectivity an
294
293
1 . Create a Remote Models Provider
295
294
2 . Name it appropriately
296
295
3 . The Service Endpoint is ` http://llm-host:11434 `
297
- 4 . "Fetch Models" (that are already installed, in this case ` llama3.2 ` )
298
- 5 . This step can be repeated as new models are added
296
+ 4 . "Fetch Models" (that are already installed, in this case ` mistral ` )
297
+ 5 . This step can be repeated as new models are added from the Ollama model repository
299
298
300
299
Example output as follows:
301
300
0 commit comments