1- * Last Update: 23 November 2024*
1+ * Last Update: 27 November 2024*
22
33<br ><h1 align =" center " >Local LLM Inferencing and Interaction<br >Using the Ollama Open Source Tool</h1 >
44<p align =" center " ><img align =" centre " src =" ./images/ollama-logo.png " width =" 10% " style =" float :right " /></p >
@@ -65,8 +65,7 @@ Create a VM in a public subnet following these guidelines:
656510 . Ensure that the VM is accessible via ` ssh `
666611 . Configure the proxy setup if required (described below)
676712 . Update your local host's ` /etc/hosts ` file to reflect your public IP address for ` llm-host `
68- 13 . Should you have a proxy'd network follow the instruction in the "Proxy Settings" section below prior to performing the next step
69- 14 . Perform an OS update in ` llm-host ` before proceeding:
68+ 13 . Perform an OS update in ` llm-host ` before proceeding:
7069
7170```
7271sudo dnf update
@@ -267,16 +266,16 @@ ollama help serve
267266
268267Download and test your first LLM (and you will notice the population of ` /mnt/llm-repo ` with data by running ` ls -lR /mnt/llm-repo ` ):
269268
270- <p ><img src =" ./images/ollama-pull-and-test.png " title =" Ollama pull/test Llama3.2 " width =" 75% " style =" float :right " /></p >
269+ <p ><img src =" ./images/ollama-pull-and-test.png " title =" Ollama pull/test mistral " width =" 75% " style =" float :right " /></p >
271270
272271Run some more tests from your client to test the APIs:
273272
274273```
275274$ curl http://llm-host:11434/api/tags
276275$ curl http://llm-host:11434/api/ps
277276$ curl -X POST http://llm-host:11434/api/generate -d '{
278- "model": "llama3.2 ",
279- "prompt":"Hello Llama3.2 !",
277+ "model": "mistral ",
278+ "prompt":"Hello Mistral !",
280279 "stream": false
281280 }'
282281```
@@ -294,8 +293,8 @@ Install any of the GUI clients mentioned previously and test the connectivity an
2942931 . Create a Remote Models Provider
2952942 . Name it appropriately
2962953 . The Service Endpoint is ` http://llm-host:11434 `
297- 4 . "Fetch Models" (that are already installed, in this case ` llama3.2 ` )
298- 5 . This step can be repeated as new models are added
296+ 4 . "Fetch Models" (that are already installed, in this case ` mistral ` )
297+ 5 . This step can be repeated as new models are added from the Ollama model repository
299298
300299Example output as follows:
301300
0 commit comments