Skip to content
This repository was archived by the owner on May 20, 2025. It is now read-only.

Commit c26ba87

Browse files
committed
fix minor spelling issues.
1 parent 5906648 commit c26ba87

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

docs/guides/python/serverless-llama.mdx

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -49,7 +49,7 @@ We'll use a quantized version of the lightweight Llama 1B model, specifically [L
4949
accuracy of the model.
5050
</Note>
5151

52-
The [LM Studio](https://lmstudio.ai) team provide several quantized versions of Llama 3.2 1B on [Hugging Face](https://huggingface.co/bartowski/Llama-3.2-1B-Instruct-GGUF). Consider trying different versions to find one that best fits your needs (e.g. `Q5_K_M` which is slightly larger but higher quality).
52+
The [LM Studio](https://lmstudio.ai) team provides several quantized versions of Llama 3.2 1B on [Hugging Face](https://huggingface.co/bartowski/Llama-3.2-1B-Instruct-GGUF). Consider trying different versions to find one that best fits your needs (e.g. `Q5_K_M` which is slightly larger but higher quality).
5353

5454
Let's download the chosen model and save it in a `models` directory in your project.
5555

@@ -291,4 +291,4 @@ At this point, we demonstrated how you can use a lightweight model like Llama 3.
291291

292292
As you've seen in the code example, we've setup a fairly basic prompt structure, but you can expand on this to include more complex prompts. Including system prompts that help restrict/guide the model's responses, or even more complex interactions with the model. Also, in this example we expose the model directly as an API, but this limits the response time to 30 seconds on AWS with API Gateway.
293293

294-
In future guides we'll show how you can go beyond simple one-time responses to more complex interactions, such as maintaining context between requests. We can also include Websockets and streamed responses to provider a better user experience for larger responses.
294+
In future guides we'll show how you can go beyond simple one-time responses to more complex interactions, such as maintaining context between requests. We can also include Websockets and streamed responses to provide a better user experience for larger responses.

0 commit comments

Comments
 (0)