Skip to content

Commit 0da7316

Browse files
authored
fix(ai): Updating documentations according to console copy (#4062)
1 parent b03f10a commit 0da7316

File tree

3 files changed

+9
-9
lines changed

3 files changed

+9
-9
lines changed

ai-data/managed-inference/concepts.mdx

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ content:
77
paragraph: This page explains all the concepts related to Managed Inference
88
tags:
99
dates:
10-
validation: 2024-09-10
10+
validation: 2024-11-29
1111
categories:
1212
- ai-data
1313
---
@@ -68,19 +68,19 @@ LLM Applications are applications or software tools that leverage the capabiliti
6868
## Large Language Models
6969

7070
LLMs are advanced artificial intelligence systems capable of understanding and generating human-like text on various topics.
71-
These models, such as Llama-2, are trained on vast amounts of data to learn the patterns and structures of language, enabling them to generate coherent and contextually relevant responses to queries or prompts.
71+
These models, such as Llama-3, are trained on vast amounts of data to learn the patterns and structures of language, enabling them to generate coherent and contextually relevant responses to queries or prompts.
7272
LLMs have applications in natural language processing, text generation, translation, and other tasks requiring sophisticated language understanding and production.
7373

7474
## Prompt
7575

76-
In the context of LLMs, a prompt refers to the input provided to the model to generate a desired response.
76+
In the context of generative AI models, a prompt refers to the input provided to the model to generate a desired response.
7777
It typically consists of a sentence, paragraph, or series of keywords or instructions that guide the model in producing text relevant to the given context or task.
7878
The quality and specificity of the prompt greatly influence the generated output, as the model uses it to understand the user's intent and create responses accordingly.
7979

8080
## Quantization
8181

8282
Quantization is a technique used to reduce the precision of numerical values in a model's parameters or activations to improve efficiency and reduce memory footprint during inference. It involves representing floating-point values with fewer bits while minimizing the loss of accuracy.
83-
LLMs provided for deployment are named with suffixes that denote their quantization levels, such as `:int8`, `:fp8`, and `:fp16`.
83+
AI models provided for deployment are named with suffixes that denote their quantization levels, such as `:int8`, `:fp8`, and `:fp16`.
8484

8585
## Retrieval Augmented Generation (RAG)
8686

ai-data/managed-inference/how-to/create-deployment.mdx

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -26,11 +26,11 @@ dates:
2626
- Choose the geographical **region** for the deployment.
2727
- Specify the GPU Instance type to be used with your deployment.
2828
4. Enter a **name** for the deployment, and optional tags.
29-
5. Configure the **network** settings for the deployment:
30-
- Enable **Private Network** for secure communication and restricted availability within Private Networks. Choose an existing Private Network from the drop-down list, or create a new one.
31-
- Enable **Public Network** to access resources via the public internet. Token protection is enabled by default.
29+
5. Configure the **network connectivity** settings for the deployment:
30+
- Attach to a **Private Network** for secure communication and restricted availability. Choose an existing Private Network from the drop-down list, or create a new one.
31+
- Set up **Public connectivity** to access resources via the public internet. Authentication by API key is enabled by default.
3232
<Message type="important">
33-
- Enabling both private and public networks will result in two distinct endpoints (public and private) for your deployment.
33+
- Enabling both private and public connectivity will result in two distinct endpoints (public and private) for your deployment.
3434
- Deployments must have at least one endpoint, either public or private.
3535
</Message>
3636
6. Click **Deploy model** to launch the deployment process. Once the model is ready, it will be listed among your deployments.

ai-data/managed-inference/index.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ meta:
77
<ProductHeader
88
productName="Managed Inference"
99
productLogo="inference"
10-
description="Dive into seamless language processing with our easy-to-use LLM endpoints. Perfect for everything from data analysis to creative tasks, all with clear pricing."
10+
description="Effortlessly deploy AI models on a sovereign infrastructure, manage and scale inference with full data privacy. Start now with a simple interface for creating dedicated endpoints."
1111
url="/ai-data/managed-inference/quickstart"
1212
label="Managed Inference Quickstart"
1313
/>

0 commit comments

Comments
 (0)