Skip to content

Commit 6e726be

Browse files
Apply suggestions from code review
Co-authored-by: Jessica <[email protected]>
1 parent f7ca636 commit 6e726be

File tree

1 file changed

+5
-5
lines changed

1 file changed

+5
-5
lines changed

faq/managed-inference.mdx

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ All models are currently hosted in a secure data center located in Paris, France
2222
- **Managed Inference**: Allows deployment of curated or custom models with chosen quantization and instances, offering predictable throughput and enhanced security features like private network isolation and access control. Managed Inference is billed by hourly usage, whether provisioned capacity is receiving traffic or not.
2323
- **Generative APIs**: A serverless service providing access to pre-configured AI models via API, billed per token usage.
2424

25-
## Where can I find information regarding the data, privacy and security policies applied to Scaleway's AI services?
25+
## Where can I find information regarding the data, privacy, and security policies applied to Scaleway's AI services?
2626
You can find detailed information regarding the policies applied to Scaleway's AI services in our [Data, privacy, and security for Scaleway's AI services](/managed-inference/reference-content/data-privacy-security-scaleway-ai-services/) documentation.
2727

2828
## Is Managed Inference compatible with Open AI APIs?
@@ -33,7 +33,7 @@ We are currently working on defining our SLAs for Managed Inference. We will pro
3333

3434
## What are the performance guarantees (vs. Generative APIs)?
3535
Managed Inference provides dedicated resources, ensuring predictable performance and lower latency compared to Generative APIs, which are a shared, serverless offering optimized for infrequent traffic with moderate peak loads. Managed Inference is ideal for workloads that require consistent response times, high availability, custom hardware configurations or generate extreme peak loads during a narrow period of time.
36-
Compared to Generative APIs, no usage quota are applied on the number of tokens per second generated, since the output is limited by the GPU Instances size and number of your Managed Inference Deployment.
36+
Compared to Generative APIs, no usage quota is applied to the number of tokens per second generated, since the output is limited by the GPU Instances size and number of your Managed Inference Deployment.
3737

3838
## What types of models can I deploy with Managed Inference?
3939
You can deploy a variety of models, including:
@@ -69,7 +69,7 @@ Yes, Managed Inference is designed for low-latency, high-throughput applications
6969
## Can I use Managed Inference with other Scaleway services?
7070
Absolutely. Managed Inference integrates seamlessly with other Scaleway services, such as [Object Storage](/object-storage/quickstart/) for model hosting, [Kubernetes](/kubernetes/quickstart/) for containerized applications, and [Scaleway IAM](/iam/quickstart/) for access management.
7171

72-
## Do model license apply when using Managed Inference?
73-
Yes, model licenses need to be complied with when using Managed Inference. Applicable licenses are available for [each model in our documentation](https://www.scaleway.com/en/docs/managed-inference/reference-content/).
72+
## Do model licenses apply when using Managed Inference?
73+
Yes, model licenses need to be complied with when using Managed Inference. Applicable licenses are available for [each model in our documentation](/managed-inference/reference-content/).
7474
- For models provided in the Scaleway catalog, you need to accept licenses (including potential EULA) before creating any Managed Inference deployment.
75-
- For custom models you choose to import on Scaleway, you are responsible to comply with model licenses (as with any software you choose to install on a GPU Instance for example).
75+
- For custom models you choose to import on Scaleway, you are responsible for complying with model licenses (as with any software you choose to install on a GPU Instance for example).

0 commit comments

Comments
 (0)