Skip to content

Commit 86635f5

Browse files
authored
feat(genapi): update faq
1 parent cbf1cb8 commit 86635f5

File tree

1 file changed

+4
-0
lines changed

1 file changed

+4
-0
lines changed

pages/managed-inference/faq.mdx

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -60,6 +60,10 @@ You can select the Instance type based on your model’s computational needs and
6060
Billing is based on the Instance type and usage duration. Unlike [Generative APIs](/generative-apis/quickstart/), which are billed per token, Managed Inference provides predictable costs based on the allocated infrastructure.
6161
Pricing details can be found on the [Scaleway pricing page](https://www.scaleway.com/en/pricing/model-as-a-service/#managed-inference).
6262

63+
## Can I pause Managed Inference billing when the instance is not in use ?
64+
When a Managed Inference deployment is running, corresponding resources are provisioned and thus billed. Resources can therefore not be paused.
65+
However, you can still optimize your Managed Inference deployment to fit within specific time range (such as during working hours). To do so, you can automate deployment creation and deletion using [Managed Inference API](https://www.scaleway.com/en/developers/api/inference/), [Terraform](https://registry.terraform.io/providers/scaleway/scaleway/latest/docs/resources/inference_deployment) or [Scaleway SDKs](https://www.scaleway.com/en/docs/scaleway-sdk/). These actions can be programmed using [Serverless Jobs]() to be performed periodically in an automated way.
66+
6367
## Can I run inference on private models?
6468
Yes, Managed Inference allows you to deploy private models with access control settings. You can restrict access to specific users, teams, or networks.
6569

0 commit comments

Comments
 (0)