You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/hub/advanced-compute-options.md
-59Lines changed: 0 additions & 59 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -31,62 +31,3 @@ ZeroGPU is a dynamic GPU allocation system that optimizes AI deployment on Huggi
31
31
-**Enterprise-Grade Infrastructure**: Access to high-performance NVIDIA A100 GPUs with 40GB VRAM per workload
32
32
33
33
[Learn more about ZeroGPU →](https://huggingface.co/docs/hub/spaces-zerogpu)
34
-
35
-
## Train on NVIDIA DGX Cloud
36
-
37
-
Train on NVIDIA DGX Cloud offers a simple no-code training job creation experience powered by Hugging Face AutoTrain and Hugging Face Spaces. Instantly access NVIDIA GPUs and avoid the time-consuming work of writing, testing, and debugging training scripts for AI models.
38
-
39
-
### How it works
40
-
41
-
Read the [blogpost for Train on NVIDIA DGX Cloud](https://huggingface.co/blog/train-dgx-cloud#how-it-works).
42
-
43
-
### Supported architectures
44
-
45
-
#### Transformers
46
-
47
-
| Architecture |
48
-
| ------------ |
49
-
| Llama |
50
-
| Falcon |
51
-
| Mistral |
52
-
| Mixtral |
53
-
| T5 |
54
-
| gemma |
55
-
56
-
#### Diffusers
57
-
58
-
| Architecture |
59
-
| ------------------- |
60
-
| Stable Diffusion |
61
-
| Stable Diffusion XL |
62
-
63
-
### Pricing
64
-
65
-
Usage of Train on NVIDIA DGX Cloud is billed by the minute of the GPU instances used during your training jobs. Usage fees accrue to your Enterprise Hub Organizations’ current monthly billing cycle, once a job is completed. You can check your current and past usage at any time within the billing settings of your Enterprise Hub Organization.
66
-
67
-
| NVIDIA GPU | GPU Memory | On-Demand Price/hr |
68
-
| ----------- | ---------- | ------------------ |
69
-
| NVIDIA L40S | 48GB | $2.75 |
70
-
| NVIDIA H100 | 80GB | $8.25 |
71
-
72
-
## NVIDIA NIM API (serverless)
73
-
74
-
NVIDIA NIM API (serverless) offers access to [NVIDIA Inference Microservices (NIM)](https://www.nvidia.com/en-us/ai/) powered by NVIDIA H100s in a serverless way. Use standardized APIs and a few lines of code to run inference in a pay-as-you-go pricing model.
75
-
76
-
### How it works
77
-
78
-
Read the [blogpost for Serverless Inference with Hugging Face and NVIDIA NIMs](https://huggingface.co/blog/inference-dgx-cloud#how-it-works).
79
-
80
-
### Supported models
81
-
82
-
You can find all supported models in [this NVIDIA Collection](https://huggingface.co/collections/nvidia/nim-66a3c6fcdcb5bbc6e975b508).
83
-
84
-
### Pricing
85
-
86
-
Usage of NVIDIA NIM API (serverless) is billed based on the compute time spent per request. Usage fees accrue to your Enterprise Hub Organizations’ current monthly billing cycle, once a job is completed. You can check your current and past usage at any time within the billing settings of your Enterprise Hub Organization.
87
-
88
-
| NVIDIA GPU | GPU Memory | On-Demand Price/hr |
89
-
| ----------- | ---------- | ------------------ |
90
-
| NVIDIA H100 | 80GB | $8.25 |
91
-
92
-
The total cost for a request will depend on the model size, the number of GPUs required, and the time taken to process the request. For each model, you can find which hardware configuration is used in the notes of [this NVIDIA Collection](https://huggingface.co/collections/nvidia/nim-66a3c6fcdcb5bbc6e975b508).
0 commit comments