Skip to content

Commit ef483fe

Browse files
daisyfaithaumathatsKevinJain
authored andcommitted
[Workers AI]loras limits (#21608)
* loras limits * Update src/content/docs/workers-ai/features/fine-tunes/loras.mdx Co-authored-by: Kevin Jain <[email protected]> --------- Co-authored-by: Kevin Jain <[email protected]>
1 parent 89f1408 commit ef483fe

File tree

1 file changed

+13
-6
lines changed
  • src/content/docs/workers-ai/features/fine-tunes

1 file changed

+13
-6
lines changed

src/content/docs/workers-ai/features/fine-tunes/loras.mdx

Lines changed: 13 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -18,12 +18,19 @@ Workers AI supports fine-tuned inference with adapters trained with [Low-Rank Ad
1818
## Limitations
1919

2020
- We only support LoRAs for the following models (must not be quantized):
21-
- `@cf/meta-llama/llama-2-7b-chat-hf-lora`
22-
- `@cf/mistral/mistral-7b-instruct-v0.2-lora`
23-
- `@cf/google/gemma-2b-it-lora`
24-
- `@cf/google/gemma-7b-it-lora`
25-
- Adapter must be trained with rank `r <=8`. You can check the rank of a pre-trained LoRA adapter through the adapter's `config.json` file
26-
- LoRA adapter file must be < 100MB
21+
22+
- `@cf/meta/llama-3.2-11b-vision-instruct`
23+
- `@cf/meta/llama-3.3-70b-instruct-fp8-fast`
24+
- `@cf/meta/llama-guard-3-8b`
25+
- `@cf/meta/llama-3.1-8b-instruct-fast (soon)`
26+
- `@cf/deepseek-ai/deepseek-r1-distill-qwen-32b`
27+
- `@cf/qwen/qwen2.5-coder-32b-instruct`
28+
- `@cf/qwen/qwq-32b`
29+
- `@cf/mistralai/mistral-small-3.1-24b-instruct`
30+
- `@cf/google/gemma-3-12b-it`
31+
32+
- Adapter must be trained with rank `r <=8` as well as larger ranks if up to 32. You can check the rank of a pre-trained LoRA adapter through the adapter's `config.json` file
33+
- LoRA adapter file must be < 300MB
2734
- LoRA adapter files must be named `adapter_config.json` and `adapter_model.safetensors` exactly
2835
- You can test up to 30 LoRA adapters per account
2936

0 commit comments

Comments
 (0)