Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
7 changes: 1 addition & 6 deletions src/content/docs/workers-ai/features/batch-api/index.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -33,9 +33,4 @@ This will create a repository in your GitHub account and deploy a ready-to-use W

## Supported Models

- [@cf/meta/llama-3.3-70b-instruct-fp8-fast](/workers-ai/models/llama-3.3-70b-instruct-fp8-fast/)
- [@cf/baai/bge-small-en-v1.5](/workers-ai/models/bge-small-en-v1.5/)
- [@cf/baai/bge-base-en-v1.5](/workers-ai/models/bge-base-en-v1.5/)
- [@cf/baai/bge-large-en-v1.5](/workers-ai/models/bge-large-en-v1.5/)
- [@cf/baai/bge-m3](/workers-ai/models/bge-m3/)
- [@cf/meta/m2m100-1.2b](/workers-ai/models/m2m100-1.2b/)
Refer to our [model catalog](/workers-ai/models/?capabilities=Batch) for supported models.
13 changes: 1 addition & 12 deletions src/content/docs/workers-ai/features/fine-tunes/loras.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -17,18 +17,7 @@ Workers AI supports fine-tuned inference with adapters trained with [Low-Rank Ad

## Limitations

- We only support LoRAs for the following models (must not be quantized):

- `@cf/meta/llama-3.2-11b-vision-instruct`
- `@cf/meta/llama-3.3-70b-instruct-fp8-fast`
- `@cf/meta/llama-guard-3-8b`
- `@cf/meta/llama-3.1-8b-instruct-fast (soon)`
- `@cf/deepseek-ai/deepseek-r1-distill-qwen-32b`
- `@cf/qwen/qwen2.5-coder-32b-instruct`
- `@cf/qwen/qwq-32b`
- `@cf/mistralai/mistral-small-3.1-24b-instruct`
- `@cf/google/gemma-3-12b-it`

- We only support LoRAs for a [variety of models](/workers-ai/models/?capabilities=LoRA) (must not be quantized)
- Adapter must be trained with rank `r <=8` as well as larger ranks if up to 32. You can check the rank of a pre-trained LoRA adapter through the adapter's `config.json` file
- LoRA adapter file must be < 300MB
- LoRA adapter files must be named `adapter_config.json` and `adapter_model.safetensors` exactly
Expand Down