Skip to content

Commit c7adb2f

Browse files
committed
fix(models): Update available models documentation and add new Arcee AI models with detailed descriptions
1 parent 6fe99d7 commit c7adb2f

File tree

2 files changed

+275
-4
lines changed

2 files changed

+275
-4
lines changed

docs/guides/available-models.md

Lines changed: 15 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -46,10 +46,10 @@ For more information about working with models, see the [Working with Models](/g
4646
## Models by Type
4747
{: .d-inline-block }
4848

49-
Last updated: 2025-05-05
49+
Last updated: 2025-05-06
5050
{: .label .label-green }
5151

52-
### Chat Models (458)
52+
### Chat Models (464)
5353

5454
| ID | Type | Name | Provider | Context | MaxTok | Family | In$/M | Out$/M |
5555
| :-- | :-- | :-- | :-- | --: | --: | :-- | --: | --: |
@@ -238,6 +238,12 @@ Last updated: 2025-05-05
238238
| anthropic/claude-3.7-sonnet:beta | chat | Anthropic: Claude 3.7 Sonnet (self-moderated) | openrouter | 200000 | 128000 | other | 3.00 | 15.00 |
239239
| anthropic/claude-3.7-sonnet:thinking | chat | Anthropic: Claude 3.7 Sonnet (thinking) | openrouter | 200000 | 64000 | other | 3.00 | 15.00 |
240240
| arcee-ai/arcee-blitz | chat | Arcee AI: Arcee Blitz | openrouter | 32768 | | other | 0.45 | 0.75 |
241+
| arcee-ai/caller-large | chat | Arcee AI: Caller Large | openrouter | 32768 | | other | 0.55 | 0.85 |
242+
| arcee-ai/coder-large | chat | Arcee AI: Coder Large | openrouter | 32768 | | other | 0.50 | 0.80 |
243+
| arcee-ai/maestro-reasoning | chat | Arcee AI: Maestro Reasoning | openrouter | 131072 | | other | 0.90 | 3.30 |
244+
| arcee-ai/spotlight | chat | Arcee AI: Spotlight | openrouter | 131072 | | other | 0.18 | 0.18 |
245+
| arcee-ai/virtuoso-large | chat | Arcee AI: Virtuoso Large | openrouter | 131072 | | other | 0.75 | 1.20 |
246+
| arcee-ai/virtuoso-medium-v2 | chat | Arcee AI: Virtuoso Medium V2 | openrouter | 131072 | | other | 0.50 | 0.80 |
241247
| arliai/qwq-32b-arliai-rpr-v1:free | chat | ArliAI: QwQ 32B RpR v1 (free) | openrouter | 32768 | | other | 0.00 | 0.00 |
242248
| bytedance-research/ui-tars-72b:free | chat | Bytedance: UI-TARS 72B (free) | openrouter | 32768 | | other | 0.00 | 0.00 |
243249
| cognitivecomputations/dolphin-mixtral-8x22b | chat | Dolphin 2.9.2 Mixtral 8x22B 🐬 | openrouter | 16000 | | other | 0.90 | 0.90 |
@@ -740,7 +746,7 @@ Last updated: 2025-05-05
740746
| anthropic.claude-v2:1:200k | chat | Claude | bedrock | 200000 | 4096 | claude2 | 8.00 | 24.00 |
741747
| us.anthropic.claude-3-7-sonnet-20250219-v1:0 | chat | Claude 3.7 Sonnet | bedrock | 200000 | 4096 | claude3_sonnet | 3.00 | 15.00 |
742748

743-
### Openrouter Models (319)
749+
### Openrouter Models (325)
744750

745751
| ID | Type | Name | Provider | Context | MaxTok | Family | In$/M | Out$/M |
746752
| :-- | :-- | :-- | :-- | --: | --: | :-- | --: | --: |
@@ -790,6 +796,12 @@ Last updated: 2025-05-05
790796
| anthropic/claude-3.7-sonnet:beta | chat | Anthropic: Claude 3.7 Sonnet (self-moderated) | openrouter | 200000 | 128000 | other | 3.00 | 15.00 |
791797
| anthropic/claude-3.7-sonnet:thinking | chat | Anthropic: Claude 3.7 Sonnet (thinking) | openrouter | 200000 | 64000 | other | 3.00 | 15.00 |
792798
| arcee-ai/arcee-blitz | chat | Arcee AI: Arcee Blitz | openrouter | 32768 | | other | 0.45 | 0.75 |
799+
| arcee-ai/caller-large | chat | Arcee AI: Caller Large | openrouter | 32768 | | other | 0.55 | 0.85 |
800+
| arcee-ai/coder-large | chat | Arcee AI: Coder Large | openrouter | 32768 | | other | 0.50 | 0.80 |
801+
| arcee-ai/maestro-reasoning | chat | Arcee AI: Maestro Reasoning | openrouter | 131072 | | other | 0.90 | 3.30 |
802+
| arcee-ai/spotlight | chat | Arcee AI: Spotlight | openrouter | 131072 | | other | 0.18 | 0.18 |
803+
| arcee-ai/virtuoso-large | chat | Arcee AI: Virtuoso Large | openrouter | 131072 | | other | 0.75 | 1.20 |
804+
| arcee-ai/virtuoso-medium-v2 | chat | Arcee AI: Virtuoso Medium V2 | openrouter | 131072 | | other | 0.50 | 0.80 |
793805
| arliai/qwq-32b-arliai-rpr-v1:free | chat | ArliAI: QwQ 32B RpR v1 (free) | openrouter | 32768 | | other | 0.00 | 0.00 |
794806
| bytedance-research/ui-tars-72b:free | chat | Bytedance: UI-TARS 72B (free) | openrouter | 32768 | | other | 0.00 | 0.00 |
795807
| cognitivecomputations/dolphin-mixtral-8x22b | chat | Dolphin 2.9.2 Mixtral 8x22B 🐬 | openrouter | 16000 | | other | 0.90 | 0.90 |

lib/ruby_llm/models.json

Lines changed: 260 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5634,7 +5634,7 @@
56345634
"output_price_per_million": 0.75,
56355635
"metadata": {
56365636
"object": null,
5637-
"description": "Arcee-Blitz (24B) is a 24B-parameter instruct model built on Mistral-Small-24B and distilled from DeepSeek-V3. It emphasizes speed and practicality, with strong performance across general tasks and notable gains in world knowledge benchmarks. The model uses DeepSeek logits as pretraining targets and underwent additional fine-tuning to improve instruction-following and reasoning capabilities.\n\nBenchmarks show consistent improvements over Mistral-Small-3, particularly on MMLU-Pro, Math, and MixEval. Arcee-Blitz supports a 32k context length and is released under the Apache 2.0 license.",
5637+
"description": "Arcee Blitz is a 24 B‑parameter dense model distilled from DeepSeek and built on Mistral architecture for \"everyday\" chat. The distillation‑plus‑refinement pipeline trims compute while keeping DeepSeek‑style reasoning, so Blitz punches above its weight on MMLU, GSM‑8K and BBH compared with other mid‑size open models. With a default 128 k context window and competitive throughput, it serves as a cost‑efficient workhorse for summarization, brainstorming and light code help. Internally, Arcee uses Blitz as the default writer in Conductor pipelines when the heavier Virtuoso line is not required. Users therefore get near‑70 B quality at ~⅓ the latency and price. ",
56385638
"architecture": {
56395639
"modality": "text->text",
56405640
"input_modalities": [
@@ -5661,6 +5661,265 @@
56615661
}
56625662
}
56635663
},
5664+
{
5665+
"id": "arcee-ai/caller-large",
5666+
"created_at": "2025-05-06T01:31:09+02:00",
5667+
"display_name": "Arcee AI: Caller Large",
5668+
"provider": "openrouter",
5669+
"context_window": 32768,
5670+
"max_tokens": null,
5671+
"type": "chat",
5672+
"family": "other",
5673+
"supports_vision": false,
5674+
"supports_functions": true,
5675+
"supports_json_mode": true,
5676+
"input_price_per_million": 0.55,
5677+
"output_price_per_million": 0.85,
5678+
"metadata": {
5679+
"object": null,
5680+
"description": "Caller Large is Arcee's specialist \"function‑calling\" SLM built to orchestrate external tools and APIs. Instead of maximizing next‑token accuracy, training focuses on structured JSON outputs, parameter extraction and multi‑step tool chains, making Caller a natural choice for retrieval‑augmented generation, robotic process automation or data‑pull chatbots. It incorporates a routing head that decides when (and how) to invoke a tool versus answering directly, reducing hallucinated calls. The model is already the backbone of Arcee Conductor's auto‑tool mode, where it parses user intent, emits clean function signatures and hands control back once the tool response is ready. Developers thus gain an OpenAI‑style function‑calling UX without handing requests to a frontier‑scale model. ",
5681+
"architecture": {
5682+
"modality": "text->text",
5683+
"input_modalities": [
5684+
"text"
5685+
],
5686+
"output_modalities": [
5687+
"text"
5688+
],
5689+
"tokenizer": "Other",
5690+
"instruct_type": null
5691+
},
5692+
"pricing": {
5693+
"prompt": "0.00000055",
5694+
"completion": "0.00000085",
5695+
"request": "0",
5696+
"image": "0",
5697+
"web_search": "0",
5698+
"internal_reasoning": "0"
5699+
},
5700+
"top_provider": {
5701+
"context_length": 32768,
5702+
"max_completion_tokens": null,
5703+
"is_moderated": false
5704+
}
5705+
}
5706+
},
5707+
{
5708+
"id": "arcee-ai/coder-large",
5709+
"created_at": "2025-05-05T22:57:43+02:00",
5710+
"display_name": "Arcee AI: Coder Large",
5711+
"provider": "openrouter",
5712+
"context_window": 32768,
5713+
"max_tokens": null,
5714+
"type": "chat",
5715+
"family": "other",
5716+
"supports_vision": false,
5717+
"supports_functions": true,
5718+
"supports_json_mode": true,
5719+
"input_price_per_million": 0.5,
5720+
"output_price_per_million": 0.7999999999999999,
5721+
"metadata": {
5722+
"object": null,
5723+
"description": "Coder‑Large is a 32 B‑parameter offspring of Qwen 2.5‑Instruct that has been further trained on permissively‑licensed GitHub, CodeSearchNet and synthetic bug‑fix corpora. It supports a 32k context window, enabling multi‑file refactoring or long diff review in a single call, and understands 30‑plus programming languages with special attention to TypeScript, Go and Terraform. Internal benchmarks show 5–8 pt gains over CodeLlama‑34 B‑Python on HumanEval and competitive BugFix scores thanks to a reinforcement pass that rewards compilable output. The model emits structured explanations alongside code blocks by default, making it suitable for educational tooling as well as production copilot scenarios. Cost‑wise, Together AI prices it well below proprietary incumbents, so teams can scale interactive coding without runaway spend. ",
5724+
"architecture": {
5725+
"modality": "text->text",
5726+
"input_modalities": [
5727+
"text"
5728+
],
5729+
"output_modalities": [
5730+
"text"
5731+
],
5732+
"tokenizer": "Other",
5733+
"instruct_type": null
5734+
},
5735+
"pricing": {
5736+
"prompt": "0.0000005",
5737+
"completion": "0.0000008",
5738+
"request": "0",
5739+
"image": "0",
5740+
"web_search": "0",
5741+
"internal_reasoning": "0"
5742+
},
5743+
"top_provider": {
5744+
"context_length": 32768,
5745+
"max_completion_tokens": null,
5746+
"is_moderated": false
5747+
}
5748+
}
5749+
},
5750+
{
5751+
"id": "arcee-ai/maestro-reasoning",
5752+
"created_at": "2025-05-05T23:41:09+02:00",
5753+
"display_name": "Arcee AI: Maestro Reasoning",
5754+
"provider": "openrouter",
5755+
"context_window": 131072,
5756+
"max_tokens": null,
5757+
"type": "chat",
5758+
"family": "other",
5759+
"supports_vision": false,
5760+
"supports_functions": true,
5761+
"supports_json_mode": true,
5762+
"input_price_per_million": 0.8999999999999999,
5763+
"output_price_per_million": 3.3000000000000003,
5764+
"metadata": {
5765+
"object": null,
5766+
"description": "Maestro Reasoning is Arcee's flagship analysis model: a 32 B‑parameter derivative of Qwen 2.5‑32 B tuned with DPO and chain‑of‑thought RL for step‑by‑step logic. Compared to the earlier 7 B preview, the production 32 B release widens the context window to 128 k tokens and doubles pass‑rate on MATH and GSM‑8K, while also lifting code completion accuracy. Its instruction style encourages structured \"thought → answer\" traces that can be parsed or hidden according to user preference. That transparency pairs well with audit‑focused industries like finance or healthcare where seeing the reasoning path matters. In Arcee Conductor, Maestro is automatically selected for complex, multi‑constraint queries that smaller SLMs bounce. ",
5767+
"architecture": {
5768+
"modality": "text->text",
5769+
"input_modalities": [
5770+
"text"
5771+
],
5772+
"output_modalities": [
5773+
"text"
5774+
],
5775+
"tokenizer": "Other",
5776+
"instruct_type": null
5777+
},
5778+
"pricing": {
5779+
"prompt": "0.0000009",
5780+
"completion": "0.0000033",
5781+
"request": "0",
5782+
"image": "0",
5783+
"web_search": "0",
5784+
"internal_reasoning": "0"
5785+
},
5786+
"top_provider": {
5787+
"context_length": 131072,
5788+
"max_completion_tokens": null,
5789+
"is_moderated": false
5790+
}
5791+
}
5792+
},
5793+
{
5794+
"id": "arcee-ai/spotlight",
5795+
"created_at": "2025-05-05T23:45:52+02:00",
5796+
"display_name": "Arcee AI: Spotlight",
5797+
"provider": "openrouter",
5798+
"context_window": 131072,
5799+
"max_tokens": null,
5800+
"type": "chat",
5801+
"family": "other",
5802+
"supports_vision": true,
5803+
"supports_functions": true,
5804+
"supports_json_mode": true,
5805+
"input_price_per_million": 0.18,
5806+
"output_price_per_million": 0.18,
5807+
"metadata": {
5808+
"object": null,
5809+
"description": "Spotlight is a 7‑billion‑parameter vision‑language model derived from Qwen 2.5‑VL and fine‑tuned by Arcee AI for tight image‑text grounding tasks. It offers a 32 k‑token context window, enabling rich multimodal conversations that combine lengthy documents with one or more images. Training emphasized fast inference on consumer GPUs while retaining strong captioning, visual‐question‑answering, and diagram‑analysis accuracy. As a result, Spotlight slots neatly into agent workflows where screenshots, charts or UI mock‑ups need to be interpreted on the fly. Early benchmarks show it matching or out‑scoring larger VLMs such as LLaVA‑1.6 13 B on popular VQA and POPE alignment tests. ",
5810+
"architecture": {
5811+
"modality": "text+image->text",
5812+
"input_modalities": [
5813+
"image",
5814+
"text"
5815+
],
5816+
"output_modalities": [
5817+
"text"
5818+
],
5819+
"tokenizer": "Other",
5820+
"instruct_type": null
5821+
},
5822+
"pricing": {
5823+
"prompt": "0.00000018",
5824+
"completion": "0.00000018",
5825+
"request": "0",
5826+
"image": "0",
5827+
"web_search": "0",
5828+
"internal_reasoning": "0"
5829+
},
5830+
"top_provider": {
5831+
"context_length": 131072,
5832+
"max_completion_tokens": null,
5833+
"is_moderated": false
5834+
}
5835+
}
5836+
},
5837+
{
5838+
"id": "arcee-ai/virtuoso-large",
5839+
"created_at": "2025-05-05T23:01:25+02:00",
5840+
"display_name": "Arcee AI: Virtuoso Large",
5841+
"provider": "openrouter",
5842+
"context_window": 131072,
5843+
"max_tokens": null,
5844+
"type": "chat",
5845+
"family": "other",
5846+
"supports_vision": false,
5847+
"supports_functions": true,
5848+
"supports_json_mode": true,
5849+
"input_price_per_million": 0.75,
5850+
"output_price_per_million": 1.2,
5851+
"metadata": {
5852+
"object": null,
5853+
"description": "Virtuoso‑Large is Arcee's top‑tier general‑purpose LLM at 72 B parameters, tuned to tackle cross‑domain reasoning, creative writing and enterprise QA. Unlike many 70 B peers, it retains the 128 k context inherited from Qwen 2.5, letting it ingest books, codebases or financial filings wholesale. Training blended DeepSeek R1 distillation, multi‑epoch supervised fine‑tuning and a final DPO/RLHF alignment stage, yielding strong performance on BIG‑Bench‑Hard, GSM‑8K and long‑context Needle‑In‑Haystack tests. Enterprises use Virtuoso‑Large as the \"fallback\" brain in Conductor pipelines when other SLMs flag low confidence. Despite its size, aggressive KV‑cache optimizations keep first‑token latency in the low‑second range on 8× H100 nodes, making it a practical production‑grade powerhouse.",
5854+
"architecture": {
5855+
"modality": "text->text",
5856+
"input_modalities": [
5857+
"text"
5858+
],
5859+
"output_modalities": [
5860+
"text"
5861+
],
5862+
"tokenizer": "Other",
5863+
"instruct_type": null
5864+
},
5865+
"pricing": {
5866+
"prompt": "0.00000075",
5867+
"completion": "0.0000012",
5868+
"request": "0",
5869+
"image": "0",
5870+
"web_search": "0",
5871+
"internal_reasoning": "0"
5872+
},
5873+
"top_provider": {
5874+
"context_length": 131072,
5875+
"max_completion_tokens": null,
5876+
"is_moderated": false
5877+
}
5878+
}
5879+
},
5880+
{
5881+
"id": "arcee-ai/virtuoso-medium-v2",
5882+
"created_at": "2025-05-05T22:53:54+02:00",
5883+
"display_name": "Arcee AI: Virtuoso Medium V2",
5884+
"provider": "openrouter",
5885+
"context_window": 131072,
5886+
"max_tokens": null,
5887+
"type": "chat",
5888+
"family": "other",
5889+
"supports_vision": false,
5890+
"supports_functions": true,
5891+
"supports_json_mode": true,
5892+
"input_price_per_million": 0.5,
5893+
"output_price_per_million": 0.7999999999999999,
5894+
"metadata": {
5895+
"object": null,
5896+
"description": "Virtuoso‑Medium‑v2 is a 32 B model distilled from DeepSeek‑v3 logits and merged back onto a Qwen 2.5 backbone, yielding a sharper, more factual successor to the original Virtuoso Medium. The team harvested ~1.1 B logit tokens and applied \"fusion‑merging\" plus DPO alignment, which pushed scores past Arcee‑Nova 2024 and many 40 B‑plus peers on MMLU‑Pro, MATH and HumanEval. With a 128 k context and aggressive quantization options (from BF16 down to 4‑bit GGUF), it balances capability with deployability on single‑GPU nodes. Typical use cases include enterprise chat assistants, technical writing aids and medium‑complexity code drafting where Virtuoso‑Large would be overkill. ",
5897+
"architecture": {
5898+
"modality": "text->text",
5899+
"input_modalities": [
5900+
"text"
5901+
],
5902+
"output_modalities": [
5903+
"text"
5904+
],
5905+
"tokenizer": "Other",
5906+
"instruct_type": null
5907+
},
5908+
"pricing": {
5909+
"prompt": "0.0000005",
5910+
"completion": "0.0000008",
5911+
"request": "0",
5912+
"image": "0",
5913+
"web_search": "0",
5914+
"internal_reasoning": "0"
5915+
},
5916+
"top_provider": {
5917+
"context_length": 131072,
5918+
"max_completion_tokens": null,
5919+
"is_moderated": false
5920+
}
5921+
}
5922+
},
56645923
{
56655924
"id": "arliai/qwq-32b-arliai-rpr-v1:free",
56665925
"created_at": "2025-04-13T16:53:02+02:00",

0 commit comments

Comments
 (0)