这是一个关于免费的大模型api的合集,并精选了一部分模型
This is a collection of free LLM apis, and selected some models
我会尽可能更新维护这个项目(目前只有我一个人)
I will keep maintaining and updating this project to the best of my ability
入选原则是:限制请求速率而不是token > 尽可能多的来源 > 尽可能新且好的模型 > 足够用的请求速率
The selection criteria are: limit request rate over token count > more sources > newer and better models > sufficient rate limits
主要是有一定热度的文本模型
Primarily text models that have gained some popularity
目前只接受提供了OpenAI格式的API
At present, only accepted OpenAI-formated API
欢迎大家分享更多api
Welcome to share more apis
这个表格是由Gemini 2.5 Pro生成的,由Taple渲染
This table was generated by Gemini 2.5 Pro, Rendered by Taple
名称 / Name | API | 模型 / Models | 请求速率 / Rate Limits | 后台 / Dashboard | 注 / Tips |
---|---|---|---|---|---|
ChatAnywhere | https://api.chatanywhere.tech |
gpt-4o-mini |
Not Limited | https://api.chatanywhere.org/ |
|
硅基流动 / SiliconFlow | https://api.siliconflow.cn/v1 |
deepseek-ai/DeepSeek-R1-0528-Qwen3-8B Qwen/Qwen3-8B THUDM/glm-4-9b-chat``THUDM/GLM-4-9B-0414``THUDM/GLM-Z1-9B-0414 THUDM/GLM-4.1V-9B-Thinking |
1000 RPM (each model) | https://cloud.siliconflow.cn/bills |
|
OpenRouter | https://openrouter.ai/api/v1 |
deepseek/deepseek-r1-0528:free deepseek/deepseek-chat-v3-0324:free thudm/glm-z1-32b:free tencent/hunyuan-a13b-instruct:free moonshotai/kimi-k2:free z-ai/glm-4.5-air:free qwen/qwen3-coder:free qwen/qwen3-235b-a22b:free openai/gpt-oss-20b:free `` |
20 RPM / 200 RPD (each model) | https://openrouter.ai/activity |
|
书生 / Intern AI | https://chat.intern-ai.org.cn/api/v1 |
intern-latest |
10 RPM | https://internlm.intern-ai.org.cn/api/callDetail |
密钥有效期6个月 / The key is vailed for 6 months |
共享算力 / suanli.com | https://api.suanli.cn/v1 |
free:QwQ-32B |
Unknown | https://api.suanli.cn/detail |
算力由他人设备共享提供 / Shared computing by other people's devices |
Google Gemini | https://generativelanguage.googleapis.com/v1beta/openai |
gemini-2.5-pro |
5 RPM / 100 RPD | https://console.cloud.google.com/apis/api/generativelanguage.googleapis.com/metrics |
GFW |
↑ | ↑ | gemini-2.5-flash |
10 RPM / 250 RPD | ↑ | |
↑ | ↑ | gemini-2.5-flash-lite |
15 RPM / 1000 RPD | ↑ | |
Cohere | https://api.cohere.ai/compatibility/v1 |
command-a-03-2025 command-a-vision-07-2025 |
20 RPM | https://dashboard.cohere.com/billing |
绑定支付方式可以使用速率限制跟宽松的Production Key / Binding payment methods can use rate limiting and relaxed Production Key GFW |
Bigmodel | https://open.bigmodel.cn/api/paas/v4/ |
GLM-4-Flash-250414 GLM-Z1-Flash GLM-4.5-Flash |
只有并发数限制(均为30) / Only the number of concurrent transactions is limited (both 30). | ? | |
Github Models | https://models.github.ai/inference |
openai/gpt-4.1-mini openai/gpt-4.1 openai/gpt-4o |
15 RPM / 150 RPD | ? | 如果使用Azure API,可以使用更多模型 / If used Azure API, more models available |
这是我的另一个项目,建议配套使用
My another project, recommended for use together
示例配置文件/Sample Configuration File : asak.json
asak: for-the-zero/asak
- llm_benchmark:个人评测榜单,可信度高,而且收录更全 / A personal review list, it is highly credible, and it is more comprehensive
- Artifical Analysis
- lmsys lmarena