-
Notifications
You must be signed in to change notification settings - Fork 6.4k
Open
Labels
CLIIssues related to the Codex CLIIssues related to the Codex CLIbugSomething isn't workingSomething isn't workingcustom-modelIssues related to custom model providers (including local models)Issues related to custom model providers (including local models)
Description
What version of Codex is running?
codex-cli 0.63.0
What subscription do you have?
ChatGPT Plus
Which model were you using?
qwen3-coder-30b
What platform is your computer?
Microsoft Windows NT 10.0.26100.0 x64
What issue are you seeing?
When using a local model with LM Studio, which I have hosted on a different PC in the same network, the base_url, which contains the IP adress of the other computer, is ignored. This is my config.toml:
model_provider="lmstudio"
model = "qwen3-coder-30b"
[model_providers.lmstudio]
name = "lmstudio"
base_url = "http://xxx.xxx.xx.xx:1234/v1"
[profiles.qwen3-coder-30b]
model_provider = "lmstudio"
model = "qwen/qwen3-coder-30b"
After the CLI fails to connect it shows this error: Connection failed: error sending request for url (http://localhost:1234/v1/responses), which means that it didn't use the IP from the base_url but the localhost URL.
What steps can reproduce the bug?
Host a local model in you local network on a different machine and do the config as above.
What is the expected behavior?
It is supposed to use the base_url, otherwise it is not able to connect, since the URL is false.
Additional information
No response
otterammo
Metadata
Metadata
Assignees
Labels
CLIIssues related to the Codex CLIIssues related to the Codex CLIbugSomething isn't workingSomething isn't workingcustom-modelIssues related to custom model providers (including local models)Issues related to custom model providers (including local models)