See the project root AGENTS.md for repository-wide policies and workflows.
This directory contains model configuration and evaluation setup for the OpenHands SDK.
-
resolve_model_config.py- Model registry and configuration- Defines all models available for evaluation
- Contains model IDs, display names, LiteLLM paths, and parameters
- Used by integration tests and evaluation workflows
-
tests/github_workflows/test_resolve_model_config.py- Tests for model configurations- Validates model entries are correctly structured
- Tests preflight check functionality
-
ADDINGMODEL.md- Detailed guide for adding models (see below)
→ See ADDINGMODEL.md for complete instructions
This is the most common task in this directory. The guide covers:
- Required steps and files to modify
- Model feature categories and when to use them
- Integration testing requirements
- Common issues and troubleshooting
- Critical rules to prevent breaking existing models
If a model is failing in evaluations:
- Check the model configuration in
resolve_model_config.py - Review parameter compatibility (especially
temperature+top_pfor Claude) - Check if model is in correct feature categories in
openhands-sdk/openhands/sdk/llm/utils/model_features.py - Run preflight check:
MODEL_IDS="model-id" python resolve_model_config.py
Warning: Only update existing models if there's a confirmed issue. Working configurations should not be changed.
If you must update:
- Document why the change is needed (link to issue/PR showing the problem)
- Test thoroughly before and after the change
- Run integration tests to verify no regressions
This directory bridges model definitions with the evaluation system:
- Models defined here are available for integration tests
- Configuration includes LiteLLM routing and SDK-specific parameters
- Preflight checks validate model accessibility before expensive evaluation runs
- Tests ensure all models are correctly structured and resolvable