Skip to content

Conversation

@abrichr
Copy link
Member

@abrichr abrichr commented Apr 5, 2025

  • Adds OMNIPARSER_DOWNSAMPLE_FACTOR config variable to allow scaling
    screenshots before parsing (default 1.0). Implemented via new
    utils.downsample_image function called from visual_state.py.
  • Adds LLM_PROVIDER and ANTHROPIC_DEFAULT_MODEL config variables.
  • Updates completions.py to use config, add tenacity retries, and
    restore/use format_chat_messages (guarded by new DEBUG_FULL_PROMPTS config flag).
  • Fixes test failures in test_visual_state.py related to mock data and patches.

Agent planning sequence for calculator task now appears correct with Sonnet
and downsample_factor=1.0, although perception (~7-11s) and planning
(~5-9s) steps remain slow. Downsampling factor < 1.0 can be enabled via
.env for performance testing, but may affect accuracy.

abrichr added 3 commits April 5, 2025 15:11
- Adds OMNIPARSER_DOWNSAMPLE_FACTOR config variable to allow scaling
  screenshots before parsing (default 1.0). Implemented via new
  utils.downsample_image function called from visual_state.py.
- Adds LLM_PROVIDER and ANTHROPIC_DEFAULT_MODEL config variables.
- Updates completions.py to use config, add tenacity retries, and
  restore/use format_chat_messages (guarded by new DEBUG_FULL_PROMPTS config flag).
- Fixes test failures in test_visual_state.py related to mock data and patches.

Agent planning sequence for calculator task now appears correct with Sonnet
and downsample_factor=1.0, although perception (~7-11s) and planning
(~5-9s) steps remain slow. Downsampling factor < 1.0 can be enabled via
.env for performance testing, but may affect accuracy.
@abrichr abrichr merged commit 7a562a1 into main Apr 5, 2025
1 check passed
@abrichr abrichr deleted the feat/downsampling branch April 5, 2025 19:19
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants