Skip to content

Releases: evalstate/fast-agent

v0.6.7

22 Mar 20:53

Choose a tag to compare

What's new

  • Reduce Markdown streaming flickering
  • Fix Opus/Sonnet 4.6 structured content API warning
  • Display README files from card pack installs/updates (--pack) option
  • Fix by @phucly95 to fix per-agent skill filtering
  • Fix by @peachgabba22 to fix MCP connectivity for non-persistent Servers.
  • Suppress uvloop / prompt_toolkit warnings on Python 3.14

New Contributors

Full Changelog: v0.6.1...v0.6.7

v0.6.1

17 Mar 20:02

Choose a tag to compare

What's Changed

  • GPT-5.4-mini/nano support
  • Remove SSE server support / migrate to FastMCP3 by @evalstate in #724

Full Changelog: v0.6.0...v0.6.1

v0.6.0 ❤️llama.cpp

16 Mar 19:51

Choose a tag to compare

fast-agent 0.6.0

llama.cpp support

  • Added llama.cpp model discovery, import, and launch-oriented picker flows to make local model setup much easier. Use with fast-agent model llamacpp.
  • Reads model settings (available context window size) and modalities, and creates a configuration file (model overlay). If you want to include sampling parameters from the server use --include-sampling-defaults.
image

Model Overlays

  • Added local model overlays so teams can define reusable local model aliases and runtime settings outside the built-in catalog.
  • Expanded model selection and switching UX across CLI, prompt UI, and ACP, with better cataloging, diagnostics, and reference handling.
  • https://fast-agent.ai/models/model_overlays/

Anthropic 1M Context, Google Native Refresh

  • Support for Anthropic 1M context window for 4.6 class models. Older models still require the extension flag e.g. claude-sonnet-4-5?context=1m
  • Improved Google Gemini support and general Gemini provider behavior.
  • Added support for Anthropic 1M-context workflows and broader Anthropic capability improvements.
  • Introduced a dedicated skills CLI and reworked the skills subsystem into clearer service and operations layers.
  • Improved ACP/MCP command handling, auth passthrough, session behavior, and command-surface parity.
  • Upgraded prompt and interactive shell experiences, including better completions, tool display, history views, and input/runtime structure.
  • Added more detailed history summaries, response telemetry, tool tracking, and streaming behavior coverage.
  • Refactored core runtime, agent loading, validation, and model resolution code to support the newer workflows more cleanly.

⚠️WARNING

  • Model Aliases are now known as Presets (e.g. kimi25).
  • Configuration Aliases are now known as References (e.g. $system.default, $system.fast). Configuration files will need to be updated if they defined aliases.
  • Large internal refactorings -- no major API changes but be careful upgrading regardless.

What's Changed

Full Changelog: v0.5.7...v0.6.0

v0.5.7

08 Mar 00:34

Choose a tag to compare

What's Changed

2026-03-07-gpt-5 4

OpenAI Responses Improvements and gpt-5.4 support

  • Support for gpt-5.4 (codexplan alias updated) and gpt-5.3-chat-instant (chatgpt alias)
  • Support for assistant "phase"
  • SDK version update
  • Service Tier selection fast and flex for supported models (use Shift+TAB in UI for convenience)
  • WebSocket is now the default transport - (use ?transport=sse for SSE connections) NB: OpenResponses remains SSE

Shell Improvements

  • read_text_file and write_text_file are supplied with shell access (Qwen 3.5/ACP compatibility). Configurable with write_text_file_mode (auto|on|off|apply_patch)
  • write_text_file is replaced with apply_patch for Responses models
  • apply_patch output is coloured post streaming
  • read_text_file previews enhanced and syntax highlight aware.
  • Skill prompt generation now uses explicit read-tool naming (read_text_file vs read_skill) based on
    runtime capability.

Tool Result passthrough

  • Option in request_params and model card to bypass LLM processing of tool outputs (intended for use with code-mode style subagents and MCP Servers).

Other UI/UX Improvements

  • service_tier, reasoning, verbosity, web_search and web_fetch are toggled with Shift+TAB, F6-F9 keys)
  • Added fast-agent model setup and fast-agent model doctor to manage model configuration and aliases
  • Added Codex login link in the model picker
  • fast-agent model setup now guides users through missing, invalid, and configured aliases
  • fresh model setup now defaults to $system.default
  • Use ##<agent> to send a message to an agent with silent return to the input buffer.
  • Interactive UI toolbar/model chip was upgraded with service-tier, reasoning, verbosity, web-search, and
    web-fetch indicators.

New prompt shortcuts and controls:

  • Shift+Tab service-tier cycle
  • F6 reasoning cycle
  • F7 verbosity cycle
  • F8 web search toggle
  • F9 web fetch toggle

Internal / Refactor / Tests

  • Shared formatting and marketplace-argument parsing helpers for command handlers.
  • Unified A3 header rendering utility used across CLI/command surfaces.
  • Docs layout switched to a docs git submodule pointer.
  • Broad test expansion for passthrough flow, filesystem runtime, apply_patch display/preview, quiet hash-
    send, display suppression, model controls, and ACP response-mode handling.
  • Improved cancellation/interrupt handling across interactive, CLI runner, ACP prompt handling, and tool-
    loop history reconciliation.
  • Better post-cancel recovery: stale progress row cleanup, clearer cancelled-turn messaging, and safer
    task cancellation state handling.
  • Improved read_text_file and shell result display behavior (cleaner summaries, truncation behavior,
    display suppression integration).
  • Refined MCP connection/reconnect handling and stream transport tracking hooks.
  • Result export reliability improvements for one-shot message runs (including transient result history
    capture when needed).

v0.5.6

01 Mar 21:34

Choose a tag to compare

What's Changed

  • Huge improvements to Markdown Streaming Performance
  • Major WebSockets improvements for Responses API.
  • Minimax M2.5 and Qwen 3.5 Support
  • Model Picker
  • MCP experimental sessions demonstrations
  • Other display enhancements

Full Changelog: v0.5.1...v0.5.6

fast-agent 0.5.1 codex and config

24 Feb 23:12

Choose a tag to compare

Codex and Config

  • GPT-5.3-codex support via API Key
  • All OpenAI Responses models can now be configured to use WebSocket (add ?transport=ws to the model string e.g. fast-agent -x --model responses.GPT-5.2?transport=ws)
  • Improvements to WebSocket handling/planning
  • fastagent.config.yaml and AgentCards can both use target style configuration for MCP
  • MCP Experimental Sessions support and demos

Full Changelog: v0.5.0...v0.5.1

fast-agent `0.5.0`

22 Feb 23:04

Choose a tag to compare

fast-agent 0.5.0 series

  • Support for "card packs" - adding, updating, publishing from command line and ACP/TUI. (fast-agent cards or /cards).

  • Improved model handling with support for aliases (e.g. $system.code or $system.fast). New /models command and CLI option.

  • Improved Smart Agent MCP functionality.

  • MCP Completion Support

  • NB This does not break API compatibility but fast-agent setup is now fast-agent scaffold and other automations may need command line updates.

What's Changed

Full Changelog: v0.4.54...v0.5.0

v0.4.54

18 Feb 00:11

Choose a tag to compare

fast-agent 0.4.54

Sonnet 4.6 and Web Search/Web Fetch features

Support for claude-sonnet-46, as well as Web Search and Fetch features (all supported models). Enable with ?web_search=on&web_fetch=on.

NB The older web tools appear to perform better, I will continue testing and revert/fix as needed.

Agent Card MCP enhancements / Smart Agent improvements.

Specify url, npx or uvx MCP connections directly from the Agent Card:

mcp_connect:
  - target: "https://hf.co/mcp"
    name: "hf_hub"
  - target: "npx @modelcontextprotocol/server-everything"
    name: "everything"

Smart Agent instructions can be added to prompts with:

{{internal:smart_prompt}} for the default smart agent prompt, {{internal:smart_agent_cards}} for the description section. Read more at https://fast-agent.ai/agents/instructions/#template-variables for other new template vars.

Web Socket Improvements

Improved Web Socket handling for codexplan models.

What's Changed

Full Changelog: v0.4.53...v0.4.54

v0.4.53

15 Feb 23:09

Choose a tag to compare

fast-agent 0.4.53 -- hot sockets and skills

What's Changed

  • Experimental: Codex Spark support. Use model string codexspark
  • Experimental: Websockets for OpenAI Plan - Use codexspark?transport=ws or codexplan?transport=ws to enable Web Socket connection.
  • Skill Update mechanism -- uses a small manifest to track git source and allow updating without manifest/marketplace peeking.
  • GLM-5 support via Hugging Face inference providers.
  • Introduce --agent and --noenv flags to make use in evaluations/testing simpler
  • ACP Registry compatibility and nested agent notifications (thanks @iqdoctor)
  • /mcp connect command to allow dynamic loading/reloading of MCP Servers. This works similar to the command line so /mcp connect https://huggingface.co and /mcp connect npx @modelcontextprotocol/server-everything wrok.
  • Improvements to URL Elicitation handling

Full Changelog: v0.4.49...v0.4.53

v0.4.49

07 Feb 17:38

Choose a tag to compare

fast-agent 0.4.49

GPT-5.3-Codex and Opus 4.6

Use GPT-5.3-Codex with the codexplan model (e.g. fast-agent --model codexplan. Alias added for GPT-5.2-Codex codexplan52.

Use Opus 4.6 with Adaptive Reasoning and 1m Context Window support. Alias opus now points to Opus 4.6. Adaptive reasoning is on by default, use opus?reasoning=[off|low|med|high|max to set effort hints. Use [?|&]longcontext=1m to enable extended context window.

.agentskills support.

Includes .agentskills/skills as a default skills search location.

Open Telemetry and other fixes

Patches and library updates to restore OpenTelemetry for Anthropic and MCP. Thanks @yarisoy .

Numerous defect fixes thanks to the ever diligent @iqdoctor .

What's Changed

Full Changelog: v0.4.45...v0.4.49