Skip to content

Releases: evalstate/fast-agent

v0.6.22 - HF Trace Export and Kimi 2.6

21 Apr 22:32
80af6c0

Choose a tag to compare

HF Trace Export

Export session traces in a Hugging Face Trace Viewer compatible format - and directly upload to a dataset. Available from /session export or fast-agent export --hf-dataset <repo>.

Kimi 2.6 and Kimi 2.6 Instant

Model aliases added for above models via Novita.

Other Changes

  • Improvements to ACP session, error and system prompt handling.

Full Changelog: v0.6.21...v0.6.22

v0.6.17 - opus 4.7 and more

16 Apr 21:55
b5c344f

Choose a tag to compare

What's Changed

Opus 4.7 Support

Includes "summarized display" feature to show reasoning summaries.

  • Adaptive reasoning is on by default. Set Reasoning Level and Task Budget with /model commands or
  • via model string - e.g. opus?reasoning=xhigh&task_budget=64000.

Access earlier versions with presets opus46 and opus45 or the full model slug.

Note that task_budget is only available for Opus 4.7

Assistant Response Banner

Long responses now show a banner at the final assistant message, to make location in the scrollback buffer easier. Use fast-agent config display to disable (Stream Reprint Banner).

image

Other Changes

  • Added Kimi25 instant mode alias
  • Enhanced session snapshotting

colour themes, llama router, tweaks

12 Apr 17:15

Choose a tag to compare

  • Experimental support for auto-configuration with llama.cpp as a router (shortcut added to model picker)
  • Updated default colour scheme (all still ANSI)
  • Fixes to hook display code, session new now resets counters.
  • MCP Servers can now be supplied by the ACP Client

v0.6.13 Overdue Release (Note) - 🚨 IMPORTANT DEPENDENCY UPDATE

09 Apr 23:42

Choose a tag to compare

fast-agent 0.6.13 -- 🚨FastMCP Update to 0.3.2🚨

Please update to this version ASAP.

OpenAI Responses API fix and FastMCP v0.3.2

This release contains a fix for a server-side change on the OpenAI WebSockets API and important security fixes for FastMCP.

Major Features

What's Changed

New Contributors

Full Changelog: v0.6.7...v0.6.13

v0.6.7

22 Mar 20:53

Choose a tag to compare

What's new

  • Reduce Markdown streaming flickering
  • Fix Opus/Sonnet 4.6 structured content API warning
  • Display README files from card pack installs/updates (--pack) option
  • Fix by @phucly95 to fix per-agent skill filtering
  • Fix by @peachgabba22 to fix MCP connectivity for non-persistent Servers.
  • Suppress uvloop / prompt_toolkit warnings on Python 3.14

New Contributors

Full Changelog: v0.6.1...v0.6.7

v0.6.1

17 Mar 20:02

Choose a tag to compare

What's Changed

  • GPT-5.4-mini/nano support
  • Remove SSE server support / migrate to FastMCP3 by @evalstate in #724

Full Changelog: v0.6.0...v0.6.1

v0.6.0 ❤️llama.cpp

16 Mar 19:51

Choose a tag to compare

fast-agent 0.6.0

llama.cpp support

  • Added llama.cpp model discovery, import, and launch-oriented picker flows to make local model setup much easier. Use with fast-agent model llamacpp.
  • Reads model settings (available context window size) and modalities, and creates a configuration file (model overlay). If you want to include sampling parameters from the server use --include-sampling-defaults.
image

Model Overlays

  • Added local model overlays so teams can define reusable local model aliases and runtime settings outside the built-in catalog.
  • Expanded model selection and switching UX across CLI, prompt UI, and ACP, with better cataloging, diagnostics, and reference handling.
  • https://fast-agent.ai/models/model_overlays/

Anthropic 1M Context, Google Native Refresh

  • Support for Anthropic 1M context window for 4.6 class models. Older models still require the extension flag e.g. claude-sonnet-4-5?context=1m
  • Improved Google Gemini support and general Gemini provider behavior.
  • Added support for Anthropic 1M-context workflows and broader Anthropic capability improvements.
  • Introduced a dedicated skills CLI and reworked the skills subsystem into clearer service and operations layers.
  • Improved ACP/MCP command handling, auth passthrough, session behavior, and command-surface parity.
  • Upgraded prompt and interactive shell experiences, including better completions, tool display, history views, and input/runtime structure.
  • Added more detailed history summaries, response telemetry, tool tracking, and streaming behavior coverage.
  • Refactored core runtime, agent loading, validation, and model resolution code to support the newer workflows more cleanly.

⚠️WARNING

  • Model Aliases are now known as Presets (e.g. kimi25).
  • Configuration Aliases are now known as References (e.g. $system.default, $system.fast). Configuration files will need to be updated if they defined aliases.
  • Large internal refactorings -- no major API changes but be careful upgrading regardless.

What's Changed

Full Changelog: v0.5.7...v0.6.0

v0.5.7

08 Mar 00:34

Choose a tag to compare

What's Changed

2026-03-07-gpt-5 4

OpenAI Responses Improvements and gpt-5.4 support

  • Support for gpt-5.4 (codexplan alias updated) and gpt-5.3-chat-instant (chatgpt alias)
  • Support for assistant "phase"
  • SDK version update
  • Service Tier selection fast and flex for supported models (use Shift+TAB in UI for convenience)
  • WebSocket is now the default transport - (use ?transport=sse for SSE connections) NB: OpenResponses remains SSE

Shell Improvements

  • read_text_file and write_text_file are supplied with shell access (Qwen 3.5/ACP compatibility). Configurable with write_text_file_mode (auto|on|off|apply_patch)
  • write_text_file is replaced with apply_patch for Responses models
  • apply_patch output is coloured post streaming
  • read_text_file previews enhanced and syntax highlight aware.
  • Skill prompt generation now uses explicit read-tool naming (read_text_file vs read_skill) based on
    runtime capability.

Tool Result passthrough

  • Option in request_params and model card to bypass LLM processing of tool outputs (intended for use with code-mode style subagents and MCP Servers).

Other UI/UX Improvements

  • service_tier, reasoning, verbosity, web_search and web_fetch are toggled with Shift+TAB, F6-F9 keys)
  • Added fast-agent model setup and fast-agent model doctor to manage model configuration and aliases
  • Added Codex login link in the model picker
  • fast-agent model setup now guides users through missing, invalid, and configured aliases
  • fresh model setup now defaults to $system.default
  • Use ##<agent> to send a message to an agent with silent return to the input buffer.
  • Interactive UI toolbar/model chip was upgraded with service-tier, reasoning, verbosity, web-search, and
    web-fetch indicators.

New prompt shortcuts and controls:

  • Shift+Tab service-tier cycle
  • F6 reasoning cycle
  • F7 verbosity cycle
  • F8 web search toggle
  • F9 web fetch toggle

Internal / Refactor / Tests

  • Shared formatting and marketplace-argument parsing helpers for command handlers.
  • Unified A3 header rendering utility used across CLI/command surfaces.
  • Docs layout switched to a docs git submodule pointer.
  • Broad test expansion for passthrough flow, filesystem runtime, apply_patch display/preview, quiet hash-
    send, display suppression, model controls, and ACP response-mode handling.
  • Improved cancellation/interrupt handling across interactive, CLI runner, ACP prompt handling, and tool-
    loop history reconciliation.
  • Better post-cancel recovery: stale progress row cleanup, clearer cancelled-turn messaging, and safer
    task cancellation state handling.
  • Improved read_text_file and shell result display behavior (cleaner summaries, truncation behavior,
    display suppression integration).
  • Refined MCP connection/reconnect handling and stream transport tracking hooks.
  • Result export reliability improvements for one-shot message runs (including transient result history
    capture when needed).

v0.5.6

01 Mar 21:34

Choose a tag to compare

What's Changed

  • Huge improvements to Markdown Streaming Performance
  • Major WebSockets improvements for Responses API.
  • Minimax M2.5 and Qwen 3.5 Support
  • Model Picker
  • MCP experimental sessions demonstrations
  • Other display enhancements

Full Changelog: v0.5.1...v0.5.6

fast-agent 0.5.1 codex and config

24 Feb 23:12

Choose a tag to compare

Codex and Config

  • GPT-5.3-codex support via API Key
  • All OpenAI Responses models can now be configured to use WebSocket (add ?transport=ws to the model string e.g. fast-agent -x --model responses.GPT-5.2?transport=ws)
  • Improvements to WebSocket handling/planning
  • fastagent.config.yaml and AgentCards can both use target style configuration for MCP
  • MCP Experimental Sessions support and demos

Full Changelog: v0.5.0...v0.5.1