< cd ..

Plandex

~/plandex-ai/plandex

AI coding agent for large projects

●●● stats.json

latest

server/v2.2.1

total

30

stable

30

scraped

Mar 2, 2026

#ai #cli #coding-agent #large-projects

/changelog

cli/v2.2.1 [Jul 16, 2025] details → github →

# Release cli/v2.2.1

🖇️  Connect your Claude Pro or Max subscription

If you have a Claude Pro or Max subscription, Plandex can use it when calling Anthropic models. You can use it in either Integrated Models Mode on Plandex Cloud, or in BYO Key Mode (whether on Cloud or self-hosting).

Assuming you're using Anthropic models (which the default model pack does), you'll be asked if you want to connect your Claude subscription the first time you run Plandex. Follow the instructions to connect.

Learn more in the docs.

🐞  Bug fixes

Fixed an issue with custom models and providers.

cli/v2.2.0 [Jul 1, 2025] details → github →

# Release cli/v2.2.0

This is a big release that is mainly focused on Plandex's model provider and model config system. It significantly increases model provider flexibility, makes custom model configuration much easier, reduces costs on Cloud, and adds built-in support for Ollama.

🔌  Model provider flexibility for BYO key mode

  • Now when using Plandex in BYO key mode (either Cloud or self-hosted), you can easily use Plandex with a wide range of built-in model providers.
  • Apart from OpenRouter and the OpenAI API (which were already built-in), built-in providers now include:
  • Anthropic API
  • Google AI Studio
  • Google Vertex AI
  • Azure OpenAI
  • AWS Bedrock
  • DeepSeek API
  • Perplexity API
  • Ollama (for local models—see below for details)
!plandex-model-providers

🛟  Provider fallback

  • When API keys/credentials are provided for multiple providers for a model, Plandex will fail over to the last valid provider if the first one fails. This is especially useful when using a direct provider (like OpenAI or Anthropic) alongside OpenRouter. If the direct call fails, Plandex will fall back to OpenRouter, which has its own internal fallback system across multiple providers. You get the best of both worlds: direct access by default, with no additional cost or latency, plus multi-layered resilience in case of stability issues.

💰  5.5% price reduction for Plandex Cloud Integrated Models mode

  • Thanks to the new model provider system described above, Plandex Cloud with Integrated Models mode can now make direct provider calls under the hood rather than defaulting to OpenRouter, which allows us to avoid OpenRouter's 5.5% surcharge on model calls and pass the savings on to you. OpenRouter is still used as a fallback for added resilience.

⚙️  JSON-based model config with IDE support

  • Plandex now supports JSON-based model config, which make it much easier to try out different models, different settings, and to use custom models, providers, and model packs.
  • The new JSON model config system integrates cleanly with your IDE or editor. When you first edit model settings, Plandex will prompt you to set a preferred editor, and the JSON config file will be opened in that editor.
  • Model config files use JSON schemas, which allows for autocomplete, validation, and inline documentation/type hints in most IDEs and editors.

set-model and set-model default commands

  • set-model has been simplified to work with the new system. If run without arguments, you'll be prompted to either select a built-in or custom model pack, or to directly edit the current plan's model config inline in JSON. You can also pass it a model pack name (set-model daily-driver) or jump straight to the JSON settings with set-model --json.
  • set-model default works the same way, but allows you to configure the default model settings for new plans.
!plandex-model-settings-json

models custom command

  • models custom is a new all-in-one command for managing custom providers, models, and model packs in one place. It replaces the models add, models delete, model-packs create, model-packs update, and model-packs delete commands.
  • The first time you run it, if you haven't already configured any custom models or model packs, an example config file will be created to get you started. If you do already have custom models or model packs configured, the config file will be populated with those models and model packs.
!plandex-custom-models-json

models and models default commands

  • The models and models default commands now show simplified output by default, with a new --all flag to show all properties.
  • These commands also now show all fallback models (for large context, large output, errors, etc.) for each role, including multiple levels of fallback, which previously weren't always included in the output.

🦙  Built-in Ollama support

  • Plandex now offers built-in support for Ollama, which makes it much easier to use local models. Check out the new Ollama quickstart for details.
!plandex-ollama

📖  Built-in models and models packs

  • The docs now include all built-in models and model packs, making it easier to see what's available, and the settings for each model/model pack.

🧠  New built-in models

  • mistral/devstral-small, with both OpenRouter and Ollama providers.
  • The Qwen 3 series of models, from 8B to 235B, available with cloud variants for OpenRouter and local variants for Ollama.
  • Distilled local versions of deepseek/r1, from 8b to 70b, available with Ollama provider.

🎛️  New model packs

  • The gemini-exp model pack has been removed, and in its place there's now a new gemini-planner model pack, which uses Gemini 2.5 Pro for planning and context selection, and the default models for other roles, as well as a new google model pack, which uses either Gemini 2.5 Pro or Gemini 2.5 Flash for all roles.
  • A new o3-planner model pack has been added, which uses OpenAI o3-medium for planning and context selection, and the default models for other roles.

🔄  Other built-in model updates

  • Built-in Gemini 2.5 Pro and Gemini 2.5 Flash models now use the latest model identifiers (replacing old 2.5 Pro Preview and 2.5 Flash Preview identifiers)
  • The gemini-preview model pack has been removed, and a new gemini-planner model pack has been added, which uses Gemini 2.5 Pro for planning and context selection, and the default models for other roles.
  • OpenAI o3 models have had their pricing drastically reduced when using Plandex Cloud with Integrated Models mode—input tokens now cost $2/M, output $8/M, an 80% reduction.
  • The deepseek/r1 model has been updated to use the latest model identifier (deepseek/deepseek-r1-0528) on OpenRouter.

🐞  Bug fixes

  • Fixed a file mapping bug for TypeScript files that caused directly exported symbols like export const foo = 'bar' to be omitted from map files. Also improved TypeScript mapping support for some other constructs like declare global, namespace, and enum blocks, and improved handling of arrow functions. Thanks to @mnahkies for the PR identifying this.

🔧  Other changes

  • plandex checkout now has a --yes/-y flag to auto-confirm creating a new branch if it doesn't exist, so the command can be used for scripting with no user interaction.
  • plandex tell, plandex continue, and plandex build all now support a --skip-menu flag to skip the interactive menu that appears when the response finishes and changes are pending. There's also a new skip-changes-menu config setting that can be set to true to skip this menu by default.
server/v2.1.8 [Jun 8, 2025] details → github →

# Release server/v2.1.8

  • Fix for potential hang in file map queue
server/v2.1.7 [Jun 4, 2025] details → github →

# Release server/v2.1.7

  • Fix for "conversation summary timestamp not found in conversation" error (https://github.com/plandex-ai/plandex/issues/274)
  • Fix for potential panic/crash during plan stream (https://github.com/plandex-ai/plandex/issues/275)
  • Better protection against panics/crashes in server goroutines across the board
cli/v2.1.6+1 [May 27, 2025] details → github →

# Release cli/v2.1.6+1

  • Error handling fix
  • Fix for some roles in the daily-driver model pack that weren't correctly updated to Sonnet 4 in 2.1.6
  • Added fallback from Sonnet 4 to Sonnet 3.7 to deal with occasional provider errors and rate limit issues
cli/v2.1.6 [May 27, 2025] details → github →

# Release cli/v2.1.6

  • The newly released Claude Sonnet 4 is now stable in testing, so it now replaces Sonnet 3.7 as the default model for context sizes under 200k across all model packs where 3.7 was previously used.
  • A new strong-opus model pack is now available. It uses Claude Opus 4 for planning and coding, and is otherwise the same as the 'strong' pack. Use it with \set-model strong-opus to try it out.
  • The opus-4-planner model pack that was introduced in 2.1.5 has been renamed to opus-planner, but the old name is still supported. This model pack uses Claude Opus 4 for planning, and the default models for other roles.
  • Fix for occasional garbled error message when the model is unresponsive.
  • Fix for occasional 'couldn't aquire lock' error after stream finishes.
  • Additional retry when model is unresponsive or hits provider rate limits—helps particularly with new Opus 4 model on OpenRouter.
cli/v2.1.5 [May 24, 2025] details → github →

# Release cli/v2.1.5

  • Added newly released Claude Sonnet 4 and Claude Opus 4 as built-in models.
  • Sonnet 4 isn't yet used in the default 'daily-driver' model pack due to sporadic errors in early testing, but it can be used with the 'sonnet-4-daily' model pack (use '\set-model sonnet-4-daily' to use it). It will be promoted to the default model pack soon.
  • Opus 4 can be used with the 'opus-4-planner' model pack ( '\set-model opus-4-planner'), which uses Opus 4 for planning and Sonnet 4 for coding.
  • Removed error fallbacks for o4-mini and gemini-2.5-pro-preview.
server/v2.1.4 [May 20, 2025] details → github →

# Release server/v2.1.4

  • Fix to remove occasional extraneous blank lines from start/end of edited files.
server/v2.1.3 [May 19, 2025] details → github →

# Release server/v2.1.3

  • Fix for 'panic in execTellPlan' error when using a model pack that doesn't explicitly set the 'coder' or 'whole-file-builder' roles
cli/v2.1.3 [May 15, 2025] details → github →

# Release cli/v2.1.3

  • Fix for default model pack not being correctly applied to new plans
  • Fix for potential crash on Linux when applying a plan
server/v2.1.2 [May 8, 2025] details → github →

# Release server/v2.1.2

  • Fix for auto-load context error: 'Error decoding response → EOF'
cli/v2.1.2 [May 8, 2025] details → github →

# Release cli/v2.1.2

  • Fix for rare auto-load context timeout error when no files are loaded.
server/v2.1.1+1 [May 5, 2025] details → github →

# Release server/v2.1.1+1

  • Improve error handling to catch yet another "context length exceeded" error message variation from Anthropic.
cli/v2.1.1 [May 2, 2025] details → github →

# Release cli/v2.1.1

  • Fix for free Gemini 2.5 Pro Experimental OpenRouter endpoint.
  • Retries for "No endpoints found that support cache control" error that showed up when OpenRouter temporarily disabled caching for Gemini 2.5 Pro Preview.
  • Other minor improvements to error handling and retries.
server/v2.1.0+1 [Apr 30, 2025] details → github →

# Release server/v2.1.0+1

  • Fix for context length exceeded error that still wasn't being caught and retried by the fallback correctly.
cli/v2.1.0+1 [Apr 30, 2025] details → github →

# Release cli/v2.1.0+1

  • Fix for potential encoding issue when loading files into context.
cli/v2.1.0 [Apr 29, 2025] details → github →

# Release cli/v2.1.0

🚀  OpenRouter only for BYO key

  • When using a BYO key mode (either cloud or self-hosted), you can now use Plandex with only an OpenRouter.ai account and OPENROUTER_API_KEY set. A separate OpenAI account is no longer required.
  • You can still use a separate OpenAI account if desired by setting the OPENAI_API_KEY environment variable in addition to OPENROUTER_API_KEY. This will cause OpenAI models to make direct calls to OpenAI, which is slightly faster and cheaper.

🧠  New Models

Gemini

  • Google's Gemini 2.5 Pro Preview is now available as a built-in model, and is the new default model when context is between 200k and 1M tokens.
  • A new gemini-preview model pack has been added, which uses Gemini 2.5 Pro Preview for planning and coding, and default models for other roles. You can use this pack by running the REPL with the --gemini-preview flag (plandex --gemini-preview), or with \set-model gemini-preview from inside the REPL. Because this model is still in preview, a fallback to Gemini 1.5 Pro is used on failure.
  • Google's Gemini Flash 2.5 Preview is also now available as a built-in model. While it's not currently used by default in any built-in model packs, you can use with \set-model or a custom model pack.

OpenAI

  • OpenAI's o4-mini is now available as a built-in model with high, medium, and low reasoning effort levels. o3-mini has been replaced by the corresponding o4-mini models across all model packs, with a fallback to o3-mini on failure. This improves Plandex's file edit reliability and performance with no increase in costs. o4-mini-medium is also the new default planning model for the cheap model pack.
  • OpenAI's o3 is now available as a built-in model with high, medium, and low reasoning effort levels. Note that if you're using Plandex in BYO key mode, OpenAI requires an organization verification step before you can use o3.
  • o3-high is the new default planning model for the strong model pack, replacing o1. Due to the verification requirements for o3, the strong pack falls back to o4-mini-high for planning if o3 is not available.
  • OpenAI's gpt-4.1, gpt-4.1-mini, and gpt-4.1-nano have been added as built-in models, replacing gpt-4o and gpt-4o-mini in all model packs that used them previously.
  • gpt-4.1 is now used as a large context fallback for the default coder role, effectively increasing the context limit for the implementation phase from 200k to 1M tokens.
  • gpt-4.1 is also the new coder model in the cheap model pack, and is also the new main planning and coding model in the openai model pack.

🛟  Model Fallbacks

  • In order to better incorporate newly released models and preview models that may have initial reliability or capacity issues, a more robust fallback and retry system has been implemented. This will allow for faster introduction of new models in the future while still maintaining a high level of reliability.
  • Fallbacks for 'context length exceeded' errors have also been improved, so that these errors will now trigger an automatic fallback to a model with a larger context limit if one is defined in the model pack. This will fix issues like https://github.com/plandex-ai/plandex/issues/232 where the stream errors with a 400 or 413 error when context is exceeded instead of falling back correctly.

💰  Gemini Caching

  • Gemini models now support prompt caching, significantly reducing costs and latency during planning, implementation, and builds when using Gemini models.

🤫  Quieter Reasoning

  • When using Claude 3.7 Sonnet thinking model in the reasoning AND strong model packs, reasoning is no longer included by default. This clears up some issues that were caused by output with specific formatting that Plandex takes action on being duplicated between the reasoning and the main output. It also feels a bit more relaxed to keep the reasoning behind-the-scenes, even though there can be a longer wait for the initial output.

💻  REPL Improvements

  • Additional handling of possibly incorrect or mistyped commands in the REPL. Now apart from suggesting commands only based on possibly mistyped backslash commands, any likely command with or without the backslash will suggest possible commands rather than sending the prompt straight to the AI model, which can waste tokens due to minor typos or a missing backslash.

☁️  Plandex Cloud

  • If you started a free trial of Plandex Cloud with BYO Key mode, you can now switch to a trial of Integrated Models mode if desired from your billing dashboard (use \billing from the REPL to open the dashboard).
  • When doing a trial in Integrated Models mode, you will now be warned when your trial credits balance goes below $1.00.
  • In Integrated Models mode, the required number of credits to send a prompt is now much lower, so you can use more credits before getting an 'Insufficient credits' message.

🐞  Bug Fixes

  • Fix for 'Plan replacement failed' error during file edits on Windows that was caused by mismatched line endings.
  • Fix for 'tool calls not supported' error for custom models that use the XML output format (https://github.com/plandex-ai/plandex/issues/238).
  • Fix for errors in some roles with Anthropic models when only a single system message was sent (https://github.com/plandex-ai/plandex/issues/208).
  • Fix for potential back-pressure issue with large/concurrent project map operations.
  • Plandex Cloud: fix for JSON parsing error on payment form when the card is declined. It will now show the proper error message.
cli/v2.0.7+1 [Apr 8, 2025] details → github →

# Release cli/v2.0.7+1

  • Small adjustment to previous release: in the REPL, select the first auto-complete suggestion on 'enter' if any suggestions are listed.
cli/v2.0.7 [Apr 8, 2025] details → github →

# Release cli/v2.0.7

  • Better handling of partial or mistyped commands in the REPL. Rather than falling through to the AI model, a partial \ command that matches only a single option will default to that command. If multiple commands could match, you'll be given a list of options. For input that begins with a \ but doesn't match any command, there is now a confirmation step. This helps to prevent accidentally sending mistyped commands the model and burning tokens.
server/v2.0.6 [Apr 3, 2025] details → github →

# Release server/v2.0.6

  • Improvements to process management and cleanup for command execution
  • Remove extraneous model request logging
cli/v2.0.6 [Apr 3, 2025] details → github →

# Release cli/v2.0.6

  • Timeout for 'plandex browser' log capture command
  • Better failure handling for 'plandex browser' command
server/v2.0.5 [Apr 2, 2025] details → github →

# Release server/v2.0.5

  • Fix for a bug that was causing occasional model errors. Model calls should be much more reliable now.
  • Better error handling and error messages for model errors (rate limits or other errors).
  • No error retries for rate limit errors.
  • Fixed bug that caused retries to add the prompt to the conversation multiple times.
  • Error responses with no output no longer create a log entry.
cli/v2.0.5 [Apr 2, 2025] details → github →

# Release cli/v2.0.5

  • Consolidated to a single model pack for Gemini 2.5 Pro Experimental: 'gemini-exp'. Use it with 'plandex --gemini-exp' or '\set-model gemini-exp' in the REPL.
  • Prevent the '\send' command from being included in the prompt when using multi-line mode in the REPL.