Model Providers
By default, Plandex uses a mix of Anthropic, OpenAI, and Google models. OpenAI models call the OpenAI API directly, while all other models are called via OpenRouter.ai.
You can also use Plandex with models from any provider that provides an OpenAI-compatible API, like the aforementioned OpenRouter.ai, Together.ai, Replicate, Ollama, and more.
Model Performance
While you can use Plandex with many different providers and models as described above, Plandex's prompts have mainly been written and tested against the built-in models and model packs, so you should expect them to give the best results.
Built-in Models and Model Packs
Plandex provides a curated set of built-in models and model packs.
You can see the list of available model packs with:
\model-packs # REPL
plandex model-packs # CLI
You can see the list of available models with:
\models available # REPL
plandex models available # CLI
Integrated Models
If you use Plandex Cloud, you have the option of using Integrated Models Mode which allows you to use Plandex credits to pay for AI models. No separate accounts or API keys are required in this case.
If, alternatively, you use BYO API Key Mode with Plandex Cloud, or if you self-host Plandex, you'll need to generate API keys for the providers you want to use.
OpenRouter
Account
If you don't have an OpenRouter account, first sign up here.
API Key
Once you've created an OpenRouter account, generate an API key here.
OpenAI
Account
If you don't have an OpenAI account, first sign up here.
API Key
Once you've created an OpenAI account, generate an API key here.
Other Providers
Apart from those listed above, Plandex can use models from any provider that is compatible with the OpenAI API, like Together.ai, Replicate, Ollama, and more. You'll need to create an account and generate an API key for any other providers you plan on using.
Environment Variables
Now that you've generated API keys for your providers, export them as environment variables in your terminal.
export OPENROUTER_API_KEY=...
export OPENAI_API_KEY=...
# optional - set api keys for any other providers you're using
export TOGETHER_API_KEY...
If you're using OpenAI as a model provider, you can also set a different base URL for API calls:
export OPENAI_API_BASE=... # optional - set a different base url for OpenAI calls e.g. https://<your-proxy>/v1
If you have multiple OpenAI orgs, you can specify which org to use:
export OPENAI_ORG_ID=... # optional - set the OpenAI OrgID if you have multiple orgs