Skip to content

AI Provider Reference

AeroAgent supports 19 AI providers. All providers use the OpenAI-compatible chat completions API unless noted otherwise.

Provider Overview

#ProviderTypeBase URLAuth
1OpenAIopenaihttps://api.openai.com/v1API Key
2Anthropicanthropichttps://api.anthropic.com/v1API Key
3Google Geminigooglehttps://generativelanguage.googleapis.com/v1betaAPI Key
4xAI (Grok)xaihttps://api.x.ai/v1API Key
5OpenRouteropenrouterhttps://openrouter.ai/api/v1API Key
6Ollama (Local)ollamahttp://localhost:11434None
7Kimi (Moonshot)kimihttps://api.moonshot.cn/v1API Key
8Qwen (Alibaba)qwenhttps://dashscope-intl.aliyuncs.com/compatible-mode/v1API Key
9DeepSeekdeepseekhttps://api.deepseek.comAPI Key
10Mistralmistralhttps://api.mistral.ai/v1API Key
11Groqgroqhttps://api.groq.com/openai/v1API Key
12Perplexityperplexityhttps://api.perplexity.aiAPI Key
13Coherecoherehttps://api.cohere.com/compatibility/v1API Key
14Together AItogetherhttps://api.together.xyz/v1API Key
15AI21 Labsai21https://api.ai21.com/studio/v1API Key
16Cerebrascerebrashttps://api.cerebras.ai/v1API Key
17SambaNovasambanovahttps://api.sambanova.ai/v1API Key
18Fireworks AIfireworkshttps://api.fireworks.ai/inference/v1API Key
19CustomcustomUser-definedUser-defined

Feature Matrix

ProviderStreamingToolsVisionThinkingStructured Output
OpenAIYesYesYesYes (o3)Yes (strict:true)
AnthropicYesYesYesYesNo
Google GeminiYesYesYesYesNo
xAI (Grok)YesYesYesNoYes (strict:true)
OpenRouterYesYesYesVariesYes (strict:true)
OllamaYesYesYesVariesNo
KimiYesYesNoNoNo
QwenYesYesYesNoNo
DeepSeekYesYesNoYesNo
MistralYesYesYesNoNo
GroqYesYesYesNoNo
PerplexityYesNoNoNoNo
CohereYesTextNoNoNo
Together AIYesYesVariesNoNo
AI21 LabsYesYesNoNoNo
CerebrasYesYesNoNoNo
SambaNovaYesYesNoNoNo
Fireworks AIYesYesVariesNoNo
CustomYesConfigurableConfigurableNoNo

Models

AeroFTP does not ship with hardcoded default models. Use the Models button in AI Settings to fetch available models from each provider dynamically.

For Ollama, the Detect button queries GET /api/tags to list locally installed models. You can also pull new models directly from the UI via POST /api/pull with streaming progress.

Provider-Specific Notes

Anthropic

  • Uses native Anthropic API format (not OpenAI-compatible)
  • Prompt caching: Supports cache_control: { type: "ephemeral" } for 90% read discount on cached prefixes
  • System prompt: Sent as top-level system field
  • Thinking: Native thinking blocks with configurable budget (0-100K tokens)

Google Gemini

  • Uses Gemini-native API format
  • System instruction: Sent as top-level system_instruction field (not as a message)
  • Code execution: Supports executableCode / codeExecutionResult blocks
  • Context caching: gemini_create_cache for reusing large context prefixes

Ollama

  • Runs locally, no API key required
  • Model families: 8 detected profiles (llama, mistral, phi, gemma, qwen, deepseek, codellama, vicuna) with family-specific prompt styles
  • GPU monitoring: ollama_list_running shows active models and VRAM usage
  • Model pull: Download models from the Ollama registry with streaming progress bar

OpenRouter

  • Aggregator: routes to 100+ models from multiple providers
  • Uses OpenAI-compatible format
  • Structured output support depends on the underlying model

DeepSeek

Endpoint Format

DeepSeek's base URL is https://api.deepseek.com without a trailing /v1. The /v1/chat/completions path is appended by the client.

Cohere

Compatibility Endpoint

Cohere uses the compatibility endpoint at https://api.cohere.com/compatibility/v1, not the native /v2 API. Tool format is text-based, not native function calling.

stream_options

The stream_options: { include_usage: true } field is included in streaming requests to receive token usage in the final chunk. This field is excluded for Cohere and Perplexity, which reject unknown fields.

Provider Profiles

Each provider has a built-in personality profile that optimizes the system prompt for that provider's strengths:

  • Parameter presets: Creative, Balanced, and Precise temperature/top_p configurations
  • Capability awareness: The system prompt adapts based on whether the provider supports tools, vision, and thinking
  • Ollama model-specific templates: Prompt format adapts to the detected model family

Thinking Budget

For providers that support reasoning/thinking (Anthropic, OpenAI o3, Gemini, DeepSeek), AeroAgent offers 5 presets:

PresetToken BudgetUse Case
Off0Simple queries
Light1,024Quick reasoning
Balanced8,192General use
Deep32,768Complex analysis
Maximum100,000Full reasoning chain

A range slider allows fine-tuning between 0 and 100,000 tokens.

Adding a Custom Provider

Any OpenAI-compatible API can be added as a Custom provider:

  1. Go to Settings > AeroAgent > Providers
  2. Click Add Provider or use the Provider Marketplace
  3. Select Custom
  4. Enter the base URL and API key
  5. Use the Models button to fetch available models

The base URL should point to the root of the API (e.g., https://my-provider.com/v1). AeroFTP appends /chat/completions automatically.

Released under the GPL-3.0 License.