Skip to main content
Every agent needs a brain. On Abundly, you can choose which large language model (LLM) powers your agent—or simply let the platform pick the best one for you.
Agent settings showing model selection with Claude Sonnet Latest selected

Default behavior

If you don’t have a preference, leave the model set to (no preference). This is the default, and it means Abundly will use whatever model we think works best for general agentic behavior. We continuously evaluate models and update this default as better options become available. This is the right choice for most users—you get great performance without having to think about model selection at all.

Model aliases vs specific versions

When selecting a model, you can choose between:
  • Aliases like “Claude Sonnet Latest” — automatically points to the latest version that we have verified works well
  • Specific versions like “Claude Sonnet 4.5” — locked to that exact version
We recommend using aliases in most cases. Models improve constantly, and when a new version of Claude Sonnet arrives, it usually makes sense to upgrade. By selecting an alias, you won’t need to remember to do that—Abundly will point the alias to the latest version when we’ve verified it works well (typically after some internal testing). However, if you want full control and want to minimize the risk of surprises, pick a specific model version. Although newer models usually perform better, agent behavior can sometimes change subtly. Locking to a specific version lets you manually decide when to upgrade.

Available models

Abundly integrates with the leading AI providers. You can switch models at any time—even mid-conversation.

Comparing models

Not sure which model to choose? We tend to default to the Claude models (Sonnet or Opus). We find they work very well for agentic behaviour. But the other models have improved a lot lately, and some have specific strenghts that you may want to leverage. This is a changing landscape, so you’re probably best off researching online yourself (or asking an Agent to do it for you…). But here’s a high level overview of how the models from each provider compare against each other.

Anthropic Claude models

ModelBest forSpeedCostNotes
Claude Opus 4.5Complex reasoning, high-stakes tasksModerate$$$Most capable Claude model. Best for difficult problems that defeat Sonnet.
Claude Sonnet 4.5Everyday work, coding, analysisFast$$Recommended default. Excellent balance of capability and speed.
Claude Haiku 4.5High-volume, speed-critical tasksFastest$~3x faster than Sonnet at 1/3 the cost. Near-Sonnet quality for most tasks.
Start with Claude Sonnet for most use cases. Use Opus for your hardest problems and Haiku when you need speed at scale.

Google Gemini models

ModelBest forSpeedCostNotes
Gemini Pro 3Maximum reasoning depthModerate$$$Strongest reasoning. 1M token context for huge documents.
Gemini Flash 2.5Balanced performanceFast$$Good multimodal capabilities. Native audio support.
Gemini Flash Lite 2.5High-volume processingFastest$Most cost-effective. Great for bulk summarization and routing.
Gemini Flash is the workhorse—use it for most tasks. Choose Pro for complex reasoning and Flash Lite for high-volume, simple operations.

OpenAI GPT models

ModelBest forSpeedCostNotes
GPT 5.2 ProMission-critical, complex reasoningModerate$$$$Extended reasoning for highest accuracy. Best for high-stakes decisions.
GPT 5.2Latest improvements, reduced hallucinationsFast$$$30% fewer hallucinations. Strong on coding (55.6% SWE-Bench Pro) and reasoning.
GPT 5 MiniWell-defined tasks, interactive appsFaster$$Great balance of speed and capability. Ideal for chat and coding assistants.
GPT 5 NanoHigh-volume, simple tasksFastest$Ultra-cheap. Best for summarization, classification, and bulk processing.
GPT 5.2 is the recommended default for most tasks. Use Mini for speed-sensitive applications and Nano for high-volume simple operations.

Thinking tokens

For Claude models, you can enable thinking—an advanced reasoning mode where the model thinks through problems step by step before responding.
Thinking settings showing toggles for triggers and new chats, plus thinking budget configuration
Thinking makes your agent smarter at complex reasoning tasks, but it also increases processing time and credit usage. You can configure:
  • Thinking for triggers — Enable thinking when the agent responds to emails, schedules, and other triggers
  • Default thinking for new chats — Enable thinking by default in chat conversations
  • Thinking budget — Control how many tokens the agent can use for thinking (1,024–32,000)
Thinking is currently only available for Claude Sonnet and Opus models.

When to change models

ScenarioRecommendation
Just getting startedLeave it on “(no preference)“
Complex reasoning, planningClaude Opus, Gemini Pro, or GPT 5.2 Pro
Everyday tasks, codingClaude Sonnet, GPT 5.2, or GPT 5 Mini
High-volume, simple tasksClaude Haiku, Gemini Flash Lite, or GPT 5 Nano
Speed is criticalClaude Haiku, Gemini Flash, or GPT 5 Nano
Lowest hallucination ratesGPT 5.2 or GPT 5.2 Pro
Working with long documentsGemini Pro (1M token context)

Unified interface

Regardless of which model you choose, the platform provides a consistent experience:
  • Switch models even mid-conversation
  • Same capabilities work across all models
  • Consistent behavior and tool usage
  • No need to learn different APIs
You can switch models at any time. If you’re not happy with the results, try a different model—your agent’s instructions and capabilities stay the same.
Each model has different credit costs per token. More capable models like Opus and GPT 5.2 cost more per request, while Haiku and Flash Lite are very cost-effective. Check your usage dashboard to monitor credit consumption.
Yes! Each agent can have its own model preference. You might use a fast, cheap model for a high-volume support agent and a powerful reasoning model for a research agent.

Learn more