AI
Which Claude model runs each AI feature
Cheap features use Haiku, complex synthesis uses Sonnet, and document extraction can fall back to Opus.
Last updated May 3, 2026
Journey Builder picks the right model per feature so you don't overpay. The split:
Fast (Haiku 4.5)
Used for: Persona generation, Stage suggestions, Action plan generation. These are repetitive, structured tasks where Haiku produces the same quality as Sonnet for ~3× lower cost.
Pricing: $1 per million input tokens, $5 per million output. Cached reads cost ~10% of regular input.
Balanced (Sonnet 4.6)
Used for: Insights synthesis (combining all stages, personas, and pain points into a single strategic narrative). Sonnet is meaningfully better than Haiku for this kind of cross-context reasoning.
Pricing: $3 per million input tokens, $15 per million output.
Document extraction
By default, document import uses ANTHROPIC_MODEL from the environment (defaults to Opus 4.7 for backwards compatibility with the original implementation). For cost-sensitive deployments, override with Haiku — quality is acceptable for most CX docs.
Prompt caching
All AI routes use Anthropic's prompt caching: the system prompt and stable journey context are cached for ~5 minutes after each call. Re-running the same feature on the same journey within that window saves ~70% of input cost and ~50% of latency.
Rate limits
Anthropic's free tier is 5 RPM per model. If you hit it, Journey Builder waits 15 seconds and retries automatically — you'll see a brief "Rate limited" toast but the call usually succeeds. Heavy users on a paid Anthropic plan rarely see this.