AI

Which Claude model runs each AI feature

Cheap features use Haiku, complex synthesis uses Sonnet, and document extraction can fall back to Opus.

Last updated May 3, 2026

Journey Builder picks the right model per feature so you don't overpay. The split:

Fast (Haiku 4.5)

Used for: Persona generation, Stage suggestions, Action plan generation. These are repetitive, structured tasks where Haiku produces the same quality as Sonnet for ~3× lower cost.

Pricing: $1 per million input tokens, $5 per million output. Cached reads cost ~10% of regular input.

Balanced (Sonnet 4.6)

Used for: Insights synthesis (combining all stages, personas, and pain points into a single strategic narrative). Sonnet is meaningfully better than Haiku for this kind of cross-context reasoning.

Pricing: $3 per million input tokens, $15 per million output.

Document extraction

By default, document import uses ANTHROPIC_MODEL from the environment (defaults to Opus 4.7 for backwards compatibility with the original implementation). For cost-sensitive deployments, override with Haiku — quality is acceptable for most CX docs.

Prompt caching

All AI routes use Anthropic's prompt caching: the system prompt and stable journey context are cached for ~5 minutes after each call. Re-running the same feature on the same journey within that window saves ~70% of input cost and ~50% of latency.

Rate limits

Anthropic's free tier is 5 RPM per model. If you hit it, Journey Builder waits 15 seconds and retries automatically — you'll see a brief "Rate limited" toast but the call usually succeeds. Heavy users on a paid Anthropic plan rarely see this.

More in AI