Overview
Conductor supports multiple AI models across different providers. Choose the right model based on your task complexity, latency requirements, and cost constraints.Anthropic (Claude)
Claude Sonnet 3.5
Model:claude-3-5-sonnet-20241022
Best for: Most tasks, best balance of intelligence and speed
Capabilities:
- Advanced reasoning and analysis
- Code generation and review
- Long-form content creation
- Multi-step problem solving
- Fast response times
Claude Opus 3
Model:claude-3-opus-20240229
Best for: Complex tasks requiring highest intelligence
Capabilities:
- Highest reasoning ability
- Complex analysis
- Expert-level tasks
- Research and synthesis
Claude Sonnet 3
Model:claude-3-sonnet-20240229
Best for: Balanced performance and cost
Capabilities:
- Strong reasoning
- General purpose tasks
- Good speed/quality balance
Claude Haiku 3
Model:claude-3-haiku-20240307
Best for: Fast, simple tasks
Capabilities:
- Fastest responses
- Simple classifications
- Quick summaries
- Low-cost operations
OpenAI (GPT)
GPT-4 Turbo
Model:gpt-4-turbo-preview
Best for: Latest GPT-4 capabilities
Capabilities:
- Advanced reasoning
- Code generation
- Complex problem solving
- Large context window (128K)
GPT-4
Model:gpt-4
Best for: Production-ready, most capable
Capabilities:
- Highest intelligence
- Reliable and stable
- Proven performance
GPT-3.5 Turbo
Model:gpt-3.5-turbo
Best for: Fast, cost-effective
Capabilities:
- Quick responses
- Simple tasks
- High throughput
- Low cost
Cloudflare Workers AI
Llama 3.1 (8B)
Model:@cf/meta/llama-3.1-8b-instruct
Best for: Free, edge-optimized, fast
Capabilities:
- No API key required
- Runs on Cloudflare’s edge
- Good for testing
- Free tier available
Llama 2 (7B)
Model:@cf/meta/llama-2-7b-chat-int8
Best for: Fast, small model
Capabilities:
- Faster responses
- Lower resource usage
- Simple tasks
Mistral 7B
Model:@cf/mistral/mistral-7b-instruct-v0.1
Best for: Alternative open model
Capabilities:
- Different approach
- Good for variety
- European model
Model Selection Guide
By Task Type
Code Generation
Recommended: Claude Sonnet 3.5, GPT-4Creative Writing
Recommended: Claude Sonnet 3.5, Claude Opus 3Classification/Tagging
Recommended: Claude Haiku 3, GPT-3.5 Turbo, Cloudflare LlamaAnalysis/Research
Recommended: Claude 3 Opus, GPT-4Summarization
Recommended: Claude Sonnet 3.5, Claude Haiku 3By Latency Requirements
Real-Time (under 1s)
- Claude Haiku 3
- GPT-3.5 Turbo
- Cloudflare Llama 3.1
Standard (1-3 seconds)
- Claude Sonnet 3.5
- GPT-4 Turbo
Complex (3-10 seconds)
- Claude Opus 3
- GPT-4
By Cost Constraints
Free Tier
- Cloudflare Workers AI models
Low Cost (under $1 per 1M tokens)
- Claude Haiku 3: 1.25/MTok
- GPT-3.5 Turbo: 1.50/MTok
Medium Cost (10 per 1M tokens)
- Claude Sonnet 3.5: 15/MTok
- Claude Sonnet 3: 15/MTok
- GPT-4 Turbo: 30/MTok
High Cost ($10+ per 1M tokens)
- Claude Opus 3: 75/MTok
- GPT-4: 60/MTok
Testing Strategy
Use cheaper models for development:Model Comparison
| Model | Provider | Speed | Quality | Cost | Best For |
|---|---|---|---|---|---|
| Claude Sonnet 3.5 | Anthropic | Fast | Excellent | Medium | Most tasks |
| Claude Opus 3 | Anthropic | Medium | Best | High | Complex analysis |
| Claude Haiku 3 | Anthropic | Fastest | Good | Low | Simple tasks |
| GPT-4 Turbo | OpenAI | Medium | Excellent | Medium-High | Latest features |
| GPT-4 | OpenAI | Medium | Excellent | High | Production |
| GPT-3.5 Turbo | OpenAI | Fast | Good | Low | High volume |
| Llama 3.1 8B | Cloudflare | Fast | Good | Free | Testing |
Best Practices
- Start with free models - Test with Cloudflare first
- Match model to task - Don’t over-spec for simple tasks
- Monitor costs - Track token usage
- Use temperature wisely - Lower for deterministic, higher for creative
- Test locally - Validate prompts before production
- Implement fallbacks - Have backup models
- Cache results - Avoid redundant calls
- Set max tokens - Control response length
- Version your prompts - Track what works
- Measure quality - A/B test different models

