Try our new intelligent model routing solution, Arcee Conductor. Sign up today and get a $20 credit (~400M free tokens).

Introducing Arcee Conductor

A new standard for intelligent model routing.

Conductor intelligently routes your prompt to the best model, to efficiently deliver precise results, for any task.

The right model for each prompt, every time.

Try now

Today’s AI is about more than the single best large language model (LLM)

With new AI models released daily, it's hard to keep up with which model is best for your business. Our pioneering work in small language models (SLMs) gives us unique insights into which model is the right one for your tasks or queries.

That's why we built Arcee Conductor, a model-agnostic platform that gives you access to a complete suite of top-performing SLMs, as well as other industry-leading LLMs.

Arcee Conductor intelligently routes your query to the optimal model based on factors like industry/specialty, complexity, efficiency, and cost–all in an easy-to-use interface that requires no technical expertise.

How Arcee Conductor works

Arcee Conductor takes intelligent model routing to a new level

Precise, efficient routing

Automatically routes your prompt or query based on complexity, type of task, industry or domain, language, and whether it involves tool / function calling.

Slash your AI spend

Cost-per-query reduced by 85%+, thanks to simpler prompts getting routed to the smaller models. You're no longer paying premium prices for routine queries, and you get live visibility into query costs.

Routes to small, large models

The intelligent router evaluates your prompt or query, then routes it to an Arcee SLM (like Arcee-Blitz, Virtuoso-Medium, or Virtuoso-Large), or to a leading LLM (like Claude-Sonnet-3.7, GPT-4o, or DeepSeek-R1).

Advanced query processing

Supports chain-of-thought reasoning for enhanced analytical capabilities, and automatic compute scaling for task-specific optimization.

One platform, diverse models

The only routing platform that utilizes open-source and closed-source models, including our proprietary Arcee SLMs and our distilled DeepSeek models.

Customizable model settings

User-defined preferences for routing parameters and model prioritization, with preset profiles for different tasks.

Try our industry-leading small language models

Arcee Al creates powerful, highly efficient small language models (SLMs). They contain significantly fewer parameters than large language models, which makes them cost-effective and fast. But because they're purpose-built for specific tasks, data, and requirements, they perform remarkably well.

Explore Arcee Conductor

Modes

With "Auto" mode, Conductor chooses the best model to use, within a defined list of available models. Model selection is based on task and domain complexity.

For complex prompts, select "Auto Reasoning" and Conductor will choose the best reasoning model to return your carefully curated output.

To initiate a function call, select "Auto Tools", which uses Caller-Large, a 32B model specially trained to use external tools such as APIs and functions.

Pricing

To unlock additional options – like volume discounts, custom model configuration, and dedicated SLAs – contact Sales to learn about our  Enterprise Tier.

Mode: AUTO
Model
Price per Million Tokens (I/O)
Anthropic
Claude 3.7 Sonnet
$3.00 / $15.00
OpenAI
GPT-4.1
$2.00 / $8.00
Arcee AI
Virtuoso-Large
$0.75/ $1.20
Arcee AI
Virtuoso-Medium
$0.50 / $0.80
Arcee AI
Arcee-Blitz
$0.45 / $0.75
Mode: AUTO REASONING
Model
Price per Million Tokens (I/O)
Anthropic
Claude 3.7 Sonnet-Thinking
$3.00 / $15.00
OpenAI
OpenAI o3-mini
$1.10 / $4.40
DeepSeek
DeepSeek-R1
$3.00 / $7.00
Arcee AI
Maestro
$0.90 / $3.30
Mode: AUTO TOOLS
Model
Price per Million Tokens (I/O)
Arcee AI
Caller-Large
$0.55 / $0.85

Frequently
Asked Questions

Why is model routing important for your business?

Currently, companies face significant challenges:

  • Foundation model performance varies significantly and in unintuitive ways
  • High-margin AI products are creating financial pressure on traditional SaaS businesses
  • Teams cannot realistically track which model performs best for each specific prompt.

While selecting the best model for your task improves response quality, that isn't enough – on its own – to solve these challenges completely. This is why intelligent model routing is essential. It automatically selects the optimal model for each prompt, helping you to reclaim your profit margins without sacrificing quality. 

What exactly is Arcee Conductor?

Arcee Conductor is an intelligent model routing platform that directs each input to its ideal AI model based on complexity, domain, cost, and other requirements. By dynamically routing between large language models (LLMs) and small language models (SLMs), Conductor maximizes cost efficiency without compromising performance. You get the right model for each prompt, every time. 

What models power Arcee Conductor?

In Arcee Conductor, your prompt is automatically routed to the most suitable model through an advanced routing mechanism.

Available models include:
Arcee AI SLMs: Virtuoso-Large, Virtuoso-Medium, Arcee-Blitz, Maestro-32b and Caller-Large
LLMs: Claude-Sonnet-3.7, GPT-4.1, Claude-Sonnet-3.7-Thinking, o3-mini, DeepSeek-R1

What is a mode on Arcee Conductor? 

A mode on Arcee Conductor is a selection of specialized AI models grouped by functionality. Each mode serves a different purpose, giving you the flexibility to use the most suitable approach for your task, whether for general inquiries, complex reasoning, or function-calling.

What modes are available on Arcee Conductor?

  • Auto Mode: Intelligently route a prompt to the most suitable general-purpose model based on complexity and efficiency. Ideal for everyday questions and tasks.
  • Auto Reasoning: Automatically selects the optimal reasoning model for complex prompts that need problem breakdown into smaller steps. Ideal for tasks requiring detailed, step-by-step solutions.
  • Auto Tools: Provides models specifically designed for function-calling, allowing integration with external systems and workflow execution.

Each mode serves a specific purpose to deliver the most effective results for your particular needs.

How do I get started with Arcee Conductor?

You can sign up here to begin using Arcee Conductor today.  First-time users receive $20 in credits (equivalent to approximately 400 million tokens) towards your Conductor usage. 

How does Arcee Conductor's billing work?

Arcee Conductor is based on a usage-based pricing model. When you sign up for Conductor for the first time,  you'll receive $20 in credits automatically. This allows you to immediately start using the platform.

After you've used your $20 in credits, charges will be applied to your payment method based on your usage. The specific rates vary depending on specific rates per token of models. 

What are the specific rates per token in Arcee Conductor?

                      Model

Input Tokens
(per million tokens)

Output Tokens
(per million tokens)

Arcee-Blitz 
(General Purpose)

$0.45

$0.75

Virtuoso-Medium
(General Purpose)

$0.50

$0.80

Virtuoso-Large
(General Purpose)

$0.75

$1.20

Maestro
(Reasoning)

$0.90

$3.30

Caller-Large
(Function Calling)

$0.55

$0.85

Claude 3.7 Sonnet

$3.00

$15.00

GPT-4.1

$2.00

$8.00

DeepSeek-R1

$3.00

$7.00

OpenAI o3-mini

$1.10

$4.40

Industry-leading SLMs & LLMs with unified API access, inference, and intelligent routing
all in one solution.

First-time users of Arcee Conductor can get started with a one-time $20 credit. Optimize output across models, reduce usage costs, and maximize performance with intelligent routing.

Get started