config.agent_model parameter. Each mode offers a different balance between capability and speed.
Agentic mode
Full agent with multi-step planning, reasoning, and tool execution. This is the default.agent_model | Description |
|---|---|
magpie-2.0 | Default. Agentic model with proactive planning and reasoning. |
magpie-2.5 | Beta. Our latest agentic model — faster, more adaptable, and built to handle a broader range of real-world tasks. |
magpie-1.1 | Previous-generation agentic model. |
Ask mode
Lightweight single-turn Q&A optimized for RAG (Retrieval-Augmented Generation) use cases. Faster than agentic mode with lower latency.agent_model | Description |
|---|---|
magpie-1.1-flash | Fast model that only supports the semantic_search tool. |
semantic_search tool is supported in Ask mode. Requests specifying other tools will be rejected. Structured outputs are not supported.
Fastest mode
Direct LLM response with no planning or tool execution. Lowest latency, best for simple conversational queries that don’t require data retrieval or actions.agent_model | Description |
|---|---|
llm-only | Direct LLM conversation with no tool calls. |
Choosing a mode
| Mode | Use when | Latency | Tools | Structured outputs |
|---|---|---|---|---|
| Agentic | You need multi-step reasoning, tool calls, or data analysis | Higher | All | Yes |
| Ask | You need fast answers from knowledge bases (RAG) | Medium | semantic_search only | No |
| Fastest | You need quick conversational responses without tools | Lowest | None | No |
config.agent_model is omitted, the API defaults to magpie-2.0 (Agentic mode).