Speed-critical tasks (3–5× faster than o1-preview)
Cost-sensitive STEM work (80% cheaper)
Streaming responses (unlike newer o-series)
Avoid For
All new development (deprecated — use o4-mini)
Non-STEM tasks (limited knowledge)
Function calling (not supported)
JSON output (not supported)
Fine-tuning (not supported)
Images/multimodal input (text only)
Long context (128K vs 200K in o4-mini)
128,000 context window
65,536 max output tokens
Oct 1, 2023 knowledge cutoff
Reasoning supported
Pricing
o1 Mini API pricing is based on token usage for input and output. Prices are listed per 1M tokens, with lower rates for cached input. Tool-specific features may add per-call fees.
Text tokens
Per 1M tokens
Input$1.10
Cached input$0.55
Output$4.40
Example costs (o1 Mini)
TASK
APPROX COST
Math problem (AIME)
~$0.01–$0.05
Competitive coding
~$0.02–$0.10
STEM reasoning
~$0.01–$0.08
Modalities
What the model can accept and produce
TextInput and output
ImagesNot supported
AudioNot supported
VideoNot supported
Features
Platform-level capabilities
StreamingSupported
Function callingNot supported
Structured outputsNot supported
Fine-tuningNot supported
DistillationNot supported
Predicted outputsNot supported
Tools
Tools supported by this model when using the Responses API.
Web searchNot supported
File searchSupported
Image generationNot supported
Code interpreterSupported
MCPNot supported
Computer useNot supported
Snapshots
o1 Mini model snapshots ensure stable behavior by locking a specific version. See all available snapshots and aliases below.