o1-mini
Faster, cost-efficient reasoning model optimized for coding and agentic applications. 80% cheaper than o1-preview with strong capabilities for complex problem-solving in tasks where context is provided within the prompt.
Specifications
- Architecture
- Decoder-only Transformer
- License
- Proprietary
- Context Window
- 128,000 tokens
- Max Output
- 65,536 tokens
- Training Data Cutoff
- Sep 30, 2023
- Type
- text
- Modalities
- text
Benchmark Scores
American Invitational Mathematics Examination (AIME) problems test advanced mathematical problem-sol...
Advanced competitive programming benchmark for evaluating large language models on algorithmic probl...
Evaluates models on their ability to solve cybersecurity challenges across various domains including...
The FACTS Grounding Leaderboard evaluates LLMs' ability to generate factually accurate long-form res...
Graduate-level Problems in Quantitative Analysis (GPQA) evaluates advanced reasoning on graduate-lev...
Evaluates code generation capabilities by asking models to complete Python functions based on docstr...
A sample of 500 diverse problems from the MATH benchmark, spanning topics like probability, algebra,...
Massive Multitask Language Understanding (MMLU) tests knowledge across 57 subjects including mathema...
Advanced Specifications
- Model Family
- o-series
- API Access
- Available
- Chat Interface
- Available
- Multilingual Support
- Yes
Capabilities & Limitations
- Capabilities
- reasoningcodemathscience
- Known Limitations
- no function callingno structured outputsno streamingno system messageslacks broad world knowledge
- Notable Use Cases
- coding tasksdebugging complex codeagentic applicationscost-efficient reasoning
- Function Calling Support
- No
- Tool Use Support
- No